[ 490.624544] env[62824]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' {{(pid=62824) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 490.624900] env[62824]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' {{(pid=62824) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 490.624940] env[62824]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' {{(pid=62824) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 490.625291] env[62824]: INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs [ 490.720524] env[62824]: DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm {{(pid=62824) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:383}} [ 490.731080] env[62824]: DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 0 in 0.010s {{(pid=62824) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:421}} [ 491.330898] env[62824]: INFO nova.virt.driver [None req-44a70650-88f1-4373-a328-642795d27196 None None] Loading compute driver 'vmwareapi.VMwareVCDriver' [ 491.401797] env[62824]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "oslo_vmware_api_lock" by "oslo_vmware.api.VMwareAPISession._create_session" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 491.401949] env[62824]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" acquired by "oslo_vmware.api.VMwareAPISession._create_session" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 491.402065] env[62824]: DEBUG oslo_vmware.service [-] Creating suds client with soap_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk' and wsdl_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk/vimService.wsdl' {{(pid=62824) __init__ /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:242}} [ 494.666078] env[62824]: DEBUG oslo_vmware.service [-] Invoking ServiceInstance.RetrieveServiceContent with opID=oslo.vmware-0c309844-6879-4346-abe7-336ae121281e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 494.683161] env[62824]: DEBUG oslo_vmware.api [-] Logging into host: vc1.osci.c.eu-de-1.cloud.sap. {{(pid=62824) _create_session /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:242}} [ 494.683307] env[62824]: DEBUG oslo_vmware.service [-] Invoking SessionManager.Login with opID=oslo.vmware-e23fa0fb-4151-44f6-85f0-66e25224babe {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 494.720658] env[62824]: INFO oslo_vmware.api [-] Successfully established new session; session ID is 4cd49. [ 494.720819] env[62824]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" "released" by "oslo_vmware.api.VMwareAPISession._create_session" :: held 3.319s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 494.721431] env[62824]: INFO nova.virt.vmwareapi.driver [None req-44a70650-88f1-4373-a328-642795d27196 None None] VMware vCenter version: 7.0.3 [ 494.724923] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96b2c3d8-167b-4ad0-a13b-b8e38003fa9d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 494.748239] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-600ea1d8-f6f9-451a-87f2-a0674f7f646c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 494.754943] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d08b4707-061c-4d37-8a99-11b4e8c11b98 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 494.762181] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f730a22a-d863-4842-b639-998db850112d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 494.776088] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bff52627-7d70-4682-9b92-79b9f9997b6a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 494.782466] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c9ca834-1e57-4fa9-a22e-0f59b144327a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 494.813357] env[62824]: DEBUG oslo_vmware.service [-] Invoking ExtensionManager.FindExtension with opID=oslo.vmware-029fae6d-6066-4e23-b9e1-5b6e03320c41 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 494.819174] env[62824]: DEBUG nova.virt.vmwareapi.driver [None req-44a70650-88f1-4373-a328-642795d27196 None None] Extension org.openstack.compute already exists. {{(pid=62824) _register_openstack_extension /opt/stack/nova/nova/virt/vmwareapi/driver.py:228}} [ 494.821911] env[62824]: INFO nova.compute.provider_config [None req-44a70650-88f1-4373-a328-642795d27196 None None] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access. [ 495.325742] env[62824]: DEBUG nova.context [None req-44a70650-88f1-4373-a328-642795d27196 None None] Found 2 cells: 00000000-0000-0000-0000-000000000000(cell0),62bd51a0-02df-41ea-b321-4bd1095178fd(cell1) {{(pid=62824) load_cells /opt/stack/nova/nova/context.py:464}} [ 495.327868] env[62824]: DEBUG oslo_concurrency.lockutils [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 495.328114] env[62824]: DEBUG oslo_concurrency.lockutils [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 495.328812] env[62824]: DEBUG oslo_concurrency.lockutils [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 495.329259] env[62824]: DEBUG oslo_concurrency.lockutils [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] Acquiring lock "62bd51a0-02df-41ea-b321-4bd1095178fd" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 495.329440] env[62824]: DEBUG oslo_concurrency.lockutils [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] Lock "62bd51a0-02df-41ea-b321-4bd1095178fd" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 495.330442] env[62824]: DEBUG oslo_concurrency.lockutils [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] Lock "62bd51a0-02df-41ea-b321-4bd1095178fd" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 495.350987] env[62824]: INFO dbcounter [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] Registered counter for database nova_cell0 [ 495.358962] env[62824]: INFO dbcounter [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] Registered counter for database nova_cell1 [ 495.362363] env[62824]: DEBUG oslo_db.sqlalchemy.engines [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=62824) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:395}} [ 495.362988] env[62824]: DEBUG oslo_db.sqlalchemy.engines [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=62824) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:395}} [ 495.368081] env[62824]: ERROR nova.db.main.api [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 495.368081] env[62824]: result = function(*args, **kwargs) [ 495.368081] env[62824]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 495.368081] env[62824]: return func(*args, **kwargs) [ 495.368081] env[62824]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 495.368081] env[62824]: result = fn(*args, **kwargs) [ 495.368081] env[62824]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 495.368081] env[62824]: return f(*args, **kwargs) [ 495.368081] env[62824]: File "/opt/stack/nova/nova/objects/service.py", line 556, in _db_service_get_minimum_version [ 495.368081] env[62824]: return db.service_get_minimum_version(context, binaries) [ 495.368081] env[62824]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 495.368081] env[62824]: _check_db_access() [ 495.368081] env[62824]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 495.368081] env[62824]: stacktrace = ''.join(traceback.format_stack()) [ 495.368081] env[62824]: [ 495.368869] env[62824]: ERROR nova.db.main.api [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 495.368869] env[62824]: result = function(*args, **kwargs) [ 495.368869] env[62824]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 495.368869] env[62824]: return func(*args, **kwargs) [ 495.368869] env[62824]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 495.368869] env[62824]: result = fn(*args, **kwargs) [ 495.368869] env[62824]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 495.368869] env[62824]: return f(*args, **kwargs) [ 495.368869] env[62824]: File "/opt/stack/nova/nova/objects/service.py", line 556, in _db_service_get_minimum_version [ 495.368869] env[62824]: return db.service_get_minimum_version(context, binaries) [ 495.368869] env[62824]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 495.368869] env[62824]: _check_db_access() [ 495.368869] env[62824]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 495.368869] env[62824]: stacktrace = ''.join(traceback.format_stack()) [ 495.368869] env[62824]: [ 495.369271] env[62824]: WARNING nova.objects.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] Failed to get minimum service version for cell 00000000-0000-0000-0000-000000000000 [ 495.369395] env[62824]: WARNING nova.objects.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] Failed to get minimum service version for cell 62bd51a0-02df-41ea-b321-4bd1095178fd [ 495.369817] env[62824]: DEBUG oslo_concurrency.lockutils [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] Acquiring lock "singleton_lock" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 495.369980] env[62824]: DEBUG oslo_concurrency.lockutils [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] Acquired lock "singleton_lock" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 495.370235] env[62824]: DEBUG oslo_concurrency.lockutils [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] Releasing lock "singleton_lock" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 495.370564] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] Full set of CONF: {{(pid=62824) _wait_for_exit_or_signal /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/service.py:363}} [ 495.370708] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] ******************************************************************************** {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2804}} [ 495.370839] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] Configuration options gathered from: {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2805}} [ 495.371025] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] command line args: ['--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-cpu-common.conf', '--config-file', '/etc/nova/nova-cpu-1.conf'] {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2806}} [ 495.371233] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2807}} [ 495.371366] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] ================================================================================ {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2809}} [ 495.371574] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] allow_resize_to_same_host = True {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.371749] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] arq_binding_timeout = 300 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.371884] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] backdoor_port = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.372028] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] backdoor_socket = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.372207] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] block_device_allocate_retries = 60 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.372366] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] block_device_allocate_retries_interval = 3 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.372535] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cert = self.pem {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.372702] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] compute_driver = vmwareapi.VMwareVCDriver {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.372871] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] compute_monitors = [] {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.373050] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] config_dir = [] {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.373230] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] config_drive_format = iso9660 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.373369] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.373535] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] config_source = [] {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.373702] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] console_host = devstack {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.373867] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] control_exchange = nova {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.374033] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cpu_allocation_ratio = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.374236] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] daemon = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.374406] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] debug = True {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.374565] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] default_access_ip_network_name = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.374731] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] default_availability_zone = nova {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.374889] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] default_ephemeral_format = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.375067] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] default_green_pool_size = 1000 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.375362] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.375543] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] default_schedule_zone = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.375706] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] disk_allocation_ratio = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.375867] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] enable_new_services = True {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.376059] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] enabled_apis = ['osapi_compute'] {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.376231] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] enabled_ssl_apis = [] {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.376396] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] flat_injected = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.376557] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] force_config_drive = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.376717] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] force_raw_images = True {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.376890] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] graceful_shutdown_timeout = 5 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.377067] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] heal_instance_info_cache_interval = 60 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.377313] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] host = cpu-1 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.377492] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] initial_cpu_allocation_ratio = 4.0 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.377659] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] initial_disk_allocation_ratio = 1.0 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.377820] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] initial_ram_allocation_ratio = 1.0 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.378048] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] injected_network_template = /opt/stack/nova/nova/virt/interfaces.template {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.378222] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] instance_build_timeout = 0 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.378389] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] instance_delete_interval = 300 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.378558] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] instance_format = [instance: %(uuid)s] {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.378727] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] instance_name_template = instance-%08x {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.378889] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] instance_usage_audit = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.379070] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] instance_usage_audit_period = month {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.379242] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] instance_uuid_format = [instance: %(uuid)s] {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.379412] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] instances_path = /opt/stack/data/nova/instances {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.379579] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] internal_service_availability_zone = internal {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.379737] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] key = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.379900] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] live_migration_retry_count = 30 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.380079] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] log_color = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.380270] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] log_config_append = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.380426] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] log_date_format = %Y-%m-%d %H:%M:%S {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.380587] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] log_dir = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.380747] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] log_file = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.380876] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] log_options = True {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.381053] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] log_rotate_interval = 1 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.381231] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] log_rotate_interval_type = days {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.381404] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] log_rotation_type = none {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.381534] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] logging_context_format_string = %(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.381660] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] logging_debug_format_suffix = {{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}} {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.381829] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] logging_default_format_string = %(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.381995] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] logging_exception_prefix = ERROR %(name)s %(instance)s {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.382142] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.382310] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] long_rpc_timeout = 1800 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.382471] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] max_concurrent_builds = 10 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.382633] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] max_concurrent_live_migrations = 1 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.382792] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] max_concurrent_snapshots = 5 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.382950] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] max_local_block_devices = 3 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.383126] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] max_logfile_count = 30 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.383298] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] max_logfile_size_mb = 200 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.383455] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] maximum_instance_delete_attempts = 5 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.383624] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] metadata_listen = 0.0.0.0 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.383794] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] metadata_listen_port = 8775 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.383961] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] metadata_workers = 2 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.384156] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] migrate_max_retries = -1 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.384334] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] mkisofs_cmd = genisoimage {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.384548] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] my_block_storage_ip = 10.180.1.21 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.384684] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] my_ip = 10.180.1.21 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.384890] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] my_shared_fs_storage_ip = 10.180.1.21 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.385069] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] network_allocate_retries = 0 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.385278] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.385492] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] osapi_compute_listen = 0.0.0.0 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.385665] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] osapi_compute_listen_port = 8774 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.385835] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] osapi_compute_unique_server_name_scope = {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.386014] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] osapi_compute_workers = 2 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.386189] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] password_length = 12 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.386356] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] periodic_enable = True {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.386516] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] periodic_fuzzy_delay = 60 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.386683] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] pointer_model = usbtablet {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.386851] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] preallocate_images = none {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.387020] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] publish_errors = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.387155] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] pybasedir = /opt/stack/nova {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.387317] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] ram_allocation_ratio = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.387478] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] rate_limit_burst = 0 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.387650] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] rate_limit_except_level = CRITICAL {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.387812] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] rate_limit_interval = 0 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.387971] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] reboot_timeout = 0 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.388146] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] reclaim_instance_interval = 0 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.388309] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] record = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.388480] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] reimage_timeout_per_gb = 60 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.388646] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] report_interval = 120 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.388810] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] rescue_timeout = 0 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.388969] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] reserved_host_cpus = 0 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.389143] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] reserved_host_disk_mb = 0 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.389309] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] reserved_host_memory_mb = 512 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.389469] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] reserved_huge_pages = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.389629] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] resize_confirm_window = 0 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.389788] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] resize_fs_using_block_device = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.389945] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] resume_guests_state_on_host_boot = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.390126] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] rootwrap_config = /etc/nova/rootwrap.conf {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.390296] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] rpc_response_timeout = 60 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.390453] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] run_external_periodic_tasks = True {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.390620] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] running_deleted_instance_action = reap {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.390781] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] running_deleted_instance_poll_interval = 1800 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.390939] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] running_deleted_instance_timeout = 0 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.391111] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] scheduler_instance_sync_interval = 120 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.391283] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] service_down_time = 720 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.391453] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] servicegroup_driver = db {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.391608] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] shell_completion = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.391768] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] shelved_offload_time = 0 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.391925] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] shelved_poll_interval = 3600 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.392106] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] shutdown_timeout = 0 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.392269] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] source_is_ipv6 = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.392428] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] ssl_only = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.392684] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] state_path = /opt/stack/data/n-cpu-1 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.392860] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] sync_power_state_interval = 600 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.393036] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] sync_power_state_pool_size = 1000 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.393214] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] syslog_log_facility = LOG_USER {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.393376] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] tempdir = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.393539] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] timeout_nbd = 10 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.393709] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] transport_url = **** {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.393872] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] update_resources_interval = 0 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.394047] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] use_cow_images = True {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.394235] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] use_journal = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.394409] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] use_json = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.394577] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] use_rootwrap_daemon = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.394738] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] use_stderr = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.394898] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] use_syslog = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.395071] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vcpu_pin_set = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.395305] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vif_plugging_is_fatal = True {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.395532] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vif_plugging_timeout = 300 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.395669] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] virt_mkfs = [] {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.395834] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] volume_usage_poll_interval = 0 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.396000] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] watch_log_file = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.396191] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] web = /usr/share/spice-html5 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 495.396380] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] os_brick.lock_path = /opt/stack/data/n-cpu-1 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.396550] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] os_brick.wait_mpath_device_attempts = 4 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.396716] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] os_brick.wait_mpath_device_interval = 1 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.396887] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_concurrency.disable_process_locking = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.397454] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_concurrency.lock_path = /opt/stack/data/n-cpu-1 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.397656] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_messaging_metrics.metrics_buffer_size = 1000 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.397833] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_messaging_metrics.metrics_enabled = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.398026] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_messaging_metrics.metrics_process_name = {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.398248] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.398432] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.398621] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] api.auth_strategy = keystone {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.398792] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] api.compute_link_prefix = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.399036] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.399254] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] api.dhcp_domain = novalocal {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.399472] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] api.enable_instance_password = True {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.399674] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] api.glance_link_prefix = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.399861] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] api.instance_list_cells_batch_fixed_size = 100 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.400052] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] api.instance_list_cells_batch_strategy = distributed {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.400229] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] api.instance_list_per_project_cells = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.400399] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] api.list_records_by_skipping_down_cells = True {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.400565] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] api.local_metadata_per_cell = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.400739] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] api.max_limit = 1000 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.400910] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] api.metadata_cache_expiration = 15 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.401103] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] api.neutron_default_tenant_id = default {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.401284] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] api.response_validation = warn {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.401462] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] api.use_neutron_default_nets = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.401635] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] api.vendordata_dynamic_connect_timeout = 5 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.401800] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] api.vendordata_dynamic_failure_fatal = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.401971] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] api.vendordata_dynamic_read_timeout = 5 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.402194] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] api.vendordata_dynamic_ssl_certfile = {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.402381] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] api.vendordata_dynamic_targets = [] {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.402548] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] api.vendordata_jsonfile_path = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.402733] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] api.vendordata_providers = ['StaticJSON'] {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.402929] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cache.backend = dogpile.cache.memcached {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.403116] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cache.backend_argument = **** {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.403283] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cache.backend_expiration_time = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.403458] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cache.config_prefix = cache.oslo {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.403630] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cache.dead_timeout = 60.0 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.403797] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cache.debug_cache_backend = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.403961] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cache.enable_retry_client = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.404161] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cache.enable_socket_keepalive = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.404341] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cache.enabled = True {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.404512] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cache.enforce_fips_mode = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.404681] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cache.expiration_time = 600 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.404846] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cache.hashclient_retry_attempts = 2 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.405027] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cache.hashclient_retry_delay = 1.0 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.405201] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cache.memcache_dead_retry = 300 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.405391] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cache.memcache_password = **** {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.405565] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cache.memcache_pool_connection_get_timeout = 10 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.405729] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cache.memcache_pool_flush_on_reconnect = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.405893] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cache.memcache_pool_maxsize = 10 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.406096] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cache.memcache_pool_unused_timeout = 60 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.406290] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cache.memcache_sasl_enabled = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.406478] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cache.memcache_servers = ['localhost:11211'] {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.406651] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cache.memcache_socket_timeout = 1.0 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.406814] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cache.memcache_username = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.406985] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cache.proxies = [] {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.407171] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cache.redis_db = 0 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.407340] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cache.redis_password = **** {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.407514] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cache.redis_sentinel_service_name = mymaster {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.407694] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cache.redis_sentinels = ['localhost:26379'] {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.407875] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cache.redis_server = localhost:6379 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.408090] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cache.redis_socket_timeout = 1.0 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.408263] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cache.redis_username = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.408435] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cache.retry_attempts = 2 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.408604] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cache.retry_delay = 0.0 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.408771] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cache.socket_keepalive_count = 1 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.408987] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cache.socket_keepalive_idle = 1 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.409221] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cache.socket_keepalive_interval = 1 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.409399] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cache.tls_allowed_ciphers = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.409607] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cache.tls_cafile = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.409793] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cache.tls_certfile = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.410009] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cache.tls_enabled = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.410221] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cache.tls_keyfile = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.410414] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cinder.auth_section = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.410677] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cinder.auth_type = password {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.410872] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cinder.cafile = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.411077] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cinder.catalog_info = volumev3::publicURL {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.411251] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cinder.certfile = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.411440] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cinder.collect_timing = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.411672] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cinder.cross_az_attach = True {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.411891] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cinder.debug = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.412095] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cinder.endpoint_template = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.412274] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cinder.http_retries = 3 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.412446] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cinder.insecure = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.412609] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cinder.keyfile = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.412784] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cinder.os_region_name = RegionOne {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.412951] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cinder.split_loggers = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.413128] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cinder.timeout = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.413309] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] compute.consecutive_build_service_disable_threshold = 10 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.413475] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] compute.cpu_dedicated_set = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.413637] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] compute.cpu_shared_set = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.413807] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] compute.image_type_exclude_list = [] {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.413971] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] compute.live_migration_wait_for_vif_plug = True {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.414210] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] compute.max_concurrent_disk_ops = 0 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.414396] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] compute.max_disk_devices_to_attach = -1 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.414568] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] compute.packing_host_numa_cells_allocation_strategy = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.414744] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] compute.provider_config_location = /etc/nova/provider_config/ {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.414914] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] compute.resource_provider_association_refresh = 300 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.415156] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] compute.sharing_providers_max_uuids_per_request = 200 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.415372] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] compute.shutdown_retry_interval = 10 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.415569] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.415755] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] conductor.workers = 2 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.415939] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] console.allowed_origins = [] {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.416118] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] console.ssl_ciphers = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.416296] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] console.ssl_minimum_version = default {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.416474] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] consoleauth.enforce_session_timeout = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.416645] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] consoleauth.token_ttl = 600 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.416817] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cyborg.cafile = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.416978] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cyborg.certfile = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.417161] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cyborg.collect_timing = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.417327] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cyborg.connect_retries = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.417487] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cyborg.connect_retry_delay = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.417647] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cyborg.endpoint_override = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.417811] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cyborg.insecure = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.417972] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cyborg.keyfile = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.418148] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cyborg.max_version = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.418312] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cyborg.min_version = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.418469] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cyborg.region_name = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.418628] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cyborg.retriable_status_codes = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.418788] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cyborg.service_name = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.418959] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cyborg.service_type = accelerator {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.419139] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cyborg.split_loggers = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.419302] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cyborg.status_code_retries = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.419461] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cyborg.status_code_retry_delay = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.419623] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cyborg.timeout = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.419806] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cyborg.valid_interfaces = ['internal', 'public'] {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.419970] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] cyborg.version = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.420157] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] database.asyncio_connection = **** {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.420324] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] database.asyncio_slave_connection = **** {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.420497] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] database.backend = sqlalchemy {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.420670] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] database.connection = **** {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.420865] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] database.connection_debug = 0 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.421058] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] database.connection_parameters = {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.422114] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] database.connection_recycle_time = 3600 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.422114] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] database.connection_trace = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.422114] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] database.db_inc_retry_interval = True {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.422114] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] database.db_max_retries = 20 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.422114] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] database.db_max_retry_interval = 10 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.422306] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] database.db_retry_interval = 1 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.422384] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] database.max_overflow = 50 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.422534] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] database.max_pool_size = 5 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.422700] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] database.max_retries = 10 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.422875] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] database.mysql_sql_mode = TRADITIONAL {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.423047] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] database.mysql_wsrep_sync_wait = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.423212] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] database.pool_timeout = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.423378] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] database.retry_interval = 10 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.423538] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] database.slave_connection = **** {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.423721] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] database.sqlite_synchronous = True {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.423904] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] database.use_db_reconnect = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.424088] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] api_database.asyncio_connection = **** {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.424329] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] api_database.asyncio_slave_connection = **** {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.428014] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] api_database.backend = sqlalchemy {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.428014] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] api_database.connection = **** {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.428014] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] api_database.connection_debug = 0 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.428014] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] api_database.connection_parameters = {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.428014] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] api_database.connection_recycle_time = 3600 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.428014] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] api_database.connection_trace = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.428014] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] api_database.db_inc_retry_interval = True {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.428254] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] api_database.db_max_retries = 20 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.428254] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] api_database.db_max_retry_interval = 10 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.428254] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] api_database.db_retry_interval = 1 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.428254] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] api_database.max_overflow = 50 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.428254] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] api_database.max_pool_size = 5 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.428254] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] api_database.max_retries = 10 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.428254] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] api_database.mysql_sql_mode = TRADITIONAL {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.428441] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] api_database.mysql_wsrep_sync_wait = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.428441] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] api_database.pool_timeout = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.428441] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] api_database.retry_interval = 10 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.428441] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] api_database.slave_connection = **** {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.428441] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] api_database.sqlite_synchronous = True {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.428441] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] devices.enabled_mdev_types = [] {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.428441] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] ephemeral_storage_encryption.cipher = aes-xts-plain64 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.428610] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] ephemeral_storage_encryption.default_format = luks {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.428610] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] ephemeral_storage_encryption.enabled = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.428661] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] ephemeral_storage_encryption.key_size = 512 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.428807] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] glance.api_servers = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.428969] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] glance.cafile = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.429142] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] glance.certfile = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.429302] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] glance.collect_timing = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.429457] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] glance.connect_retries = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.429611] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] glance.connect_retry_delay = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.429770] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] glance.debug = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.429932] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] glance.default_trusted_certificate_ids = [] {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.430103] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] glance.enable_certificate_validation = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.430275] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] glance.enable_rbd_download = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.430514] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] glance.endpoint_override = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.430718] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] glance.insecure = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.430884] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] glance.keyfile = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.431055] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] glance.max_version = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.431215] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] glance.min_version = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.431376] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] glance.num_retries = 3 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.431540] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] glance.rbd_ceph_conf = {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.431700] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] glance.rbd_connect_timeout = 5 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.431864] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] glance.rbd_pool = {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.432042] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] glance.rbd_user = {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.432220] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] glance.region_name = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.432415] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] glance.retriable_status_codes = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.432902] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] glance.service_name = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.432902] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] glance.service_type = image {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.433027] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] glance.split_loggers = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.433119] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] glance.status_code_retries = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.433282] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] glance.status_code_retry_delay = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.433444] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] glance.timeout = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.433706] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] glance.valid_interfaces = ['internal', 'public'] {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.433902] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] glance.verify_glance_signatures = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.434085] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] glance.version = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.434291] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] guestfs.debug = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.434475] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] manila.auth_section = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.434641] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] manila.auth_type = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.434804] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] manila.cafile = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.434964] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] manila.certfile = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.435148] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] manila.collect_timing = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.435312] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] manila.connect_retries = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.435474] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] manila.connect_retry_delay = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.435635] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] manila.endpoint_override = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.435811] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] manila.insecure = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.435973] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] manila.keyfile = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.436152] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] manila.max_version = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.436315] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] manila.min_version = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.436477] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] manila.region_name = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.436637] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] manila.retriable_status_codes = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.436797] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] manila.service_name = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.436970] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] manila.service_type = shared-file-system {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.437153] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] manila.share_apply_policy_timeout = 10 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.437323] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] manila.split_loggers = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.437484] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] manila.status_code_retries = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.437644] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] manila.status_code_retry_delay = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.437802] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] manila.timeout = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.437986] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] manila.valid_interfaces = ['internal', 'public'] {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.438165] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] manila.version = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.438339] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] mks.enabled = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.438693] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] mks.mksproxy_base_url = http://127.0.0.1:6090/ {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.438889] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] image_cache.manager_interval = 2400 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.439076] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] image_cache.precache_concurrency = 1 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.439255] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] image_cache.remove_unused_base_images = True {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.439451] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] image_cache.remove_unused_original_minimum_age_seconds = 86400 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.439638] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] image_cache.remove_unused_resized_minimum_age_seconds = 3600 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.439820] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] image_cache.subdirectory_name = _base {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.439999] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] ironic.api_max_retries = 60 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.440184] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] ironic.api_retry_interval = 2 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.440349] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] ironic.auth_section = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.440515] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] ironic.auth_type = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.440683] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] ironic.cafile = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.440840] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] ironic.certfile = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.441012] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] ironic.collect_timing = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.441187] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] ironic.conductor_group = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.441351] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] ironic.connect_retries = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.441511] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] ironic.connect_retry_delay = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.441670] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] ironic.endpoint_override = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.441835] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] ironic.insecure = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.441996] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] ironic.keyfile = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.442189] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] ironic.max_version = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.442366] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] ironic.min_version = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.442536] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] ironic.peer_list = [] {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.442698] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] ironic.region_name = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.442859] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] ironic.retriable_status_codes = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.443042] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] ironic.serial_console_state_timeout = 10 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.443209] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] ironic.service_name = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.443386] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] ironic.service_type = baremetal {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.443550] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] ironic.shard = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.443715] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] ironic.split_loggers = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.443874] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] ironic.status_code_retries = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.444045] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] ironic.status_code_retry_delay = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.444238] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] ironic.timeout = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.444436] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] ironic.valid_interfaces = ['internal', 'public'] {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.444601] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] ironic.version = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.444789] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] key_manager.backend = nova.keymgr.conf_key_mgr.ConfKeyManager {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.444966] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] key_manager.fixed_key = **** {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.445171] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] barbican.auth_endpoint = http://localhost/identity/v3 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.445340] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] barbican.barbican_api_version = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.445502] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] barbican.barbican_endpoint = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.445674] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] barbican.barbican_endpoint_type = public {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.445834] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] barbican.barbican_region_name = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.445996] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] barbican.cafile = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.446173] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] barbican.certfile = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.446340] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] barbican.collect_timing = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.446519] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] barbican.insecure = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.446690] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] barbican.keyfile = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.446858] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] barbican.number_of_retries = 60 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.447033] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] barbican.retry_delay = 1 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.447204] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] barbican.send_service_user_token = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.447370] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] barbican.split_loggers = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.447527] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] barbican.timeout = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.447689] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] barbican.verify_ssl = True {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.447847] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] barbican.verify_ssl_path = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.448023] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] barbican_service_user.auth_section = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.448194] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] barbican_service_user.auth_type = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.448357] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] barbican_service_user.cafile = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.448517] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] barbican_service_user.certfile = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.448681] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] barbican_service_user.collect_timing = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.448844] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] barbican_service_user.insecure = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.449008] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] barbican_service_user.keyfile = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.449183] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] barbican_service_user.split_loggers = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.449343] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] barbican_service_user.timeout = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.449546] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vault.approle_role_id = **** {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.449713] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vault.approle_secret_id = **** {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.449889] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vault.kv_mountpoint = secret {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.450065] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vault.kv_path = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.450247] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vault.kv_version = 2 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.450416] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vault.namespace = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.450579] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vault.root_token_id = **** {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.450741] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vault.ssl_ca_crt_file = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.450909] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vault.timeout = 60.0 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.451090] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vault.use_ssl = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.451268] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vault.vault_url = http://127.0.0.1:8200 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.451442] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] keystone.cafile = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.451603] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] keystone.certfile = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.451770] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] keystone.collect_timing = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.451931] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] keystone.connect_retries = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.452105] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] keystone.connect_retry_delay = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.452298] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] keystone.endpoint_override = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.452470] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] keystone.insecure = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.452632] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] keystone.keyfile = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.452791] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] keystone.max_version = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.452951] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] keystone.min_version = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.453128] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] keystone.region_name = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.453302] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] keystone.retriable_status_codes = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.453477] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] keystone.service_name = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.453652] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] keystone.service_type = identity {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.453817] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] keystone.split_loggers = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.453979] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] keystone.status_code_retries = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.454175] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] keystone.status_code_retry_delay = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.454354] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] keystone.timeout = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.454541] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] keystone.valid_interfaces = ['internal', 'public'] {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.454706] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] keystone.version = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.454900] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.ceph_mount_options = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.455242] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.ceph_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.455430] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.connection_uri = {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.455597] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.cpu_mode = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.455767] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.cpu_model_extra_flags = [] {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.455936] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.cpu_models = [] {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.456123] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.cpu_power_governor_high = performance {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.456298] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.cpu_power_governor_low = powersave {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.456479] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.cpu_power_management = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.456668] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.cpu_power_management_strategy = cpu_state {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.456841] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.device_detach_attempts = 8 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.457013] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.device_detach_timeout = 20 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.457193] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.disk_cachemodes = [] {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.457358] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.disk_prefix = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.457556] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.enabled_perf_events = [] {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.457735] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.file_backed_memory = 0 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.457904] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.gid_maps = [] {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.458078] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.hw_disk_discard = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.458244] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.hw_machine_type = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.458416] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.images_rbd_ceph_conf = {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.458584] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.images_rbd_glance_copy_poll_interval = 15 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.458749] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.images_rbd_glance_copy_timeout = 600 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.458919] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.images_rbd_glance_store_name = {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.459101] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.images_rbd_pool = rbd {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.459277] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.images_type = default {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.459440] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.images_volume_group = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.459603] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.inject_key = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.459767] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.inject_partition = -2 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.459930] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.inject_password = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.460109] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.iscsi_iface = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.460274] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.iser_use_multipath = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.460442] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.live_migration_bandwidth = 0 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.460605] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.live_migration_completion_timeout = 800 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.460770] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.live_migration_downtime = 500 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.460941] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.live_migration_downtime_delay = 75 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.461112] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.live_migration_downtime_steps = 10 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.461272] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.live_migration_inbound_addr = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.461461] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.live_migration_permit_auto_converge = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.461643] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.live_migration_permit_post_copy = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.461807] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.live_migration_scheme = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.461981] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.live_migration_timeout_action = abort {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.462162] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.live_migration_tunnelled = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.462324] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.live_migration_uri = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.462489] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.live_migration_with_native_tls = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.462649] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.max_queues = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.462814] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.mem_stats_period_seconds = 10 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.463185] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.migration_inbound_addr = 10.180.1.21 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.463462] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.nfs_mount_options = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.463912] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.nfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.464287] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.num_aoe_discover_tries = 3 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.464577] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.num_iser_scan_tries = 5 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.464851] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.num_memory_encrypted_guests = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.465159] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.num_nvme_discover_tries = 5 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.465405] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.num_pcie_ports = 0 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.465605] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.num_volume_scan_tries = 5 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.465779] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.pmem_namespaces = [] {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.465946] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.quobyte_client_cfg = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.466277] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.quobyte_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.466457] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.rbd_connect_timeout = 5 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.466626] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.rbd_destroy_volume_retries = 12 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.466791] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.rbd_destroy_volume_retry_interval = 5 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.466951] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.rbd_secret_uuid = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.467129] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.rbd_user = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.467297] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.realtime_scheduler_priority = 1 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.467471] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.remote_filesystem_transport = ssh {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.467632] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.rescue_image_id = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.467792] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.rescue_kernel_id = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.467962] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.rescue_ramdisk_id = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.468152] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.rng_dev_path = /dev/urandom {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.468316] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.rx_queue_size = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.468576] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.smbfs_mount_options = {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.468899] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.smbfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.469096] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.snapshot_compression = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.469292] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.snapshot_image_format = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.469514] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.snapshots_directory = /opt/stack/data/nova/instances/snapshots {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.469684] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.sparse_logical_volumes = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.469848] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.swtpm_enabled = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.470030] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.swtpm_group = tss {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.470206] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.swtpm_user = tss {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.470378] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.sysinfo_serial = unique {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.470536] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.tb_cache_size = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.470704] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.tx_queue_size = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.470872] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.uid_maps = [] {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.471047] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.use_virtio_for_bridges = True {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.471224] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.virt_type = kvm {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.471395] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.volume_clear = zero {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.471561] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.volume_clear_size = 0 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.471725] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.volume_use_multipath = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.471883] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.vzstorage_cache_path = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.472072] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.472269] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.vzstorage_mount_group = qemu {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.472445] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.vzstorage_mount_opts = [] {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.472615] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.vzstorage_mount_perms = 0770 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.472901] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.vzstorage_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.473095] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.vzstorage_mount_user = stack {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.473268] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] libvirt.wait_soft_reboot_seconds = 120 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.473449] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] neutron.auth_section = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.473625] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] neutron.auth_type = password {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.473788] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] neutron.cafile = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.473946] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] neutron.certfile = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.474138] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] neutron.collect_timing = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.474311] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] neutron.connect_retries = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.474495] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] neutron.connect_retry_delay = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.474683] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] neutron.default_floating_pool = public {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.474844] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] neutron.endpoint_override = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.475024] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] neutron.extension_sync_interval = 600 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.475194] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] neutron.http_retries = 3 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.475353] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] neutron.insecure = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.475512] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] neutron.keyfile = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.475669] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] neutron.max_version = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.475844] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] neutron.metadata_proxy_shared_secret = **** {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.476009] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] neutron.min_version = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.476200] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] neutron.ovs_bridge = br-int {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.476362] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] neutron.physnets = [] {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.476533] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] neutron.region_name = RegionOne {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.476692] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] neutron.retriable_status_codes = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.476859] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] neutron.service_metadata_proxy = True {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.477029] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] neutron.service_name = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.477203] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] neutron.service_type = network {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.477368] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] neutron.split_loggers = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.477527] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] neutron.status_code_retries = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.477684] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] neutron.status_code_retry_delay = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.477842] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] neutron.timeout = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.478032] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] neutron.valid_interfaces = ['internal', 'public'] {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.478206] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] neutron.version = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.478377] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] notifications.bdms_in_notifications = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.478554] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] notifications.default_level = INFO {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.478720] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] notifications.include_share_mapping = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.478897] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] notifications.notification_format = unversioned {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.479076] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] notifications.notify_on_state_change = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.479258] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] notifications.versioned_notifications_topics = ['versioned_notifications'] {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.479440] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] pci.alias = [] {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.479607] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] pci.device_spec = [] {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.479771] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] pci.report_in_placement = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.479941] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] placement.auth_section = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.480130] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] placement.auth_type = password {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.480304] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] placement.auth_url = http://10.180.1.21/identity {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.480468] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] placement.cafile = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.480627] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] placement.certfile = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.480788] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] placement.collect_timing = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.480947] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] placement.connect_retries = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.481122] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] placement.connect_retry_delay = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.481285] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] placement.default_domain_id = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.481444] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] placement.default_domain_name = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.481604] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] placement.domain_id = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.481759] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] placement.domain_name = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.481916] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] placement.endpoint_override = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.482089] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] placement.insecure = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.482277] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] placement.keyfile = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.482444] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] placement.max_version = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.482603] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] placement.min_version = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.482774] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] placement.password = **** {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.482934] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] placement.project_domain_id = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.483121] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] placement.project_domain_name = Default {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.483294] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] placement.project_id = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.483472] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] placement.project_name = service {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.483642] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] placement.region_name = RegionOne {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.483803] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] placement.retriable_status_codes = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.483960] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] placement.service_name = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.484159] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] placement.service_type = placement {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.484349] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] placement.split_loggers = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.484521] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] placement.status_code_retries = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.484683] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] placement.status_code_retry_delay = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.484844] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] placement.system_scope = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.485010] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] placement.timeout = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.485177] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] placement.trust_id = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.485335] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] placement.user_domain_id = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.485505] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] placement.user_domain_name = Default {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.485666] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] placement.user_id = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.485842] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] placement.username = nova {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.486035] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] placement.valid_interfaces = ['internal', 'public'] {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.486204] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] placement.version = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.486384] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] quota.cores = 20 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.486610] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] quota.count_usage_from_placement = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.486817] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] quota.driver = nova.quota.DbQuotaDriver {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.487026] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] quota.injected_file_content_bytes = 10240 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.487209] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] quota.injected_file_path_length = 255 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.487380] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] quota.injected_files = 5 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.487550] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] quota.instances = 10 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.487717] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] quota.key_pairs = 100 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.487886] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] quota.metadata_items = 128 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.488066] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] quota.ram = 51200 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.488237] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] quota.recheck_quota = True {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.488409] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] quota.server_group_members = 10 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.488576] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] quota.server_groups = 10 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.488787] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] quota.unified_limits_resource_list = ['servers'] {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.488974] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] quota.unified_limits_resource_strategy = require {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.489161] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] scheduler.discover_hosts_in_cells_interval = -1 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.489328] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] scheduler.enable_isolated_aggregate_filtering = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.489491] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] scheduler.image_metadata_prefilter = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.489651] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] scheduler.limit_tenants_to_placement_aggregate = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.489813] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] scheduler.max_attempts = 3 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.489975] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] scheduler.max_placement_results = 1000 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.490150] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] scheduler.placement_aggregate_required_for_tenants = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.490313] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] scheduler.query_placement_for_image_type_support = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.490473] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] scheduler.query_placement_for_routed_network_aggregates = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.490649] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] scheduler.workers = 2 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.490819] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] filter_scheduler.aggregate_image_properties_isolation_namespace = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.490988] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] filter_scheduler.aggregate_image_properties_isolation_separator = . {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.491185] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.491355] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] filter_scheduler.build_failure_weight_multiplier = 1000000.0 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.491518] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] filter_scheduler.cpu_weight_multiplier = 1.0 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.491681] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.491843] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] filter_scheduler.disk_weight_multiplier = 1.0 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.492042] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', 'SameHostFilter', 'DifferentHostFilter'] {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.492220] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] filter_scheduler.host_subset_size = 1 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.492388] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] filter_scheduler.hypervisor_version_weight_multiplier = 1.0 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.492565] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] filter_scheduler.image_properties_default_architecture = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.492744] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] filter_scheduler.io_ops_weight_multiplier = -1.0 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.492911] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] filter_scheduler.isolated_hosts = [] {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.493087] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] filter_scheduler.isolated_images = [] {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.493271] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] filter_scheduler.max_instances_per_host = 50 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.493436] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] filter_scheduler.max_io_ops_per_host = 8 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.493597] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] filter_scheduler.num_instances_weight_multiplier = 0.0 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.493759] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] filter_scheduler.pci_in_placement = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.493918] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] filter_scheduler.pci_weight_multiplier = 1.0 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.494092] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] filter_scheduler.ram_weight_multiplier = 1.0 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.494292] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.494465] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] filter_scheduler.shuffle_best_same_weighed_hosts = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.494630] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] filter_scheduler.soft_affinity_weight_multiplier = 1.0 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.494793] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.494955] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] filter_scheduler.track_instance_changes = True {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.495149] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.495322] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] metrics.required = True {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.495486] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] metrics.weight_multiplier = 1.0 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.495649] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] metrics.weight_of_unavailable = -10000.0 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.495813] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] metrics.weight_setting = [] {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.496194] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] serial_console.base_url = ws://127.0.0.1:6083/ {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.496390] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] serial_console.enabled = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.496598] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] serial_console.port_range = 10000:20000 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.496781] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] serial_console.proxyclient_address = 127.0.0.1 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.496955] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] serial_console.serialproxy_host = 0.0.0.0 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.497145] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] serial_console.serialproxy_port = 6083 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.497319] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] service_user.auth_section = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.497494] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] service_user.auth_type = password {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.497655] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] service_user.cafile = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.497813] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] service_user.certfile = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.497974] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] service_user.collect_timing = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.498149] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] service_user.insecure = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.498310] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] service_user.keyfile = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.498480] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] service_user.send_service_user_token = True {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.498642] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] service_user.split_loggers = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.498800] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] service_user.timeout = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.498968] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] spice.agent_enabled = True {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.499142] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] spice.enabled = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.499458] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.499662] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] spice.html5proxy_host = 0.0.0.0 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.499835] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] spice.html5proxy_port = 6082 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.499997] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] spice.image_compression = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.500175] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] spice.jpeg_compression = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.500335] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] spice.playback_compression = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.500497] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] spice.require_secure = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.500665] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] spice.server_listen = 127.0.0.1 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.500835] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] spice.server_proxyclient_address = 127.0.0.1 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.500992] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] spice.streaming_mode = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.501166] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] spice.zlib_compression = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.501333] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] upgrade_levels.baseapi = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.501504] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] upgrade_levels.compute = auto {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.501665] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] upgrade_levels.conductor = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.501825] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] upgrade_levels.scheduler = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.501991] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vendordata_dynamic_auth.auth_section = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.502168] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vendordata_dynamic_auth.auth_type = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.502326] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vendordata_dynamic_auth.cafile = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.502484] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vendordata_dynamic_auth.certfile = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.502644] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vendordata_dynamic_auth.collect_timing = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.502804] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vendordata_dynamic_auth.insecure = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.502964] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vendordata_dynamic_auth.keyfile = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.503138] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vendordata_dynamic_auth.split_loggers = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.503301] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vendordata_dynamic_auth.timeout = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.503474] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vmware.api_retry_count = 10 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.503634] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vmware.ca_file = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.503806] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vmware.cache_prefix = devstack-image-cache {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.503972] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vmware.cluster_name = testcl1 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.504173] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vmware.connection_pool_size = 10 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.504352] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vmware.console_delay_seconds = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.504527] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vmware.datastore_regex = ^datastore.* {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.504768] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vmware.host_ip = vc1.osci.c.eu-de-1.cloud.sap {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.504949] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vmware.host_password = **** {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.505135] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vmware.host_port = 443 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.505330] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vmware.host_username = administrator@vsphere.local {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.505507] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vmware.insecure = True {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.505671] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vmware.integration_bridge = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.505835] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vmware.maximum_objects = 100 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.505994] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vmware.pbm_default_policy = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.506172] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vmware.pbm_enabled = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.506330] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vmware.pbm_wsdl_location = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.506518] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vmware.serial_log_dir = /opt/vmware/vspc {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.506693] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vmware.serial_port_proxy_uri = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.506854] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vmware.serial_port_service_uri = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.507030] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vmware.task_poll_interval = 0.5 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.507207] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vmware.use_linked_clone = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.507384] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vmware.vnc_keymap = en-us {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.507555] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vmware.vnc_port = 5900 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.507721] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vmware.vnc_port_total = 10000 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.507906] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vnc.auth_schemes = ['none'] {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.508095] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vnc.enabled = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.508399] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vnc.novncproxy_base_url = http://127.0.0.1:6080/vnc_auto.html {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.508588] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vnc.novncproxy_host = 0.0.0.0 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.508762] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vnc.novncproxy_port = 6080 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.508950] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vnc.server_listen = 127.0.0.1 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.509148] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vnc.server_proxyclient_address = 127.0.0.1 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.509318] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vnc.vencrypt_ca_certs = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.509478] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vnc.vencrypt_client_cert = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.509638] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vnc.vencrypt_client_key = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.509811] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] workarounds.disable_compute_service_check_for_ffu = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.509972] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] workarounds.disable_deep_image_inspection = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.510148] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] workarounds.disable_fallback_pcpu_query = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.510315] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] workarounds.disable_group_policy_check_upcall = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.510475] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] workarounds.disable_libvirt_livesnapshot = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.510637] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] workarounds.disable_rootwrap = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.510798] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] workarounds.enable_numa_live_migration = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.510958] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] workarounds.enable_qemu_monitor_announce_self = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.511132] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.511296] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] workarounds.handle_virt_lifecycle_events = True {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.511455] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] workarounds.libvirt_disable_apic = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.511615] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] workarounds.never_download_image_if_on_rbd = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.511778] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] workarounds.qemu_monitor_announce_self_count = 3 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.511939] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] workarounds.qemu_monitor_announce_self_interval = 1 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.512122] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] workarounds.reserve_disk_resource_for_image_cache = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.512285] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] workarounds.skip_cpu_compare_at_startup = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.512449] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] workarounds.skip_cpu_compare_on_dest = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.512610] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] workarounds.skip_hypervisor_version_check_on_lm = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.512769] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] workarounds.skip_reserve_in_use_ironic_nodes = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.512931] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] workarounds.unified_limits_count_pcpu_as_vcpu = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.513107] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.513296] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] wsgi.api_paste_config = /etc/nova/api-paste.ini {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.513468] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] wsgi.client_socket_timeout = 900 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.513634] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] wsgi.default_pool_size = 1000 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.513801] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] wsgi.keep_alive = True {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.513967] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] wsgi.max_header_line = 16384 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.514175] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] wsgi.secure_proxy_ssl_header = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.514346] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] wsgi.ssl_ca_file = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.514511] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] wsgi.ssl_cert_file = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.514671] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] wsgi.ssl_key_file = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.514839] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] wsgi.tcp_keepidle = 600 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.515048] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] wsgi.wsgi_log_format = %(client_ip)s "%(request_line)s" status: %(status_code)s len: %(body_length)s time: %(wall_seconds).7f {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.515237] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] zvm.ca_file = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.515421] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] zvm.cloud_connector_url = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.515726] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] zvm.image_tmp_path = /opt/stack/data/n-cpu-1/images {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.515910] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] zvm.reachable_timeout = 300 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.516120] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_versionedobjects.fatal_exception_format_errors = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.516359] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.516551] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] profiler.connection_string = messaging:// {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.516722] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] profiler.enabled = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.516893] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] profiler.es_doc_type = notification {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.517081] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] profiler.es_scroll_size = 10000 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.517257] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] profiler.es_scroll_time = 2m {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.517424] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] profiler.filter_error_trace = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.517594] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] profiler.hmac_keys = **** {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.517761] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] profiler.sentinel_service_name = mymaster {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.517927] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] profiler.socket_timeout = 0.1 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.518107] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] profiler.trace_requests = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.518271] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] profiler.trace_sqlalchemy = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.518450] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] profiler_jaeger.process_tags = {} {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.518611] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] profiler_jaeger.service_name_prefix = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.518773] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] profiler_otlp.service_name_prefix = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.518936] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] remote_debug.host = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.519112] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] remote_debug.port = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.519297] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_messaging_rabbit.amqp_auto_delete = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.519456] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_messaging_rabbit.amqp_durable_queues = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.519617] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_messaging_rabbit.conn_pool_min_size = 2 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.519778] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_messaging_rabbit.conn_pool_ttl = 1200 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.519935] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_messaging_rabbit.direct_mandatory_flag = True {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.520108] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_messaging_rabbit.enable_cancel_on_failover = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.520271] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_messaging_rabbit.heartbeat_in_pthread = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.520431] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_messaging_rabbit.heartbeat_rate = 3 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.520591] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.520760] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_messaging_rabbit.hostname = devstack {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.520919] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_messaging_rabbit.kombu_compression = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.521101] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_messaging_rabbit.kombu_failover_strategy = round-robin {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.521271] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.521437] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.521602] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_messaging_rabbit.processname = nova-compute {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.521763] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_messaging_rabbit.rabbit_ha_queues = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.521923] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_messaging_rabbit.rabbit_interval_max = 30 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.522117] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.522272] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.522436] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.522610] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.522778] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.522943] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_messaging_rabbit.rabbit_quorum_queue = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.523119] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_messaging_rabbit.rabbit_retry_backoff = 2 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.523283] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_messaging_rabbit.rabbit_retry_interval = 1 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.523445] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_messaging_rabbit.rabbit_stream_fanout = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.523605] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.523765] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_messaging_rabbit.rabbit_transient_quorum_queue = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.523927] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_messaging_rabbit.rpc_conn_pool_size = 30 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.524102] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_messaging_rabbit.ssl = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.524306] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_messaging_rabbit.ssl_ca_file = {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.524485] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_messaging_rabbit.ssl_cert_file = {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.524649] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_messaging_rabbit.ssl_enforce_fips_mode = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.524816] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_messaging_rabbit.ssl_key_file = {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.524984] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_messaging_rabbit.ssl_version = {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.525168] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_messaging_rabbit.use_queue_manager = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.525391] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_messaging_notifications.driver = ['messagingv2'] {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.525567] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_messaging_notifications.retry = -1 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.525747] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_messaging_notifications.topics = ['notifications'] {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.525919] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_messaging_notifications.transport_url = **** {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.526107] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_limit.auth_section = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.526273] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_limit.auth_type = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.526433] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_limit.cafile = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.526620] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_limit.certfile = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.526785] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_limit.collect_timing = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.526948] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_limit.connect_retries = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.527121] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_limit.connect_retry_delay = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.527283] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_limit.endpoint_id = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.527457] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_limit.endpoint_interface = publicURL {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.527618] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_limit.endpoint_override = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.527779] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_limit.endpoint_region_name = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.527935] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_limit.endpoint_service_name = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.528109] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_limit.endpoint_service_type = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.528274] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_limit.insecure = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.528433] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_limit.keyfile = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.528591] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_limit.max_version = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.528747] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_limit.min_version = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.528903] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_limit.region_name = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.529074] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_limit.retriable_status_codes = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.529237] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_limit.service_name = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.529396] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_limit.service_type = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.529559] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_limit.split_loggers = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.529715] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_limit.status_code_retries = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.529872] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_limit.status_code_retry_delay = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.530040] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_limit.timeout = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.530203] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_limit.valid_interfaces = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.530361] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_limit.version = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.530527] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_reports.file_event_handler = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.530691] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_reports.file_event_handler_interval = 1 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.530853] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] oslo_reports.log_dir = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.531035] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vif_plug_linux_bridge_privileged.capabilities = [12] {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.531200] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vif_plug_linux_bridge_privileged.group = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.531364] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vif_plug_linux_bridge_privileged.helper_command = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.531528] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.531694] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vif_plug_linux_bridge_privileged.thread_pool_size = 8 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.531856] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vif_plug_linux_bridge_privileged.user = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.532034] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vif_plug_ovs_privileged.capabilities = [12, 1] {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.532201] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vif_plug_ovs_privileged.group = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.532363] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vif_plug_ovs_privileged.helper_command = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.532530] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.532695] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vif_plug_ovs_privileged.thread_pool_size = 8 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.532856] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] vif_plug_ovs_privileged.user = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.533040] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] os_vif_linux_bridge.flat_interface = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.533227] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] os_vif_linux_bridge.forward_bridge_interface = ['all'] {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.533405] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] os_vif_linux_bridge.iptables_bottom_regex = {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.533579] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] os_vif_linux_bridge.iptables_drop_action = DROP {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.533753] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] os_vif_linux_bridge.iptables_top_regex = {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.533929] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] os_vif_linux_bridge.network_device_mtu = 1500 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.534111] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] os_vif_linux_bridge.use_ipv6 = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.534312] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] os_vif_linux_bridge.vlan_interface = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.534502] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] os_vif_ovs.default_qos_type = linux-noop {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.534678] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] os_vif_ovs.isolate_vif = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.534858] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] os_vif_ovs.network_device_mtu = 1500 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.535042] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] os_vif_ovs.ovs_vsctl_timeout = 120 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.535225] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.535429] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] os_vif_ovs.ovsdb_interface = native {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.535602] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] os_vif_ovs.per_port_bridge = False {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.535774] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] privsep_osbrick.capabilities = [21] {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.535937] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] privsep_osbrick.group = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.536114] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] privsep_osbrick.helper_command = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.536284] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] privsep_osbrick.logger_name = os_brick.privileged {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.536459] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] privsep_osbrick.thread_pool_size = 8 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.536642] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] privsep_osbrick.user = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.536822] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.536983] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] nova_sys_admin.group = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.537159] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] nova_sys_admin.helper_command = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.537331] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] nova_sys_admin.logger_name = oslo_privsep.daemon {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.537495] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] nova_sys_admin.thread_pool_size = 8 {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.537657] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] nova_sys_admin.user = None {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 495.537787] env[62824]: DEBUG oslo_service.service [None req-06ec53d8-7478-454c-a17a-7309bdc364bd None None] ******************************************************************************** {{(pid=62824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2828}} [ 495.538231] env[62824]: INFO nova.service [-] Starting compute node (version 30.1.0) [ 496.042345] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-374ec9cd-0c34-444a-a880-bc2465595336 None None] Getting list of instances from cluster (obj){ [ 496.042345] env[62824]: value = "domain-c8" [ 496.042345] env[62824]: _type = "ClusterComputeResource" [ 496.042345] env[62824]: } {{(pid=62824) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 496.043500] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-089da2c8-7941-434d-aba8-4727f36ff75c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 496.052904] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-374ec9cd-0c34-444a-a880-bc2465595336 None None] Got total of 0 instances {{(pid=62824) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 496.053496] env[62824]: WARNING nova.virt.vmwareapi.driver [None req-374ec9cd-0c34-444a-a880-bc2465595336 None None] The vmwareapi driver is not tested by the OpenStack project nor does it have clear maintainer(s) and thus its quality can not be ensured. It should be considered experimental and may be removed in a future release. If you are using the driver in production please let us know via the openstack-discuss mailing list. [ 496.053975] env[62824]: INFO nova.virt.node [None req-374ec9cd-0c34-444a-a880-bc2465595336 None None] Generated node identity bf95157b-a274-42de-9ccf-9851128a44a1 [ 496.054254] env[62824]: INFO nova.virt.node [None req-374ec9cd-0c34-444a-a880-bc2465595336 None None] Wrote node identity bf95157b-a274-42de-9ccf-9851128a44a1 to /opt/stack/data/n-cpu-1/compute_id [ 496.557688] env[62824]: WARNING nova.compute.manager [None req-374ec9cd-0c34-444a-a880-bc2465595336 None None] Compute nodes ['bf95157b-a274-42de-9ccf-9851128a44a1'] for host cpu-1 were not found in the database. If this is the first time this service is starting on this host, then you can ignore this warning. [ 497.564444] env[62824]: INFO nova.compute.manager [None req-374ec9cd-0c34-444a-a880-bc2465595336 None None] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host [ 498.569784] env[62824]: WARNING nova.compute.manager [None req-374ec9cd-0c34-444a-a880-bc2465595336 None None] No compute node record found for host cpu-1. If this is the first time this service is starting on this host, then you can ignore this warning.: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host cpu-1 could not be found. [ 498.570179] env[62824]: DEBUG oslo_concurrency.lockutils [None req-374ec9cd-0c34-444a-a880-bc2465595336 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 498.570873] env[62824]: DEBUG oslo_concurrency.lockutils [None req-374ec9cd-0c34-444a-a880-bc2465595336 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 498.570873] env[62824]: DEBUG oslo_concurrency.lockutils [None req-374ec9cd-0c34-444a-a880-bc2465595336 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 498.571058] env[62824]: DEBUG nova.compute.resource_tracker [None req-374ec9cd-0c34-444a-a880-bc2465595336 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62824) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 498.571944] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8567ba1-10ab-426c-ac92-0c9e32ba7fe7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 498.580499] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd20aa6c-5398-4bd4-97b7-d1ddc42e245d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 498.594683] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ecfd0d3-01f4-4999-8770-0dd1c8682f13 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 498.601015] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b64453d-6733-494d-a1dc-a8ae28df5695 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 498.630968] env[62824]: DEBUG nova.compute.resource_tracker [None req-374ec9cd-0c34-444a-a880-bc2465595336 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181299MB free_disk=175GB free_vcpus=48 pci_devices=None {{(pid=62824) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 498.631143] env[62824]: DEBUG oslo_concurrency.lockutils [None req-374ec9cd-0c34-444a-a880-bc2465595336 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 498.631314] env[62824]: DEBUG oslo_concurrency.lockutils [None req-374ec9cd-0c34-444a-a880-bc2465595336 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 499.134145] env[62824]: WARNING nova.compute.resource_tracker [None req-374ec9cd-0c34-444a-a880-bc2465595336 None None] No compute node record for cpu-1:bf95157b-a274-42de-9ccf-9851128a44a1: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host bf95157b-a274-42de-9ccf-9851128a44a1 could not be found. [ 499.638172] env[62824]: INFO nova.compute.resource_tracker [None req-374ec9cd-0c34-444a-a880-bc2465595336 None None] Compute node record created for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 with uuid: bf95157b-a274-42de-9ccf-9851128a44a1 [ 501.146073] env[62824]: DEBUG nova.compute.resource_tracker [None req-374ec9cd-0c34-444a-a880-bc2465595336 None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 501.146438] env[62824]: DEBUG nova.compute.resource_tracker [None req-374ec9cd-0c34-444a-a880-bc2465595336 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 501.298976] env[62824]: INFO nova.scheduler.client.report [None req-374ec9cd-0c34-444a-a880-bc2465595336 None None] [req-6cdf40f9-5ca9-4d39-a2b4-2b892fa5559e] Created resource provider record via placement API for resource provider with UUID bf95157b-a274-42de-9ccf-9851128a44a1 and name domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28. [ 501.316202] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef3468c4-5eed-4aa4-980d-4d803d0faad6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 501.324711] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10cbebda-0f70-4b35-9803-9f83c0df9c5b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 501.354908] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d692fdb9-8cf0-449e-8b62-c42981215150 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 501.363260] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c81027f-1779-45a2-89d1-e616802402ed {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 501.377158] env[62824]: DEBUG nova.compute.provider_tree [None req-374ec9cd-0c34-444a-a880-bc2465595336 None None] Updating inventory in ProviderTree for provider bf95157b-a274-42de-9ccf-9851128a44a1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 501.918432] env[62824]: DEBUG nova.scheduler.client.report [None req-374ec9cd-0c34-444a-a880-bc2465595336 None None] Updated inventory for provider bf95157b-a274-42de-9ccf-9851128a44a1 with generation 0 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 501.918743] env[62824]: DEBUG nova.compute.provider_tree [None req-374ec9cd-0c34-444a-a880-bc2465595336 None None] Updating resource provider bf95157b-a274-42de-9ccf-9851128a44a1 generation from 0 to 1 during operation: update_inventory {{(pid=62824) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 501.919028] env[62824]: DEBUG nova.compute.provider_tree [None req-374ec9cd-0c34-444a-a880-bc2465595336 None None] Updating inventory in ProviderTree for provider bf95157b-a274-42de-9ccf-9851128a44a1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 501.974897] env[62824]: DEBUG nova.compute.provider_tree [None req-374ec9cd-0c34-444a-a880-bc2465595336 None None] Updating resource provider bf95157b-a274-42de-9ccf-9851128a44a1 generation from 1 to 2 during operation: update_traits {{(pid=62824) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 502.480058] env[62824]: DEBUG nova.compute.resource_tracker [None req-374ec9cd-0c34-444a-a880-bc2465595336 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62824) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 502.480527] env[62824]: DEBUG oslo_concurrency.lockutils [None req-374ec9cd-0c34-444a-a880-bc2465595336 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.849s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 502.480527] env[62824]: DEBUG nova.service [None req-374ec9cd-0c34-444a-a880-bc2465595336 None None] Creating RPC server for service compute {{(pid=62824) start /opt/stack/nova/nova/service.py:186}} [ 502.502859] env[62824]: DEBUG nova.service [None req-374ec9cd-0c34-444a-a880-bc2465595336 None None] Join ServiceGroup membership for this service compute {{(pid=62824) start /opt/stack/nova/nova/service.py:203}} [ 502.503103] env[62824]: DEBUG nova.servicegroup.drivers.db [None req-374ec9cd-0c34-444a-a880-bc2465595336 None None] DB_Driver: join new ServiceGroup member cpu-1 to the compute group, service = {{(pid=62824) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 550.507796] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._sync_power_states {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 551.010834] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-41517485-76de-4e8f-a181-a012faec848d None None] Getting list of instances from cluster (obj){ [ 551.010834] env[62824]: value = "domain-c8" [ 551.010834] env[62824]: _type = "ClusterComputeResource" [ 551.010834] env[62824]: } {{(pid=62824) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 551.012059] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-083a6b84-af5f-4453-bfb9-0ba2b086fc27 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.021052] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-41517485-76de-4e8f-a181-a012faec848d None None] Got total of 0 instances {{(pid=62824) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 551.021419] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 551.021760] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62824) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11064}} [ 551.021989] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager.update_available_resource {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 551.525019] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 551.525435] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 551.525435] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 551.525619] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62824) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 551.526543] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f46ba87-b8cc-4f79-b558-c83ca1c6ba12 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.534955] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccb04c89-b27b-4a9a-949a-9c2bed0bf797 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.548798] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b678568-ee41-40dd-99e0-ae7846711dee {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.554751] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c283c08c-f4dc-43c2-a5cd-351e67ef0c72 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.582879] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181289MB free_disk=175GB free_vcpus=48 pci_devices=None {{(pid=62824) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 551.583039] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 551.583203] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 552.600550] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 552.600818] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 552.615558] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-154c01c8-dea1-40b3-bdcc-cd14d3ebc989 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.623849] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d99ad0d-9d3e-455f-9176-8da8b4ebb24a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.653716] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-955b6062-0ee4-45a0-b3b6-fa746dc92153 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.660809] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd21822f-1fa8-482f-9a99-07503e5d3fe8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.673470] env[62824]: DEBUG nova.compute.provider_tree [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 553.176857] env[62824]: DEBUG nova.scheduler.client.report [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 553.682447] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62824) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 553.682868] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.099s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 553.682868] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._cleanup_running_deleted_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 553.683242] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-41517485-76de-4e8f-a181-a012faec848d None None] Getting list of instances from cluster (obj){ [ 553.683242] env[62824]: value = "domain-c8" [ 553.683242] env[62824]: _type = "ClusterComputeResource" [ 553.683242] env[62824]: } {{(pid=62824) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 553.684263] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9965cff7-1e8d-4efe-9a94-2d382e492f8d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.692723] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-41517485-76de-4e8f-a181-a012faec848d None None] Got total of 0 instances {{(pid=62824) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 553.953252] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 553.953479] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 553.953624] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Starting heal instance info cache {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10445}} [ 553.953742] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Rebuilding the list of instances to heal {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10449}} [ 554.456576] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Didn't find any instances for network info cache update. {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10531}} [ 554.456781] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 554.457030] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 554.457223] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 554.457422] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 554.457599] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 610.773565] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 610.773891] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 611.278636] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 611.278825] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 611.282024] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 611.282024] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62824) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11064}} [ 611.282024] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager.update_available_resource {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 611.786594] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 611.787078] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 611.787078] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 611.787209] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62824) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 611.788222] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b464e3a-1593-43ff-b415-9cfb977b0838 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.796319] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b766dee4-c8fa-45ed-99d9-c6a3d4628b21 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.810068] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c254afd0-e16b-47e0-80eb-b81584d4d4b8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.816271] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd634f80-d33f-4863-bbfe-f78a3e02b143 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.845975] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181283MB free_disk=175GB free_vcpus=48 pci_devices=None {{(pid=62824) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 611.846123] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 611.846319] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 612.866283] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 612.866520] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 612.879185] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52fdd67f-3878-4c4a-8a99-91c7467f747f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.886576] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0485d28b-77e0-4415-b098-c3546e447a8c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.916016] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6ff0140-2f01-4b44-a966-7c470c01738b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.922544] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e8537ee-e38b-4c3b-aa42-ed2bcd2c1c2f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.935178] env[62824]: DEBUG nova.compute.provider_tree [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 613.438698] env[62824]: DEBUG nova.scheduler.client.report [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 613.439954] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62824) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 613.440145] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 1.594s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 613.934808] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 613.935185] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Starting heal instance info cache {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10445}} [ 613.935185] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Rebuilding the list of instances to heal {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10449}} [ 614.438810] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Didn't find any instances for network info cache update. {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10531}} [ 614.439059] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 614.439228] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 614.439372] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 670.773137] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 670.773556] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 670.773556] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager.update_available_resource {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 671.276384] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 671.276702] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 671.276877] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 671.277041] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62824) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 671.278025] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-418ba5e0-ed65-4465-8849-84eef4065a07 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.287277] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a24147a9-c528-482e-b8b1-43e516464500 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.301444] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36c26bb3-96b2-4819-8f99-a55be56943c6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.307475] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14521a52-9e94-4260-8f12-d6bed9825723 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.335307] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181285MB free_disk=175GB free_vcpus=48 pci_devices=None {{(pid=62824) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 671.335457] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 671.335634] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 672.353266] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 672.353542] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 672.365726] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ffe5416-fee1-4ca5-83ef-8fd9c861b3f8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.373286] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-163e588c-e396-479d-90e0-3a83ef6b188c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.402417] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c85ab1d4-feae-465d-aac4-a37bc4280aab {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.409428] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98f308a2-ba98-48c0-b3ce-e4ad2a8f1cdd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.422650] env[62824]: DEBUG nova.compute.provider_tree [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 672.925773] env[62824]: DEBUG nova.scheduler.client.report [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 672.927079] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62824) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 672.927258] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 1.592s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 673.927317] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 673.927707] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 673.927707] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 673.927860] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 673.928027] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 673.928206] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62824) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11064}} [ 674.774370] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 674.774557] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Starting heal instance info cache {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10445}} [ 674.774675] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Rebuilding the list of instances to heal {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10449}} [ 675.277653] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Didn't find any instances for network info cache update. {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10531}} [ 731.272466] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 731.772776] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 731.773060] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager.update_available_resource {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 732.276080] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 732.276441] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 732.276488] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 732.276623] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62824) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 732.277585] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4eb7a2e1-4237-4930-8813-9bc1dadf583e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.285985] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09eacb02-d8b4-41f1-a43b-4c822c8068f7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.301862] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f0da065-63b4-4aad-b1be-ebdb701f3ebe {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.308163] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7087cde-6b18-4328-b52f-0de32d029c6d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.336525] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181286MB free_disk=175GB free_vcpus=48 pci_devices=None {{(pid=62824) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 732.336682] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 732.336858] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 733.355407] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 733.355644] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 733.370355] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fceb0c42-2338-4d73-ab8e-25fa16abf9ed {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.377865] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46757aa4-989a-4150-a9cf-a1a5e2058d13 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.408283] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a4de52c-52ff-4eb9-95a0-337694d1a54a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.415214] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84c4b936-6b9a-44c4-b646-8689fed54f37 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.428048] env[62824]: DEBUG nova.compute.provider_tree [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 733.931504] env[62824]: DEBUG nova.scheduler.client.report [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 733.932798] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62824) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 733.933013] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 1.596s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 734.928763] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 735.434147] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 735.434393] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 735.434599] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 735.434762] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62824) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11064}} [ 735.773363] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 735.773546] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 736.775523] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 736.775523] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Starting heal instance info cache {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10445}} [ 736.775523] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Rebuilding the list of instances to heal {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10449}} [ 737.277245] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Didn't find any instances for network info cache update. {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10531}} [ 790.773309] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 790.773745] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Cleaning up deleted instances {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11745}} [ 791.276837] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] There are 0 instances to clean {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11754}} [ 791.277091] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 791.277232] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Cleaning up deleted instances with incomplete migration {{(pid=62824) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11783}} [ 791.779740] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 793.281974] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 793.282357] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 793.282399] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 793.282542] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager.update_available_resource {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 793.785693] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 793.785947] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 793.786127] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 793.786281] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62824) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 793.787224] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44b2942f-612d-4961-82a8-a9a7da521693 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.795527] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f326130d-d667-4b4e-9aa2-f59168812e32 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.810194] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf38f930-3034-417e-9618-7b1b4c147867 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.816863] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9d436f3-600a-47fa-9815-cb5c7a35ed2f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.845479] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181174MB free_disk=175GB free_vcpus=48 pci_devices=None {{(pid=62824) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 793.845614] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 793.845773] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 794.865317] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 794.865587] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 794.878646] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19d4ac1e-67ee-4181-95e1-01283c3dbf72 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.886264] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f5b4f0b-3fe6-4f3b-9b87-cace23356590 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.915867] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-367f799a-3bf9-40cf-abf5-7eb977562cec {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.922672] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a081eb48-bde3-4e72-8421-c84848dac997 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.935527] env[62824]: DEBUG nova.compute.provider_tree [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 795.438526] env[62824]: DEBUG nova.scheduler.client.report [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 795.439792] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62824) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 795.439971] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 1.594s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 796.930109] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 796.930418] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 796.930724] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 796.930724] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62824) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11064}} [ 797.773433] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 797.773675] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Starting heal instance info cache {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10445}} [ 797.773755] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Rebuilding the list of instances to heal {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10449}} [ 798.276694] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Didn't find any instances for network info cache update. {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10531}} [ 798.277116] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 852.773886] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 852.774308] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 853.769185] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 853.772809] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager.update_available_resource {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 854.276129] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.276582] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.276582] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.276686] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62824) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 854.277602] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f70a1db-1414-46af-bca9-5cc32743be98 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.285991] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af660ca7-77c4-4baa-9233-8ca164f88f64 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.299945] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fc445c2-8e76-414b-93ce-51c3b6236cd7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.306266] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33a9be2c-bd88-4ab2-84a5-0152bb7eb545 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.335589] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181299MB free_disk=175GB free_vcpus=48 pci_devices=None {{(pid=62824) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 854.335722] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.335908] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 855.368803] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 855.369064] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 855.385427] env[62824]: DEBUG nova.scheduler.client.report [None req-41517485-76de-4e8f-a181-a012faec848d None None] Refreshing inventories for resource provider bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 855.397722] env[62824]: DEBUG nova.scheduler.client.report [None req-41517485-76de-4e8f-a181-a012faec848d None None] Updating ProviderTree inventory for provider bf95157b-a274-42de-9ccf-9851128a44a1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 855.397894] env[62824]: DEBUG nova.compute.provider_tree [None req-41517485-76de-4e8f-a181-a012faec848d None None] Updating inventory in ProviderTree for provider bf95157b-a274-42de-9ccf-9851128a44a1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 855.409908] env[62824]: DEBUG nova.scheduler.client.report [None req-41517485-76de-4e8f-a181-a012faec848d None None] Refreshing aggregate associations for resource provider bf95157b-a274-42de-9ccf-9851128a44a1, aggregates: None {{(pid=62824) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 855.425975] env[62824]: DEBUG nova.scheduler.client.report [None req-41517485-76de-4e8f-a181-a012faec848d None None] Refreshing trait associations for resource provider bf95157b-a274-42de-9ccf-9851128a44a1, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_ISO {{(pid=62824) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 855.437256] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1382e7c3-0c3e-459d-9290-223fd60b59f8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.444609] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dccb5e25-3011-474e-8d4e-d581be55e3e3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.473656] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc64ae22-ab1b-4052-a4ac-2fc4ea24563a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.480367] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b5b668f-3bf5-45eb-97a5-1353f1c9457b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.493883] env[62824]: DEBUG nova.compute.provider_tree [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 855.997374] env[62824]: DEBUG nova.scheduler.client.report [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 855.998701] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62824) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 855.998881] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 1.663s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 857.994985] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 858.502107] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 858.502107] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 858.502107] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 858.502107] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 858.502107] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62824) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11064}} [ 859.774261] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 859.774665] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Starting heal instance info cache {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10445}} [ 859.774665] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Rebuilding the list of instances to heal {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10449}} [ 860.277342] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Didn't find any instances for network info cache update. {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10531}} [ 912.773433] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 913.774202] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 913.774624] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager.update_available_resource {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 914.277532] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.277778] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.277944] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.278115] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62824) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 914.279152] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-855598a7-ae61-44ee-85a2-66b7605a9737 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.287556] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afc8c6e2-c640-420b-8691-6e5434969adc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.301265] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b249b72a-695e-44d2-ab6c-c498bf28268c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.307435] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb7b7e77-074d-4004-b145-5c3e19c8987e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.337850] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181293MB free_disk=175GB free_vcpus=48 pci_devices=None {{(pid=62824) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 914.337996] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.338185] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.356641] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 915.356896] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 915.369161] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4131cfe9-b373-4fc9-a06f-baef4eeb1416 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.376675] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86dc7b79-6302-48dd-ba1c-2d4402a0a91b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.405458] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e1af22a-ba03-49cf-929a-0cb5a909fda1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.412612] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0267edb-6892-4d75-adb5-8a81bd4634f3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.426468] env[62824]: DEBUG nova.compute.provider_tree [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 915.930051] env[62824]: DEBUG nova.scheduler.client.report [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 915.931444] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62824) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 915.931628] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 1.593s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.926269] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 918.772629] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 918.772888] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 918.773058] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 918.773214] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 918.773350] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62824) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11064}} [ 920.774623] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 920.775025] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Starting heal instance info cache {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10445}} [ 920.775025] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Rebuilding the list of instances to heal {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10449}} [ 921.277660] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Didn't find any instances for network info cache update. {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10531}} [ 974.772866] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 975.772833] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 975.773084] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager.update_available_resource {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 976.275964] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.276267] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.276385] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.276536] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62824) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 976.277478] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f60abfd-617f-4559-9d29-8903b432aa40 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.285521] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3227d953-90a9-452b-a58c-cfd808d4d0bd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.299278] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5bc1dc2-7cb6-4441-b427-c761db25381d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.305357] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fab1c33-fdd0-41bb-a083-35cfb450a209 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.334496] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181287MB free_disk=175GB free_vcpus=48 pci_devices=None {{(pid=62824) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 976.334623] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.334814] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.353204] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 977.353483] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 977.367018] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-441bdf75-95d7-40a7-94bd-5f76ea69cd85 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.374751] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d496812-c8a1-47e0-b3b8-e4a9abb641d5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.403879] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae3b0c29-c069-47e1-a85d-ae0a6abb6f6a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.410757] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d52eaa54-df8b-45bf-815b-f12df59942b3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.423444] env[62824]: DEBUG nova.compute.provider_tree [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 977.926441] env[62824]: DEBUG nova.scheduler.client.report [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 977.927760] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62824) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 977.927941] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 1.593s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.923551] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 978.923930] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 979.772853] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 980.769587] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 981.273933] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 981.274231] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 981.274312] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62824) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11064}} [ 982.773758] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 982.774281] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Starting heal instance info cache {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10445}} [ 982.774281] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Rebuilding the list of instances to heal {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10449}} [ 983.277441] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Didn't find any instances for network info cache update. {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10531}} [ 1035.772504] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1036.768739] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1036.772386] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1037.773718] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager.update_available_resource {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1038.277581] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1038.277826] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1038.278009] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.278170] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62824) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1038.279105] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8240421-ff35-4f2a-868b-39c5bfd476e9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.287442] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd8d2645-1695-464b-8bd3-e5e3de9b7d9f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.300974] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14c2d231-7640-42c3-b49f-596cff1b3e00 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.307096] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87d9fb87-7d11-444e-998b-c39c4609896e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.335411] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181305MB free_disk=175GB free_vcpus=48 pci_devices=None {{(pid=62824) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1038.335544] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1038.335723] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.354849] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1039.355105] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1039.367545] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57f129a3-eab7-447b-a941-6f03ee1a5f7b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.374831] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66df4389-7a12-49cf-adb4-a790e62f9bc7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.403614] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a52355d1-02f2-4969-8390-3599edfd2d42 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.410787] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c521b197-ed6e-45db-9ef8-a1c3d76d4622 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.423687] env[62824]: DEBUG nova.compute.provider_tree [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1039.926963] env[62824]: DEBUG nova.scheduler.client.report [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1039.928421] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62824) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1039.928607] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 1.593s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1041.928747] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1041.929123] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1041.929123] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1042.772783] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1042.772953] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62824) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11064}} [ 1044.774120] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1044.774541] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Starting heal instance info cache {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10445}} [ 1044.774541] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Rebuilding the list of instances to heal {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10449}} [ 1045.277178] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Didn't find any instances for network info cache update. {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10531}} [ 1094.774531] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1094.774945] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Cleaning up deleted instances with incomplete migration {{(pid=62824) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11783}} [ 1097.276555] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1097.773274] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager.update_available_resource {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1098.276613] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1098.276996] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1098.277073] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1098.277209] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62824) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1098.278130] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d47ac2cc-a08c-4bc6-b31e-760a130967bf {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.286455] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21d63289-01e9-45c4-ac59-5a5af47b977f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.299917] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-195a3db1-fb89-46ba-9e8d-863a61c4227a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.305837] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45f15205-2100-40cf-b9dd-3ad2ebd80285 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.334573] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181291MB free_disk=175GB free_vcpus=48 pci_devices=None {{(pid=62824) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1098.334694] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1098.334877] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1099.351882] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1099.352122] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1099.364237] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ad074b4-5a4c-49de-9630-df777865f3b5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.371576] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04061aaf-17f5-486a-82ce-bf54b24446eb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.400090] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb23aeb6-8fa7-45f7-b6dd-e17c29876045 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.406799] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14559ce1-92ac-42d0-95e3-bd89e2307c53 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.419698] env[62824]: DEBUG nova.compute.provider_tree [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1099.922549] env[62824]: DEBUG nova.scheduler.client.report [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1099.923941] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62824) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1099.924160] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 1.589s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1100.920058] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1100.920411] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1100.920676] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1101.769784] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1102.773082] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1102.773082] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1104.276449] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1104.276805] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1104.276805] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62824) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11064}} [ 1104.276968] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1104.277073] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Cleaning up deleted instances {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11745}} [ 1104.779261] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] There are 0 instances to clean {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11754}} [ 1106.276850] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1106.277266] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Starting heal instance info cache {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10445}} [ 1106.277266] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Rebuilding the list of instances to heal {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10449}} [ 1106.780191] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Didn't find any instances for network info cache update. {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10531}} [ 1150.511723] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._sync_power_states {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1151.015038] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-41517485-76de-4e8f-a181-a012faec848d None None] Getting list of instances from cluster (obj){ [ 1151.015038] env[62824]: value = "domain-c8" [ 1151.015038] env[62824]: _type = "ClusterComputeResource" [ 1151.015038] env[62824]: } {{(pid=62824) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 1151.016149] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91954c9e-f421-492c-b0d1-2f62d5d74b29 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.025170] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-41517485-76de-4e8f-a181-a012faec848d None None] Got total of 0 instances {{(pid=62824) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 1157.289746] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1157.774056] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager.update_available_resource {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1158.276851] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1158.277097] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1158.277271] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1158.277425] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62824) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1158.278344] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a4622ad-ddc6-4787-bb24-faf18317a0d3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.286603] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce63c426-b527-4394-a85e-38c38e78c570 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.300123] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-924d4a43-5b0b-4ddf-abe1-0c2263c99b0d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.306017] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9aa7584-c87e-469d-9759-2327b9b42937 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.334604] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181295MB free_disk=175GB free_vcpus=48 pci_devices=None {{(pid=62824) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1158.334736] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1158.334930] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1159.451088] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1159.451425] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1159.466793] env[62824]: DEBUG nova.scheduler.client.report [None req-41517485-76de-4e8f-a181-a012faec848d None None] Refreshing inventories for resource provider bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1159.478112] env[62824]: DEBUG nova.scheduler.client.report [None req-41517485-76de-4e8f-a181-a012faec848d None None] Updating ProviderTree inventory for provider bf95157b-a274-42de-9ccf-9851128a44a1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1159.478300] env[62824]: DEBUG nova.compute.provider_tree [None req-41517485-76de-4e8f-a181-a012faec848d None None] Updating inventory in ProviderTree for provider bf95157b-a274-42de-9ccf-9851128a44a1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1159.487380] env[62824]: DEBUG nova.scheduler.client.report [None req-41517485-76de-4e8f-a181-a012faec848d None None] Refreshing aggregate associations for resource provider bf95157b-a274-42de-9ccf-9851128a44a1, aggregates: None {{(pid=62824) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1159.501951] env[62824]: DEBUG nova.scheduler.client.report [None req-41517485-76de-4e8f-a181-a012faec848d None None] Refreshing trait associations for resource provider bf95157b-a274-42de-9ccf-9851128a44a1, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_ISO {{(pid=62824) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1159.513269] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09ee89d9-771d-4141-8b49-090c81b4913a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.520775] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dec42cfe-71d1-45e2-b568-4e7d9b743f16 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.549215] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24cf0294-90d2-4344-88ce-a2923291dfab {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.555929] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98661c48-fca0-4f1e-9713-b25a44219ca0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.569233] env[62824]: DEBUG nova.compute.provider_tree [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1160.072483] env[62824]: DEBUG nova.scheduler.client.report [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1160.073755] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62824) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1160.073935] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 1.739s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1161.074797] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1161.074797] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1162.772911] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1163.773496] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1163.773496] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62824) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11064}} [ 1164.774341] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1165.774268] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1165.774268] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Starting heal instance info cache {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10445}} [ 1165.774268] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Rebuilding the list of instances to heal {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10449}} [ 1166.277267] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Didn't find any instances for network info cache update. {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10531}} [ 1166.277512] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1217.773613] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1219.772675] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1219.773095] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager.update_available_resource {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1220.275994] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1220.276258] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1220.276422] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1220.276585] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62824) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1220.277522] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91ff60a9-e745-4da8-8e02-4b0b7865de4c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.285800] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b46e085e-b235-4b6f-b047-689dc27b2226 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.300778] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ebd84c6-87dc-4041-8c79-3c56871d24c7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.307191] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce589125-7113-4b37-b76e-c49f3632bab7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.336769] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181303MB free_disk=175GB free_vcpus=48 pci_devices=None {{(pid=62824) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1220.336889] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1220.337055] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1221.354707] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1221.354942] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1221.367724] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b121b1f-8e28-4d6e-8477-7178135dadcc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.375211] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9dc7250-b5c0-4ede-9cef-4caa343681e4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.403847] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8165283-2e26-45ba-beca-d166f08625f4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.410992] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-299ddaa8-3e3f-4c86-ae89-9a4e82634da2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.423841] env[62824]: DEBUG nova.compute.provider_tree [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1221.927271] env[62824]: DEBUG nova.scheduler.client.report [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1221.928521] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62824) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1221.928702] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 1.592s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1223.924705] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1223.925181] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1224.430494] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1224.430494] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1224.430494] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62824) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11064}} [ 1226.773869] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1226.778205] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Starting heal instance info cache {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10445}} [ 1226.778205] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Rebuilding the list of instances to heal {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10449}} [ 1227.277216] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Didn't find any instances for network info cache update. {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10531}} [ 1227.277470] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1227.772575] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1279.775131] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1279.775567] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager.update_available_resource {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1280.278993] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1280.279257] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1280.279428] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1280.279585] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62824) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1280.280541] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65f2f887-71c8-4f3b-8a0f-525f30b0323a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1280.288796] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d68be94-958c-4539-90b1-314f2a72de86 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1280.302653] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d0b55e4-2121-49b4-9702-e0419c5a9a0f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1280.308786] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b4471f5-b1d7-4588-a326-cfc6a50ef76a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1280.336884] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181284MB free_disk=175GB free_vcpus=48 pci_devices=None {{(pid=62824) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1280.337028] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1280.337215] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1281.357656] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1281.357948] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1281.372050] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47d80e3f-d322-413b-a639-47fe6e889f4a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1281.379276] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b8f06fd-3b02-4c89-bf43-00941f47b914 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1281.408089] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c51c23b-b509-4c3c-a4be-8cc57ba58630 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1281.415218] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b158cc8c-6f69-4ff8-b887-34f66aa70a77 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1281.428249] env[62824]: DEBUG nova.compute.provider_tree [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1281.930907] env[62824]: DEBUG nova.scheduler.client.report [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1281.932238] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62824) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1281.932424] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 1.595s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1283.930079] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1283.930437] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1283.930437] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1283.930595] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1283.930659] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62824) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11064}} [ 1286.773602] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1286.774016] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Starting heal instance info cache {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10445}} [ 1286.774016] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Rebuilding the list of instances to heal {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10449}} [ 1287.277186] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Didn't find any instances for network info cache update. {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10531}} [ 1287.277423] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1289.772936] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1340.773201] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1340.773683] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager.update_available_resource {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1341.276956] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1341.277240] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1341.277375] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1341.277526] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62824) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1341.278417] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee316e23-d9be-40f0-880e-6d467529e8d4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1341.286861] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57fb29e7-1485-4ed2-93d2-6c4b4466943a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1341.301347] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd55588e-1390-4d7c-9f4a-577f5626bf95 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1341.307487] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d697b2d-6aa6-43d6-b605-e09457ee34b6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1341.335362] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181286MB free_disk=175GB free_vcpus=48 pci_devices=None {{(pid=62824) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1341.335496] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1341.335685] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1342.353933] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1342.354193] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1342.366964] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16ab5058-a20d-468a-8608-5dfd67b3f2ec {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1342.374511] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7dfb80b-d4fb-4d25-998b-2c4b8be623fc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1342.404665] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f0d19bf-239f-4905-bd13-6239873511d2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1342.412460] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32074f7b-e9ee-42c0-9815-55a977c2cb23 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1342.425633] env[62824]: DEBUG nova.compute.provider_tree [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1342.929205] env[62824]: DEBUG nova.scheduler.client.report [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1342.930483] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62824) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1342.930680] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 1.595s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1345.930617] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1345.930998] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1345.930998] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1345.931421] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1345.931421] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62824) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11064}} [ 1346.770559] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1348.774426] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1348.774426] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Starting heal instance info cache {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10445}} [ 1348.774426] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Rebuilding the list of instances to heal {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10449}} [ 1349.277507] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Didn't find any instances for network info cache update. {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10531}} [ 1349.277507] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1351.773799] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1400.775640] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1400.776018] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Cleaning up deleted instances with incomplete migration {{(pid=62824) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11783}} [ 1403.275530] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1403.275911] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager.update_available_resource {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1403.782180] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1403.782400] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1403.782566] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1403.782718] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62824) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1403.783642] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a5d75e0-0d72-4866-8892-d23b13a5daaf {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1403.792485] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-024bfcf2-fcac-47db-884b-047e1e867e10 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1403.807753] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f903bdcf-737f-48fa-8e69-0e284410fc70 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1403.814374] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc780385-f9af-4650-9f66-cfd5e248c8d8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1403.843236] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181294MB free_disk=175GB free_vcpus=48 pci_devices=None {{(pid=62824) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1403.843399] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1403.843572] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1404.861214] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1404.861469] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1404.874766] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b7a2f32-1d61-409f-ac9b-7a4d57106975 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1404.882920] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d13e178-a180-4943-98d8-9a5fbaabf1bd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1404.911533] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5761b681-0ac4-43de-83c2-0100026c75d8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1404.918472] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5c37bca-68e7-44bd-873d-3fbb58bf05d0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1404.930992] env[62824]: DEBUG nova.compute.provider_tree [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1405.434196] env[62824]: DEBUG nova.scheduler.client.report [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1405.435461] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62824) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1405.435643] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 1.592s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1405.933412] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1405.933807] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1405.933807] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1405.933931] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1405.934116] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62824) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11064}} [ 1406.773271] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1409.276933] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1409.276933] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1409.276933] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Cleaning up deleted instances {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11745}} [ 1409.779688] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] There are 0 instances to clean {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11754}} [ 1411.278632] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1411.278632] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Starting heal instance info cache {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10445}} [ 1411.278632] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Rebuilding the list of instances to heal {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10449}} [ 1411.783590] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Didn't find any instances for network info cache update. {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10531}} [ 1411.783590] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1463.773751] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1464.773658] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager.update_available_resource {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1465.276922] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1465.277288] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1465.277329] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1465.277470] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62824) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1465.278365] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4a497f0-c74b-44cd-96cf-cc050be69c51 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1465.286314] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e1c6643-c6c8-45f1-a2a2-73ca54686a13 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1465.299969] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7da6d545-6d40-4bf0-8bbb-437dddc1b279 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1465.306029] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58115785-f855-43b4-87ae-9bcc51ce529d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1465.333815] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181304MB free_disk=175GB free_vcpus=48 pci_devices=None {{(pid=62824) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1465.333944] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1465.334134] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1466.372608] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1466.373022] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1466.394850] env[62824]: DEBUG nova.scheduler.client.report [None req-41517485-76de-4e8f-a181-a012faec848d None None] Refreshing inventories for resource provider bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1466.407676] env[62824]: DEBUG nova.scheduler.client.report [None req-41517485-76de-4e8f-a181-a012faec848d None None] Updating ProviderTree inventory for provider bf95157b-a274-42de-9ccf-9851128a44a1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1466.407889] env[62824]: DEBUG nova.compute.provider_tree [None req-41517485-76de-4e8f-a181-a012faec848d None None] Updating inventory in ProviderTree for provider bf95157b-a274-42de-9ccf-9851128a44a1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1466.418886] env[62824]: DEBUG nova.scheduler.client.report [None req-41517485-76de-4e8f-a181-a012faec848d None None] Refreshing aggregate associations for resource provider bf95157b-a274-42de-9ccf-9851128a44a1, aggregates: None {{(pid=62824) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1466.435133] env[62824]: DEBUG nova.scheduler.client.report [None req-41517485-76de-4e8f-a181-a012faec848d None None] Refreshing trait associations for resource provider bf95157b-a274-42de-9ccf-9851128a44a1, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_ISO {{(pid=62824) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1466.447278] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2c44f60-95c6-415a-9aa1-cb5405194a28 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1466.454674] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-985b502b-ead3-4210-bc71-e8d7049f6ccc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1466.484735] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3cfd1a0-6955-466c-8d7f-8a743fa25178 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1466.491408] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-165576b1-9bda-4123-a6dc-6c5ee20b1721 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1466.504365] env[62824]: DEBUG nova.compute.provider_tree [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1467.008048] env[62824]: DEBUG nova.scheduler.client.report [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1467.009312] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62824) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1467.009487] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 1.675s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1468.004803] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1468.005203] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1468.005203] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1468.005375] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1468.005507] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62824) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11064}} [ 1468.773592] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1470.768980] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1471.274310] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1471.274466] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Starting heal instance info cache {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10445}} [ 1471.274585] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Rebuilding the list of instances to heal {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10449}} [ 1471.777969] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Didn't find any instances for network info cache update. {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10531}} [ 1471.778340] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1523.773795] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1525.768779] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1526.772977] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1526.773424] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager.update_available_resource {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1527.276725] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1527.276942] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1527.277141] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1527.277299] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62824) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1527.278247] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36a1032c-3864-42e3-9645-1619e205c690 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1527.286489] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9128ebc2-ad94-4d84-a8cc-7fd1e4ecfd80 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1527.300783] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53be5fa8-6012-4d85-93d5-9ebee4f859d4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1527.307373] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a391daa-4fac-4e61-9ced-d26080d78c95 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1527.336633] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181292MB free_disk=175GB free_vcpus=48 pci_devices=None {{(pid=62824) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1527.336829] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1527.336989] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1528.358848] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1528.359153] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1528.372714] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b4b0baf-623e-4e8b-be51-6c806ef4e6c5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1528.380821] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffe59455-8dca-4d1b-92a8-b994640193d1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1528.411357] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2d985ad-5693-4fcd-916a-d4ce8cd0ffcc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1528.418376] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a216bfa3-8a54-4d4d-b9f2-851d00062897 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1528.431074] env[62824]: DEBUG nova.compute.provider_tree [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1528.934782] env[62824]: DEBUG nova.scheduler.client.report [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1528.936072] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62824) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1528.936260] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 1.599s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1529.936489] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1529.936821] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1529.936967] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1529.937064] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62824) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11064}} [ 1530.773626] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1530.773866] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Starting heal instance info cache {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10445}} [ 1530.773928] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Rebuilding the list of instances to heal {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10449}} [ 1531.276238] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Didn't find any instances for network info cache update. {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10531}} [ 1533.773706] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1583.774795] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1586.769285] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1587.772280] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1587.772742] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1587.772742] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62824) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11064}} [ 1587.772854] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager.update_available_resource {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1588.276062] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1588.276399] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1588.276632] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1588.276850] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62824) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1588.278167] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42b47d35-f9b7-4f73-aa95-373edc7cfc77 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1588.287332] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cbcf4d0-f9b6-4110-95b8-7fb5d8e550d4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1588.302525] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ee81be6-459a-4c36-9b3d-9ceb88033bce {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1588.309277] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de4c7dde-1ed2-4827-82e5-15a89b2a48c1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1588.337908] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181292MB free_disk=175GB free_vcpus=48 pci_devices=None {{(pid=62824) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1588.338110] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1588.338271] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1589.356331] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1589.356570] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1589.369332] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a3d54cb-6668-46ee-a8c0-b402bb416937 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1589.377386] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ec185b0-cf58-4235-a8a1-1ba717369da8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1589.407758] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d936ecc-34af-475d-92b0-4b68c3e173fe {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1589.415462] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-273ee90b-7a99-4f69-81f1-80bba50034ca {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1589.428634] env[62824]: DEBUG nova.compute.provider_tree [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1589.931297] env[62824]: DEBUG nova.scheduler.client.report [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1589.932649] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62824) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1589.932838] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 1.595s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1590.933806] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1590.934201] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1591.768791] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1592.273856] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1592.274324] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Starting heal instance info cache {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10445}} [ 1592.274324] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Rebuilding the list of instances to heal {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10449}} [ 1592.776618] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Didn't find any instances for network info cache update. {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10531}} [ 1595.774363] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1644.775798] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1647.768885] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1648.772679] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1648.773098] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1648.773175] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1648.773280] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62824) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11064}} [ 1649.773250] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager.update_available_resource {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1650.276576] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1650.276831] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1650.276965] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1650.277133] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62824) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1650.278040] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-207a723f-16eb-44d0-85bb-e8d4829771e6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1650.286207] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52efdcb3-45d4-422e-a7c9-4248b4895ec7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1650.299821] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74249bc0-68ef-4981-83db-37342be67bd6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1650.306252] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ec38604-33d4-4415-ad4d-e4601dad8d6e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1650.335555] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181286MB free_disk=175GB free_vcpus=48 pci_devices=None {{(pid=62824) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1650.335709] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1650.335904] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1651.358321] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1651.358571] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1651.371467] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dfa23ea-603a-442d-8688-65cc60e01f04 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1651.379273] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11537a0a-a42d-43ab-a80a-35ecb752a1ff {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1651.408046] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b04e11f4-d8ad-48b7-8cae-13f8d0bdc173 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1651.415632] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b26ff13c-d830-4fb0-8dbe-032688e53a39 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1651.429486] env[62824]: DEBUG nova.compute.provider_tree [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1651.933832] env[62824]: DEBUG nova.scheduler.client.report [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1651.933832] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62824) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1651.933832] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 1.598s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1653.933576] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1653.933976] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Starting heal instance info cache {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10445}} [ 1653.933976] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Rebuilding the list of instances to heal {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10449}} [ 1654.437105] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Didn't find any instances for network info cache update. {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10531}} [ 1654.437355] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1655.772738] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1704.775768] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1707.768533] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1709.773356] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1709.773724] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62824) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11064}} [ 1709.773724] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager.update_available_resource {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1710.277470] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1710.277737] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1710.277896] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1710.278069] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62824) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1710.278994] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e670be2d-18cc-445c-b721-7ddc79d605f0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1710.287411] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51f35dd5-7076-459f-bbbd-9212ebbf9728 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1710.301533] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1598d73-0e6e-4aa2-9c49-1f30aed2f62c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1710.308040] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c355b6c-9530-4a69-816a-1de06e800259 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1710.337588] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181292MB free_disk=175GB free_vcpus=48 pci_devices=None {{(pid=62824) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1710.337790] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1710.337929] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1711.358475] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1711.358720] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1711.371030] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffc1d786-85be-42cc-ad78-e17e84efd094 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1711.378661] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a21c836-e80d-4d84-9898-4f7a0a8da331 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1711.408024] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75ce0e90-9818-4af4-99c5-db2a410d6d0f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1711.414889] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a330e38c-a217-4060-afcd-13f3df36ca78 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1711.428344] env[62824]: DEBUG nova.compute.provider_tree [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1711.931746] env[62824]: DEBUG nova.scheduler.client.report [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1711.932992] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62824) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1711.933191] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 1.595s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1711.933432] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1711.933568] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Cleaning up deleted instances with incomplete migration {{(pid=62824) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11783}} [ 1713.435625] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1713.435958] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1713.773182] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1713.773182] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Starting heal instance info cache {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10445}} [ 1713.773182] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Rebuilding the list of instances to heal {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10449}} [ 1714.275689] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Didn't find any instances for network info cache update. {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10531}} [ 1714.275927] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1715.773670] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1716.768874] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1717.273652] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1723.278826] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1723.279153] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Cleaning up deleted instances {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11745}} [ 1723.782357] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] There are 0 instances to clean {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11754}} [ 1765.015143] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._sync_power_states {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1765.518625] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-41517485-76de-4e8f-a181-a012faec848d None None] Getting list of instances from cluster (obj){ [ 1765.518625] env[62824]: value = "domain-c8" [ 1765.518625] env[62824]: _type = "ClusterComputeResource" [ 1765.518625] env[62824]: } {{(pid=62824) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 1765.519737] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0b694d4-c2d3-4d52-8eea-b44b535c05cc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1765.528342] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-41517485-76de-4e8f-a181-a012faec848d None None] Got total of 0 instances {{(pid=62824) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 1767.286194] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1767.768167] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1769.773492] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager.update_available_resource {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1770.276915] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1770.277177] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1770.277336] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1770.277485] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62824) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1770.278392] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dbbdd90-929c-4536-a2d6-15d643a46dc8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1770.286915] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee06d1da-ad6b-4914-8111-de41aeb60622 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1770.301758] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6a4bab9-6a77-4ab6-99ec-0dd14f646162 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1770.308192] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c066a235-136a-4381-848b-2e50f025acf5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1770.336398] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181295MB free_disk=175GB free_vcpus=48 pci_devices=None {{(pid=62824) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1770.336568] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1770.336733] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1771.459722] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1771.460028] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1771.475349] env[62824]: DEBUG nova.scheduler.client.report [None req-41517485-76de-4e8f-a181-a012faec848d None None] Refreshing inventories for resource provider bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1771.486701] env[62824]: DEBUG nova.scheduler.client.report [None req-41517485-76de-4e8f-a181-a012faec848d None None] Updating ProviderTree inventory for provider bf95157b-a274-42de-9ccf-9851128a44a1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1771.486880] env[62824]: DEBUG nova.compute.provider_tree [None req-41517485-76de-4e8f-a181-a012faec848d None None] Updating inventory in ProviderTree for provider bf95157b-a274-42de-9ccf-9851128a44a1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1771.495982] env[62824]: DEBUG nova.scheduler.client.report [None req-41517485-76de-4e8f-a181-a012faec848d None None] Refreshing aggregate associations for resource provider bf95157b-a274-42de-9ccf-9851128a44a1, aggregates: None {{(pid=62824) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1771.510552] env[62824]: DEBUG nova.scheduler.client.report [None req-41517485-76de-4e8f-a181-a012faec848d None None] Refreshing trait associations for resource provider bf95157b-a274-42de-9ccf-9851128a44a1, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_ISO {{(pid=62824) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1771.522307] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aaac31c1-f2fb-4c7e-9b86-65ee758f76aa {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1771.530227] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a447e76-06f7-4b68-9167-017cab6b98b1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1771.560827] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8cc25d4-d2fa-4b6d-9f7b-97ae5380c058 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1771.567683] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a37c6e01-c542-417a-8b5c-9a3d512ccf47 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1771.580490] env[62824]: DEBUG nova.compute.provider_tree [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1772.085044] env[62824]: DEBUG nova.scheduler.client.report [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1772.085044] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62824) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1772.085308] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 1.748s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1774.085197] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1774.085677] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1774.085762] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1774.086038] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1774.086248] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62824) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11064}} [ 1775.774191] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1775.774606] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Starting heal instance info cache {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10445}} [ 1775.774606] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Rebuilding the list of instances to heal {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10449}} [ 1776.278028] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Didn't find any instances for network info cache update. {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10531}} [ 1777.773131] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1810.400885] env[62824]: DEBUG oslo_concurrency.lockutils [None req-84c279f8-4acc-49b0-9f33-b648872f4030 tempest-FloatingIPsAssociationTestJSON-1164183782 tempest-FloatingIPsAssociationTestJSON-1164183782-project-member] Acquiring lock "764d7b4f-49d3-4ddf-8152-c08721dbce59" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1810.401216] env[62824]: DEBUG oslo_concurrency.lockutils [None req-84c279f8-4acc-49b0-9f33-b648872f4030 tempest-FloatingIPsAssociationTestJSON-1164183782 tempest-FloatingIPsAssociationTestJSON-1164183782-project-member] Lock "764d7b4f-49d3-4ddf-8152-c08721dbce59" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1810.724837] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b73330de-7630-444b-9674-b1882b4c4da9 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Acquiring lock "579e1ca1-a6b6-4fde-8368-3e635a303032" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1810.724837] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b73330de-7630-444b-9674-b1882b4c4da9 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Lock "579e1ca1-a6b6-4fde-8368-3e635a303032" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1810.904646] env[62824]: DEBUG nova.compute.manager [None req-84c279f8-4acc-49b0-9f33-b648872f4030 tempest-FloatingIPsAssociationTestJSON-1164183782 tempest-FloatingIPsAssociationTestJSON-1164183782-project-member] [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1811.225912] env[62824]: DEBUG nova.compute.manager [None req-b73330de-7630-444b-9674-b1882b4c4da9 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1811.454693] env[62824]: DEBUG oslo_concurrency.lockutils [None req-84c279f8-4acc-49b0-9f33-b648872f4030 tempest-FloatingIPsAssociationTestJSON-1164183782 tempest-FloatingIPsAssociationTestJSON-1164183782-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1811.454991] env[62824]: DEBUG oslo_concurrency.lockutils [None req-84c279f8-4acc-49b0-9f33-b648872f4030 tempest-FloatingIPsAssociationTestJSON-1164183782 tempest-FloatingIPsAssociationTestJSON-1164183782-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1811.456597] env[62824]: INFO nova.compute.claims [None req-84c279f8-4acc-49b0-9f33-b648872f4030 tempest-FloatingIPsAssociationTestJSON-1164183782 tempest-FloatingIPsAssociationTestJSON-1164183782-project-member] [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1811.753857] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b73330de-7630-444b-9674-b1882b4c4da9 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1812.525493] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-befe225d-f52f-431a-832e-b38ab0b53caa {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1812.533317] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a2f8c32-496d-4523-a3ac-2ddba6fa8d8b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1812.564738] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71111bd2-2759-4c2c-9733-c02f253618ef {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1812.572421] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cdd5f58-88bb-4558-80e3-f2d91f4dd0b3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1812.586009] env[62824]: DEBUG nova.compute.provider_tree [None req-84c279f8-4acc-49b0-9f33-b648872f4030 tempest-FloatingIPsAssociationTestJSON-1164183782 tempest-FloatingIPsAssociationTestJSON-1164183782-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1813.092110] env[62824]: DEBUG nova.scheduler.client.report [None req-84c279f8-4acc-49b0-9f33-b648872f4030 tempest-FloatingIPsAssociationTestJSON-1164183782 tempest-FloatingIPsAssociationTestJSON-1164183782-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1813.514092] env[62824]: DEBUG oslo_concurrency.lockutils [None req-24599d6d-2f76-445f-8b64-070ec4ee0da9 tempest-TenantUsagesTestJSON-486947106 tempest-TenantUsagesTestJSON-486947106-project-member] Acquiring lock "9f016121-09f1-4da7-bbd0-7769d5104d27" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1813.514494] env[62824]: DEBUG oslo_concurrency.lockutils [None req-24599d6d-2f76-445f-8b64-070ec4ee0da9 tempest-TenantUsagesTestJSON-486947106 tempest-TenantUsagesTestJSON-486947106-project-member] Lock "9f016121-09f1-4da7-bbd0-7769d5104d27" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1813.599090] env[62824]: DEBUG oslo_concurrency.lockutils [None req-84c279f8-4acc-49b0-9f33-b648872f4030 tempest-FloatingIPsAssociationTestJSON-1164183782 tempest-FloatingIPsAssociationTestJSON-1164183782-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.144s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1813.600125] env[62824]: DEBUG nova.compute.manager [None req-84c279f8-4acc-49b0-9f33-b648872f4030 tempest-FloatingIPsAssociationTestJSON-1164183782 tempest-FloatingIPsAssociationTestJSON-1164183782-project-member] [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1813.608967] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b73330de-7630-444b-9674-b1882b4c4da9 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.855s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1813.614096] env[62824]: INFO nova.compute.claims [None req-b73330de-7630-444b-9674-b1882b4c4da9 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1814.024480] env[62824]: DEBUG nova.compute.manager [None req-24599d6d-2f76-445f-8b64-070ec4ee0da9 tempest-TenantUsagesTestJSON-486947106 tempest-TenantUsagesTestJSON-486947106-project-member] [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1814.112594] env[62824]: DEBUG nova.compute.utils [None req-84c279f8-4acc-49b0-9f33-b648872f4030 tempest-FloatingIPsAssociationTestJSON-1164183782 tempest-FloatingIPsAssociationTestJSON-1164183782-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1814.117441] env[62824]: DEBUG nova.compute.manager [None req-84c279f8-4acc-49b0-9f33-b648872f4030 tempest-FloatingIPsAssociationTestJSON-1164183782 tempest-FloatingIPsAssociationTestJSON-1164183782-project-member] [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1814.117781] env[62824]: DEBUG nova.network.neutron [None req-84c279f8-4acc-49b0-9f33-b648872f4030 tempest-FloatingIPsAssociationTestJSON-1164183782 tempest-FloatingIPsAssociationTestJSON-1164183782-project-member] [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1814.181421] env[62824]: DEBUG oslo_concurrency.lockutils [None req-6070f146-69b5-4aa1-b292-f106b5e9564b tempest-ServerAddressesNegativeTestJSON-1575558569 tempest-ServerAddressesNegativeTestJSON-1575558569-project-member] Acquiring lock "6cc60e74-0474-47a5-ba57-c1c646ce0e1d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1814.181725] env[62824]: DEBUG oslo_concurrency.lockutils [None req-6070f146-69b5-4aa1-b292-f106b5e9564b tempest-ServerAddressesNegativeTestJSON-1575558569 tempest-ServerAddressesNegativeTestJSON-1575558569-project-member] Lock "6cc60e74-0474-47a5-ba57-c1c646ce0e1d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.008s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1814.557442] env[62824]: DEBUG oslo_concurrency.lockutils [None req-24599d6d-2f76-445f-8b64-070ec4ee0da9 tempest-TenantUsagesTestJSON-486947106 tempest-TenantUsagesTestJSON-486947106-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1814.619417] env[62824]: DEBUG nova.compute.manager [None req-84c279f8-4acc-49b0-9f33-b648872f4030 tempest-FloatingIPsAssociationTestJSON-1164183782 tempest-FloatingIPsAssociationTestJSON-1164183782-project-member] [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1814.670450] env[62824]: DEBUG nova.policy [None req-84c279f8-4acc-49b0-9f33-b648872f4030 tempest-FloatingIPsAssociationTestJSON-1164183782 tempest-FloatingIPsAssociationTestJSON-1164183782-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '990501f6714f45adba13e1e08171a50e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f4e2a4a7414a4cfc816af70ecdb89c9b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 1814.685737] env[62824]: DEBUG nova.compute.manager [None req-6070f146-69b5-4aa1-b292-f106b5e9564b tempest-ServerAddressesNegativeTestJSON-1575558569 tempest-ServerAddressesNegativeTestJSON-1575558569-project-member] [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1814.764579] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-deb963ad-7c79-4bf8-8c75-ef7e691269fd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1814.774225] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa27e88f-c833-4548-a55b-21636c274338 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1814.808640] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fad44e5-7b0c-47e6-a491-1cc694d83042 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1814.817442] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b77d35ba-f831-4ea9-8319-d883168080e9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1814.832129] env[62824]: DEBUG nova.compute.provider_tree [None req-b73330de-7630-444b-9674-b1882b4c4da9 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1814.909413] env[62824]: DEBUG oslo_concurrency.lockutils [None req-250af368-8463-49f4-8a2e-de088db419fa tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Acquiring lock "8b138cc7-8bde-48c8-b1bb-cec6ead877d3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1814.909708] env[62824]: DEBUG oslo_concurrency.lockutils [None req-250af368-8463-49f4-8a2e-de088db419fa tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Lock "8b138cc7-8bde-48c8-b1bb-cec6ead877d3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1815.229933] env[62824]: DEBUG oslo_concurrency.lockutils [None req-6070f146-69b5-4aa1-b292-f106b5e9564b tempest-ServerAddressesNegativeTestJSON-1575558569 tempest-ServerAddressesNegativeTestJSON-1575558569-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1815.337112] env[62824]: DEBUG nova.scheduler.client.report [None req-b73330de-7630-444b-9674-b1882b4c4da9 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1815.415744] env[62824]: DEBUG nova.compute.manager [None req-250af368-8463-49f4-8a2e-de088db419fa tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1815.634648] env[62824]: DEBUG nova.compute.manager [None req-84c279f8-4acc-49b0-9f33-b648872f4030 tempest-FloatingIPsAssociationTestJSON-1164183782 tempest-FloatingIPsAssociationTestJSON-1164183782-project-member] [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1815.668974] env[62824]: DEBUG nova.virt.hardware [None req-84c279f8-4acc-49b0-9f33-b648872f4030 tempest-FloatingIPsAssociationTestJSON-1164183782 tempest-FloatingIPsAssociationTestJSON-1164183782-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1815.669289] env[62824]: DEBUG nova.virt.hardware [None req-84c279f8-4acc-49b0-9f33-b648872f4030 tempest-FloatingIPsAssociationTestJSON-1164183782 tempest-FloatingIPsAssociationTestJSON-1164183782-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1815.669484] env[62824]: DEBUG nova.virt.hardware [None req-84c279f8-4acc-49b0-9f33-b648872f4030 tempest-FloatingIPsAssociationTestJSON-1164183782 tempest-FloatingIPsAssociationTestJSON-1164183782-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1815.670283] env[62824]: DEBUG nova.virt.hardware [None req-84c279f8-4acc-49b0-9f33-b648872f4030 tempest-FloatingIPsAssociationTestJSON-1164183782 tempest-FloatingIPsAssociationTestJSON-1164183782-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1815.671424] env[62824]: DEBUG nova.virt.hardware [None req-84c279f8-4acc-49b0-9f33-b648872f4030 tempest-FloatingIPsAssociationTestJSON-1164183782 tempest-FloatingIPsAssociationTestJSON-1164183782-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1815.671424] env[62824]: DEBUG nova.virt.hardware [None req-84c279f8-4acc-49b0-9f33-b648872f4030 tempest-FloatingIPsAssociationTestJSON-1164183782 tempest-FloatingIPsAssociationTestJSON-1164183782-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1815.671424] env[62824]: DEBUG nova.virt.hardware [None req-84c279f8-4acc-49b0-9f33-b648872f4030 tempest-FloatingIPsAssociationTestJSON-1164183782 tempest-FloatingIPsAssociationTestJSON-1164183782-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1815.671424] env[62824]: DEBUG nova.virt.hardware [None req-84c279f8-4acc-49b0-9f33-b648872f4030 tempest-FloatingIPsAssociationTestJSON-1164183782 tempest-FloatingIPsAssociationTestJSON-1164183782-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1815.671424] env[62824]: DEBUG nova.virt.hardware [None req-84c279f8-4acc-49b0-9f33-b648872f4030 tempest-FloatingIPsAssociationTestJSON-1164183782 tempest-FloatingIPsAssociationTestJSON-1164183782-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1815.671567] env[62824]: DEBUG nova.virt.hardware [None req-84c279f8-4acc-49b0-9f33-b648872f4030 tempest-FloatingIPsAssociationTestJSON-1164183782 tempest-FloatingIPsAssociationTestJSON-1164183782-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1815.671567] env[62824]: DEBUG nova.virt.hardware [None req-84c279f8-4acc-49b0-9f33-b648872f4030 tempest-FloatingIPsAssociationTestJSON-1164183782 tempest-FloatingIPsAssociationTestJSON-1164183782-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1815.672455] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97d8bb03-0995-453c-a0f2-99aac7ee7b3c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1815.688094] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bba46960-4c92-41fe-a4d1-8b7fc008732e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1815.708236] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89255dad-844e-4967-ba57-9e72ea21a713 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1815.839589] env[62824]: DEBUG nova.network.neutron [None req-84c279f8-4acc-49b0-9f33-b648872f4030 tempest-FloatingIPsAssociationTestJSON-1164183782 tempest-FloatingIPsAssociationTestJSON-1164183782-project-member] [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] Successfully created port: 5f952c6c-084a-4768-97e4-16b121390652 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1815.844903] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b73330de-7630-444b-9674-b1882b4c4da9 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.235s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1815.845099] env[62824]: DEBUG nova.compute.manager [None req-b73330de-7630-444b-9674-b1882b4c4da9 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1815.847546] env[62824]: DEBUG oslo_concurrency.lockutils [None req-24599d6d-2f76-445f-8b64-070ec4ee0da9 tempest-TenantUsagesTestJSON-486947106 tempest-TenantUsagesTestJSON-486947106-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.290s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1815.849676] env[62824]: INFO nova.compute.claims [None req-24599d6d-2f76-445f-8b64-070ec4ee0da9 tempest-TenantUsagesTestJSON-486947106 tempest-TenantUsagesTestJSON-486947106-project-member] [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1815.943338] env[62824]: DEBUG oslo_concurrency.lockutils [None req-250af368-8463-49f4-8a2e-de088db419fa tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1816.355113] env[62824]: DEBUG nova.compute.utils [None req-b73330de-7630-444b-9674-b1882b4c4da9 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1816.357750] env[62824]: DEBUG nova.compute.manager [None req-b73330de-7630-444b-9674-b1882b4c4da9 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1816.357750] env[62824]: DEBUG nova.network.neutron [None req-b73330de-7630-444b-9674-b1882b4c4da9 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1816.603032] env[62824]: DEBUG nova.policy [None req-b73330de-7630-444b-9674-b1882b4c4da9 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '16d8bfd2c5264817a312b1911adc1a37', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dbe0dd011b0c43328d69799d09fc050c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 1816.861741] env[62824]: DEBUG nova.compute.manager [None req-b73330de-7630-444b-9674-b1882b4c4da9 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1817.009043] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76573222-f54c-4712-bc47-4ef57cdf13a4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1817.022315] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71725cc1-491a-4b10-a7c3-c4e71dfa7391 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1817.060072] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62914d18-d840-4b4b-8fd9-fd045f41eed9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1817.067847] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96ba5c7b-3773-4d95-ad03-103b6a73a2b0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1817.081784] env[62824]: DEBUG nova.compute.provider_tree [None req-24599d6d-2f76-445f-8b64-070ec4ee0da9 tempest-TenantUsagesTestJSON-486947106 tempest-TenantUsagesTestJSON-486947106-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1817.585391] env[62824]: DEBUG nova.scheduler.client.report [None req-24599d6d-2f76-445f-8b64-070ec4ee0da9 tempest-TenantUsagesTestJSON-486947106 tempest-TenantUsagesTestJSON-486947106-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1817.879966] env[62824]: DEBUG nova.compute.manager [None req-b73330de-7630-444b-9674-b1882b4c4da9 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1817.914565] env[62824]: DEBUG nova.virt.hardware [None req-b73330de-7630-444b-9674-b1882b4c4da9 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1817.914885] env[62824]: DEBUG nova.virt.hardware [None req-b73330de-7630-444b-9674-b1882b4c4da9 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1817.914950] env[62824]: DEBUG nova.virt.hardware [None req-b73330de-7630-444b-9674-b1882b4c4da9 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1817.915417] env[62824]: DEBUG nova.virt.hardware [None req-b73330de-7630-444b-9674-b1882b4c4da9 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1817.915417] env[62824]: DEBUG nova.virt.hardware [None req-b73330de-7630-444b-9674-b1882b4c4da9 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1817.915527] env[62824]: DEBUG nova.virt.hardware [None req-b73330de-7630-444b-9674-b1882b4c4da9 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1817.915663] env[62824]: DEBUG nova.virt.hardware [None req-b73330de-7630-444b-9674-b1882b4c4da9 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1817.915816] env[62824]: DEBUG nova.virt.hardware [None req-b73330de-7630-444b-9674-b1882b4c4da9 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1817.915982] env[62824]: DEBUG nova.virt.hardware [None req-b73330de-7630-444b-9674-b1882b4c4da9 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1817.916798] env[62824]: DEBUG nova.virt.hardware [None req-b73330de-7630-444b-9674-b1882b4c4da9 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1817.917026] env[62824]: DEBUG nova.virt.hardware [None req-b73330de-7630-444b-9674-b1882b4c4da9 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1817.917902] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e459d3de-7326-4743-b9cd-bd7f2a222724 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1817.926548] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6105c32e-636c-45ae-8d6a-57be83f01aa1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1818.090194] env[62824]: DEBUG oslo_concurrency.lockutils [None req-24599d6d-2f76-445f-8b64-070ec4ee0da9 tempest-TenantUsagesTestJSON-486947106 tempest-TenantUsagesTestJSON-486947106-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.242s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1818.090732] env[62824]: DEBUG nova.compute.manager [None req-24599d6d-2f76-445f-8b64-070ec4ee0da9 tempest-TenantUsagesTestJSON-486947106 tempest-TenantUsagesTestJSON-486947106-project-member] [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1818.095137] env[62824]: DEBUG oslo_concurrency.lockutils [None req-6070f146-69b5-4aa1-b292-f106b5e9564b tempest-ServerAddressesNegativeTestJSON-1575558569 tempest-ServerAddressesNegativeTestJSON-1575558569-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.865s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1818.097796] env[62824]: INFO nova.compute.claims [None req-6070f146-69b5-4aa1-b292-f106b5e9564b tempest-ServerAddressesNegativeTestJSON-1575558569 tempest-ServerAddressesNegativeTestJSON-1575558569-project-member] [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1818.444594] env[62824]: DEBUG nova.network.neutron [None req-b73330de-7630-444b-9674-b1882b4c4da9 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] Successfully created port: 3af5eeb4-c5d9-469e-9f53-3b9cc122490c {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1818.603562] env[62824]: DEBUG nova.compute.utils [None req-24599d6d-2f76-445f-8b64-070ec4ee0da9 tempest-TenantUsagesTestJSON-486947106 tempest-TenantUsagesTestJSON-486947106-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1818.608304] env[62824]: DEBUG nova.compute.manager [None req-24599d6d-2f76-445f-8b64-070ec4ee0da9 tempest-TenantUsagesTestJSON-486947106 tempest-TenantUsagesTestJSON-486947106-project-member] [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1818.614023] env[62824]: DEBUG nova.network.neutron [None req-24599d6d-2f76-445f-8b64-070ec4ee0da9 tempest-TenantUsagesTestJSON-486947106 tempest-TenantUsagesTestJSON-486947106-project-member] [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1818.927907] env[62824]: DEBUG nova.policy [None req-24599d6d-2f76-445f-8b64-070ec4ee0da9 tempest-TenantUsagesTestJSON-486947106 tempest-TenantUsagesTestJSON-486947106-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd0f6e6e8a588403ca4cd7462cae2b5d8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9524741a16e846b4afb45ba0707ad00a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 1819.110535] env[62824]: DEBUG nova.compute.manager [None req-24599d6d-2f76-445f-8b64-070ec4ee0da9 tempest-TenantUsagesTestJSON-486947106 tempest-TenantUsagesTestJSON-486947106-project-member] [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1819.393626] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a85fd1c-b07f-4096-8eaa-96f72a8bb55f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1819.409288] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26867f30-3c1a-4475-b5ea-703553a0bfce {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1819.449217] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98fd080e-e573-456b-abf3-602d509ab99b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1819.459157] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-443080fe-fd36-4f18-a088-58290f1e21dd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1819.481345] env[62824]: DEBUG nova.compute.provider_tree [None req-6070f146-69b5-4aa1-b292-f106b5e9564b tempest-ServerAddressesNegativeTestJSON-1575558569 tempest-ServerAddressesNegativeTestJSON-1575558569-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1819.986334] env[62824]: DEBUG nova.scheduler.client.report [None req-6070f146-69b5-4aa1-b292-f106b5e9564b tempest-ServerAddressesNegativeTestJSON-1575558569 tempest-ServerAddressesNegativeTestJSON-1575558569-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1820.124594] env[62824]: DEBUG nova.compute.manager [None req-24599d6d-2f76-445f-8b64-070ec4ee0da9 tempest-TenantUsagesTestJSON-486947106 tempest-TenantUsagesTestJSON-486947106-project-member] [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1820.167232] env[62824]: DEBUG nova.virt.hardware [None req-24599d6d-2f76-445f-8b64-070ec4ee0da9 tempest-TenantUsagesTestJSON-486947106 tempest-TenantUsagesTestJSON-486947106-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1820.167398] env[62824]: DEBUG nova.virt.hardware [None req-24599d6d-2f76-445f-8b64-070ec4ee0da9 tempest-TenantUsagesTestJSON-486947106 tempest-TenantUsagesTestJSON-486947106-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1820.167971] env[62824]: DEBUG nova.virt.hardware [None req-24599d6d-2f76-445f-8b64-070ec4ee0da9 tempest-TenantUsagesTestJSON-486947106 tempest-TenantUsagesTestJSON-486947106-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1820.167971] env[62824]: DEBUG nova.virt.hardware [None req-24599d6d-2f76-445f-8b64-070ec4ee0da9 tempest-TenantUsagesTestJSON-486947106 tempest-TenantUsagesTestJSON-486947106-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1820.167971] env[62824]: DEBUG nova.virt.hardware [None req-24599d6d-2f76-445f-8b64-070ec4ee0da9 tempest-TenantUsagesTestJSON-486947106 tempest-TenantUsagesTestJSON-486947106-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1820.167971] env[62824]: DEBUG nova.virt.hardware [None req-24599d6d-2f76-445f-8b64-070ec4ee0da9 tempest-TenantUsagesTestJSON-486947106 tempest-TenantUsagesTestJSON-486947106-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1820.168632] env[62824]: DEBUG nova.virt.hardware [None req-24599d6d-2f76-445f-8b64-070ec4ee0da9 tempest-TenantUsagesTestJSON-486947106 tempest-TenantUsagesTestJSON-486947106-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1820.168839] env[62824]: DEBUG nova.virt.hardware [None req-24599d6d-2f76-445f-8b64-070ec4ee0da9 tempest-TenantUsagesTestJSON-486947106 tempest-TenantUsagesTestJSON-486947106-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1820.168973] env[62824]: DEBUG nova.virt.hardware [None req-24599d6d-2f76-445f-8b64-070ec4ee0da9 tempest-TenantUsagesTestJSON-486947106 tempest-TenantUsagesTestJSON-486947106-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1820.169145] env[62824]: DEBUG nova.virt.hardware [None req-24599d6d-2f76-445f-8b64-070ec4ee0da9 tempest-TenantUsagesTestJSON-486947106 tempest-TenantUsagesTestJSON-486947106-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1820.169320] env[62824]: DEBUG nova.virt.hardware [None req-24599d6d-2f76-445f-8b64-070ec4ee0da9 tempest-TenantUsagesTestJSON-486947106 tempest-TenantUsagesTestJSON-486947106-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1820.170222] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad6bdb2b-56d2-4c1a-a35d-d3d76553cdca {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1820.182527] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30606fc2-b141-4d9c-918c-601f265e9db7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1820.495144] env[62824]: DEBUG oslo_concurrency.lockutils [None req-6070f146-69b5-4aa1-b292-f106b5e9564b tempest-ServerAddressesNegativeTestJSON-1575558569 tempest-ServerAddressesNegativeTestJSON-1575558569-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.400s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1820.495720] env[62824]: DEBUG nova.compute.manager [None req-6070f146-69b5-4aa1-b292-f106b5e9564b tempest-ServerAddressesNegativeTestJSON-1575558569 tempest-ServerAddressesNegativeTestJSON-1575558569-project-member] [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1820.499485] env[62824]: DEBUG oslo_concurrency.lockutils [None req-250af368-8463-49f4-8a2e-de088db419fa tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.556s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1820.500900] env[62824]: INFO nova.compute.claims [None req-250af368-8463-49f4-8a2e-de088db419fa tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1820.936942] env[62824]: DEBUG nova.network.neutron [None req-24599d6d-2f76-445f-8b64-070ec4ee0da9 tempest-TenantUsagesTestJSON-486947106 tempest-TenantUsagesTestJSON-486947106-project-member] [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] Successfully created port: b42c0eea-348b-430a-8119-0a16fac3dc7f {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1821.001158] env[62824]: DEBUG nova.compute.utils [None req-6070f146-69b5-4aa1-b292-f106b5e9564b tempest-ServerAddressesNegativeTestJSON-1575558569 tempest-ServerAddressesNegativeTestJSON-1575558569-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1821.004457] env[62824]: DEBUG nova.compute.manager [None req-6070f146-69b5-4aa1-b292-f106b5e9564b tempest-ServerAddressesNegativeTestJSON-1575558569 tempest-ServerAddressesNegativeTestJSON-1575558569-project-member] [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1821.005254] env[62824]: DEBUG nova.network.neutron [None req-6070f146-69b5-4aa1-b292-f106b5e9564b tempest-ServerAddressesNegativeTestJSON-1575558569 tempest-ServerAddressesNegativeTestJSON-1575558569-project-member] [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1821.262385] env[62824]: DEBUG nova.policy [None req-6070f146-69b5-4aa1-b292-f106b5e9564b tempest-ServerAddressesNegativeTestJSON-1575558569 tempest-ServerAddressesNegativeTestJSON-1575558569-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'efd2032f2837441a906f8d7eef5fbc68', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd6d073bbdd424c22817ef13d3a7dbb25', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 1821.509355] env[62824]: DEBUG nova.compute.manager [None req-6070f146-69b5-4aa1-b292-f106b5e9564b tempest-ServerAddressesNegativeTestJSON-1575558569 tempest-ServerAddressesNegativeTestJSON-1575558569-project-member] [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1821.582867] env[62824]: ERROR nova.compute.manager [None req-84c279f8-4acc-49b0-9f33-b648872f4030 tempest-FloatingIPsAssociationTestJSON-1164183782 tempest-FloatingIPsAssociationTestJSON-1164183782-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5f952c6c-084a-4768-97e4-16b121390652, please check neutron logs for more information. [ 1821.582867] env[62824]: ERROR nova.compute.manager Traceback (most recent call last): [ 1821.582867] env[62824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1996, in _allocate_network_async [ 1821.582867] env[62824]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 1821.582867] env[62824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 1821.582867] env[62824]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 1821.582867] env[62824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 1821.582867] env[62824]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 1821.582867] env[62824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1821.582867] env[62824]: ERROR nova.compute.manager self.force_reraise() [ 1821.582867] env[62824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1821.582867] env[62824]: ERROR nova.compute.manager raise self.value [ 1821.582867] env[62824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 1821.582867] env[62824]: ERROR nova.compute.manager updated_port = self._update_port( [ 1821.582867] env[62824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 1821.582867] env[62824]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 1821.583354] env[62824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 1821.583354] env[62824]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 1821.583354] env[62824]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5f952c6c-084a-4768-97e4-16b121390652, please check neutron logs for more information. [ 1821.583354] env[62824]: ERROR nova.compute.manager [ 1821.585639] env[62824]: Traceback (most recent call last): [ 1821.585719] env[62824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 1821.585719] env[62824]: listener.cb(fileno) [ 1821.585719] env[62824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 1821.585719] env[62824]: result = function(*args, **kwargs) [ 1821.585719] env[62824]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 1821.585719] env[62824]: return func(*args, **kwargs) [ 1821.585719] env[62824]: File "/opt/stack/nova/nova/compute/manager.py", line 2018, in _allocate_network_async [ 1821.585719] env[62824]: raise e [ 1821.585719] env[62824]: File "/opt/stack/nova/nova/compute/manager.py", line 1996, in _allocate_network_async [ 1821.585719] env[62824]: nwinfo = self.network_api.allocate_for_instance( [ 1821.585719] env[62824]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 1821.585719] env[62824]: created_port_ids = self._update_ports_for_instance( [ 1821.585719] env[62824]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 1821.585719] env[62824]: with excutils.save_and_reraise_exception(): [ 1821.585719] env[62824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1821.585719] env[62824]: self.force_reraise() [ 1821.585719] env[62824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1821.585719] env[62824]: raise self.value [ 1821.585719] env[62824]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 1821.585719] env[62824]: updated_port = self._update_port( [ 1821.585719] env[62824]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 1821.585719] env[62824]: _ensure_no_port_binding_failure(port) [ 1821.585719] env[62824]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 1821.585719] env[62824]: raise exception.PortBindingFailed(port_id=port['id']) [ 1821.585719] env[62824]: nova.exception.PortBindingFailed: Binding failed for port 5f952c6c-084a-4768-97e4-16b121390652, please check neutron logs for more information. [ 1821.585719] env[62824]: Removing descriptor: 14 [ 1821.588550] env[62824]: ERROR nova.compute.manager [None req-84c279f8-4acc-49b0-9f33-b648872f4030 tempest-FloatingIPsAssociationTestJSON-1164183782 tempest-FloatingIPsAssociationTestJSON-1164183782-project-member] [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5f952c6c-084a-4768-97e4-16b121390652, please check neutron logs for more information. [ 1821.588550] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] Traceback (most recent call last): [ 1821.588550] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] File "/opt/stack/nova/nova/compute/manager.py", line 2901, in _build_resources [ 1821.588550] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] yield resources [ 1821.588550] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] File "/opt/stack/nova/nova/compute/manager.py", line 2648, in _build_and_run_instance [ 1821.588550] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] self.driver.spawn(context, instance, image_meta, [ 1821.588550] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 1821.588550] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] self._vmops.spawn(context, instance, image_meta, injected_files, [ 1821.588550] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 1821.588550] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] vm_ref = self.build_virtual_machine(instance, [ 1821.588550] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 1821.588982] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] vif_infos = vmwarevif.get_vif_info(self._session, [ 1821.588982] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 1821.588982] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] for vif in network_info: [ 1821.588982] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 1821.588982] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] return self._sync_wrapper(fn, *args, **kwargs) [ 1821.588982] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 1821.588982] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] self.wait() [ 1821.588982] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 1821.588982] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] self[:] = self._gt.wait() [ 1821.588982] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 1821.588982] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] return self._exit_event.wait() [ 1821.588982] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 1821.588982] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] result = hub.switch() [ 1821.589426] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 1821.589426] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] return self.greenlet.switch() [ 1821.589426] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 1821.589426] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] result = function(*args, **kwargs) [ 1821.589426] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 1821.589426] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] return func(*args, **kwargs) [ 1821.589426] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] File "/opt/stack/nova/nova/compute/manager.py", line 2018, in _allocate_network_async [ 1821.589426] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] raise e [ 1821.589426] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] File "/opt/stack/nova/nova/compute/manager.py", line 1996, in _allocate_network_async [ 1821.589426] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] nwinfo = self.network_api.allocate_for_instance( [ 1821.589426] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 1821.589426] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] created_port_ids = self._update_ports_for_instance( [ 1821.589426] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 1821.589794] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] with excutils.save_and_reraise_exception(): [ 1821.589794] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1821.589794] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] self.force_reraise() [ 1821.589794] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1821.589794] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] raise self.value [ 1821.589794] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 1821.589794] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] updated_port = self._update_port( [ 1821.589794] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 1821.589794] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] _ensure_no_port_binding_failure(port) [ 1821.589794] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 1821.589794] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] raise exception.PortBindingFailed(port_id=port['id']) [ 1821.589794] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] nova.exception.PortBindingFailed: Binding failed for port 5f952c6c-084a-4768-97e4-16b121390652, please check neutron logs for more information. [ 1821.589794] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] [ 1821.590199] env[62824]: INFO nova.compute.manager [None req-84c279f8-4acc-49b0-9f33-b648872f4030 tempest-FloatingIPsAssociationTestJSON-1164183782 tempest-FloatingIPsAssociationTestJSON-1164183782-project-member] [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] Terminating instance [ 1821.621697] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30a0543d-8512-441b-9b98-4924016eb951 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1821.631885] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02000423-6895-4f46-9aa3-4c9ed5444de0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1821.672276] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a8f4047-6c83-4461-8ccd-33743d4e5b21 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1821.681952] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76ba07f3-ef05-4359-904b-647234055dce {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1821.698184] env[62824]: DEBUG nova.compute.provider_tree [None req-250af368-8463-49f4-8a2e-de088db419fa tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1822.095759] env[62824]: DEBUG oslo_concurrency.lockutils [None req-84c279f8-4acc-49b0-9f33-b648872f4030 tempest-FloatingIPsAssociationTestJSON-1164183782 tempest-FloatingIPsAssociationTestJSON-1164183782-project-member] Acquiring lock "refresh_cache-764d7b4f-49d3-4ddf-8152-c08721dbce59" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1822.096043] env[62824]: DEBUG oslo_concurrency.lockutils [None req-84c279f8-4acc-49b0-9f33-b648872f4030 tempest-FloatingIPsAssociationTestJSON-1164183782 tempest-FloatingIPsAssociationTestJSON-1164183782-project-member] Acquired lock "refresh_cache-764d7b4f-49d3-4ddf-8152-c08721dbce59" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1822.096210] env[62824]: DEBUG nova.network.neutron [None req-84c279f8-4acc-49b0-9f33-b648872f4030 tempest-FloatingIPsAssociationTestJSON-1164183782 tempest-FloatingIPsAssociationTestJSON-1164183782-project-member] [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1822.204736] env[62824]: DEBUG nova.scheduler.client.report [None req-250af368-8463-49f4-8a2e-de088db419fa tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1822.523164] env[62824]: DEBUG nova.compute.manager [None req-6070f146-69b5-4aa1-b292-f106b5e9564b tempest-ServerAddressesNegativeTestJSON-1575558569 tempest-ServerAddressesNegativeTestJSON-1575558569-project-member] [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1822.554775] env[62824]: DEBUG nova.virt.hardware [None req-6070f146-69b5-4aa1-b292-f106b5e9564b tempest-ServerAddressesNegativeTestJSON-1575558569 tempest-ServerAddressesNegativeTestJSON-1575558569-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1822.555083] env[62824]: DEBUG nova.virt.hardware [None req-6070f146-69b5-4aa1-b292-f106b5e9564b tempest-ServerAddressesNegativeTestJSON-1575558569 tempest-ServerAddressesNegativeTestJSON-1575558569-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1822.555288] env[62824]: DEBUG nova.virt.hardware [None req-6070f146-69b5-4aa1-b292-f106b5e9564b tempest-ServerAddressesNegativeTestJSON-1575558569 tempest-ServerAddressesNegativeTestJSON-1575558569-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1822.555515] env[62824]: DEBUG nova.virt.hardware [None req-6070f146-69b5-4aa1-b292-f106b5e9564b tempest-ServerAddressesNegativeTestJSON-1575558569 tempest-ServerAddressesNegativeTestJSON-1575558569-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1822.555701] env[62824]: DEBUG nova.virt.hardware [None req-6070f146-69b5-4aa1-b292-f106b5e9564b tempest-ServerAddressesNegativeTestJSON-1575558569 tempest-ServerAddressesNegativeTestJSON-1575558569-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1822.555881] env[62824]: DEBUG nova.virt.hardware [None req-6070f146-69b5-4aa1-b292-f106b5e9564b tempest-ServerAddressesNegativeTestJSON-1575558569 tempest-ServerAddressesNegativeTestJSON-1575558569-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1822.559455] env[62824]: DEBUG nova.virt.hardware [None req-6070f146-69b5-4aa1-b292-f106b5e9564b tempest-ServerAddressesNegativeTestJSON-1575558569 tempest-ServerAddressesNegativeTestJSON-1575558569-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1822.559691] env[62824]: DEBUG nova.virt.hardware [None req-6070f146-69b5-4aa1-b292-f106b5e9564b tempest-ServerAddressesNegativeTestJSON-1575558569 tempest-ServerAddressesNegativeTestJSON-1575558569-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1822.559914] env[62824]: DEBUG nova.virt.hardware [None req-6070f146-69b5-4aa1-b292-f106b5e9564b tempest-ServerAddressesNegativeTestJSON-1575558569 tempest-ServerAddressesNegativeTestJSON-1575558569-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1822.560137] env[62824]: DEBUG nova.virt.hardware [None req-6070f146-69b5-4aa1-b292-f106b5e9564b tempest-ServerAddressesNegativeTestJSON-1575558569 tempest-ServerAddressesNegativeTestJSON-1575558569-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1822.560352] env[62824]: DEBUG nova.virt.hardware [None req-6070f146-69b5-4aa1-b292-f106b5e9564b tempest-ServerAddressesNegativeTestJSON-1575558569 tempest-ServerAddressesNegativeTestJSON-1575558569-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1822.562422] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea68ca25-1bf4-448d-9f45-73ba3c94135a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1822.573383] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c28755e9-eafa-4f99-aefe-744ade7c6c48 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1822.654171] env[62824]: DEBUG nova.network.neutron [None req-84c279f8-4acc-49b0-9f33-b648872f4030 tempest-FloatingIPsAssociationTestJSON-1164183782 tempest-FloatingIPsAssociationTestJSON-1164183782-project-member] [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1822.710247] env[62824]: DEBUG oslo_concurrency.lockutils [None req-250af368-8463-49f4-8a2e-de088db419fa tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.210s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1822.711276] env[62824]: DEBUG nova.compute.manager [None req-250af368-8463-49f4-8a2e-de088db419fa tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1822.952574] env[62824]: DEBUG nova.network.neutron [None req-6070f146-69b5-4aa1-b292-f106b5e9564b tempest-ServerAddressesNegativeTestJSON-1575558569 tempest-ServerAddressesNegativeTestJSON-1575558569-project-member] [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] Successfully created port: 59ec24da-a159-4775-a06b-bf6a02869bed {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1822.963659] env[62824]: DEBUG nova.network.neutron [None req-84c279f8-4acc-49b0-9f33-b648872f4030 tempest-FloatingIPsAssociationTestJSON-1164183782 tempest-FloatingIPsAssociationTestJSON-1164183782-project-member] [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1823.137149] env[62824]: ERROR nova.compute.manager [None req-b73330de-7630-444b-9674-b1882b4c4da9 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3af5eeb4-c5d9-469e-9f53-3b9cc122490c, please check neutron logs for more information. [ 1823.137149] env[62824]: ERROR nova.compute.manager Traceback (most recent call last): [ 1823.137149] env[62824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1996, in _allocate_network_async [ 1823.137149] env[62824]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 1823.137149] env[62824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 1823.137149] env[62824]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 1823.137149] env[62824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 1823.137149] env[62824]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 1823.137149] env[62824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1823.137149] env[62824]: ERROR nova.compute.manager self.force_reraise() [ 1823.137149] env[62824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1823.137149] env[62824]: ERROR nova.compute.manager raise self.value [ 1823.137149] env[62824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 1823.137149] env[62824]: ERROR nova.compute.manager updated_port = self._update_port( [ 1823.137149] env[62824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 1823.137149] env[62824]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 1823.138234] env[62824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 1823.138234] env[62824]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 1823.138234] env[62824]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3af5eeb4-c5d9-469e-9f53-3b9cc122490c, please check neutron logs for more information. [ 1823.138234] env[62824]: ERROR nova.compute.manager [ 1823.138234] env[62824]: Traceback (most recent call last): [ 1823.138234] env[62824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 1823.138234] env[62824]: listener.cb(fileno) [ 1823.138234] env[62824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 1823.138234] env[62824]: result = function(*args, **kwargs) [ 1823.138234] env[62824]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 1823.138234] env[62824]: return func(*args, **kwargs) [ 1823.138234] env[62824]: File "/opt/stack/nova/nova/compute/manager.py", line 2018, in _allocate_network_async [ 1823.138234] env[62824]: raise e [ 1823.138234] env[62824]: File "/opt/stack/nova/nova/compute/manager.py", line 1996, in _allocate_network_async [ 1823.138234] env[62824]: nwinfo = self.network_api.allocate_for_instance( [ 1823.138234] env[62824]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 1823.138234] env[62824]: created_port_ids = self._update_ports_for_instance( [ 1823.138234] env[62824]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 1823.138234] env[62824]: with excutils.save_and_reraise_exception(): [ 1823.138234] env[62824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1823.138234] env[62824]: self.force_reraise() [ 1823.138234] env[62824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1823.138234] env[62824]: raise self.value [ 1823.138234] env[62824]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 1823.138234] env[62824]: updated_port = self._update_port( [ 1823.138234] env[62824]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 1823.138234] env[62824]: _ensure_no_port_binding_failure(port) [ 1823.138234] env[62824]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 1823.138234] env[62824]: raise exception.PortBindingFailed(port_id=port['id']) [ 1823.139477] env[62824]: nova.exception.PortBindingFailed: Binding failed for port 3af5eeb4-c5d9-469e-9f53-3b9cc122490c, please check neutron logs for more information. [ 1823.139477] env[62824]: Removing descriptor: 15 [ 1823.139477] env[62824]: ERROR nova.compute.manager [None req-b73330de-7630-444b-9674-b1882b4c4da9 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3af5eeb4-c5d9-469e-9f53-3b9cc122490c, please check neutron logs for more information. [ 1823.139477] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] Traceback (most recent call last): [ 1823.139477] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] File "/opt/stack/nova/nova/compute/manager.py", line 2901, in _build_resources [ 1823.139477] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] yield resources [ 1823.139477] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] File "/opt/stack/nova/nova/compute/manager.py", line 2648, in _build_and_run_instance [ 1823.139477] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] self.driver.spawn(context, instance, image_meta, [ 1823.139477] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 1823.139477] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] self._vmops.spawn(context, instance, image_meta, injected_files, [ 1823.139477] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 1823.139477] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] vm_ref = self.build_virtual_machine(instance, [ 1823.140145] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 1823.140145] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] vif_infos = vmwarevif.get_vif_info(self._session, [ 1823.140145] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 1823.140145] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] for vif in network_info: [ 1823.140145] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 1823.140145] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] return self._sync_wrapper(fn, *args, **kwargs) [ 1823.140145] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 1823.140145] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] self.wait() [ 1823.140145] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 1823.140145] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] self[:] = self._gt.wait() [ 1823.140145] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 1823.140145] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] return self._exit_event.wait() [ 1823.140145] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 1823.140759] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] result = hub.switch() [ 1823.140759] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 1823.140759] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] return self.greenlet.switch() [ 1823.140759] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 1823.140759] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] result = function(*args, **kwargs) [ 1823.140759] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 1823.140759] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] return func(*args, **kwargs) [ 1823.140759] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] File "/opt/stack/nova/nova/compute/manager.py", line 2018, in _allocate_network_async [ 1823.140759] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] raise e [ 1823.140759] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] File "/opt/stack/nova/nova/compute/manager.py", line 1996, in _allocate_network_async [ 1823.140759] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] nwinfo = self.network_api.allocate_for_instance( [ 1823.140759] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 1823.140759] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] created_port_ids = self._update_ports_for_instance( [ 1823.141131] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 1823.141131] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] with excutils.save_and_reraise_exception(): [ 1823.141131] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1823.141131] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] self.force_reraise() [ 1823.141131] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1823.141131] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] raise self.value [ 1823.141131] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 1823.141131] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] updated_port = self._update_port( [ 1823.141131] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 1823.141131] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] _ensure_no_port_binding_failure(port) [ 1823.141131] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 1823.141131] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] raise exception.PortBindingFailed(port_id=port['id']) [ 1823.141486] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] nova.exception.PortBindingFailed: Binding failed for port 3af5eeb4-c5d9-469e-9f53-3b9cc122490c, please check neutron logs for more information. [ 1823.141486] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] [ 1823.141486] env[62824]: INFO nova.compute.manager [None req-b73330de-7630-444b-9674-b1882b4c4da9 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] Terminating instance [ 1823.217401] env[62824]: DEBUG nova.compute.utils [None req-250af368-8463-49f4-8a2e-de088db419fa tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1823.222156] env[62824]: DEBUG nova.compute.manager [None req-250af368-8463-49f4-8a2e-de088db419fa tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1823.222441] env[62824]: DEBUG nova.network.neutron [None req-250af368-8463-49f4-8a2e-de088db419fa tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1823.466657] env[62824]: DEBUG oslo_concurrency.lockutils [None req-84c279f8-4acc-49b0-9f33-b648872f4030 tempest-FloatingIPsAssociationTestJSON-1164183782 tempest-FloatingIPsAssociationTestJSON-1164183782-project-member] Releasing lock "refresh_cache-764d7b4f-49d3-4ddf-8152-c08721dbce59" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1823.467383] env[62824]: DEBUG nova.compute.manager [None req-84c279f8-4acc-49b0-9f33-b648872f4030 tempest-FloatingIPsAssociationTestJSON-1164183782 tempest-FloatingIPsAssociationTestJSON-1164183782-project-member] [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1823.467792] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-84c279f8-4acc-49b0-9f33-b648872f4030 tempest-FloatingIPsAssociationTestJSON-1164183782 tempest-FloatingIPsAssociationTestJSON-1164183782-project-member] [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1823.467914] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1787b4d5-a0ed-4ac8-bf61-f31c360ab23d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1823.480022] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82dcb7e6-5ed7-4f65-81f6-1eceea82825b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1823.492623] env[62824]: DEBUG nova.policy [None req-250af368-8463-49f4-8a2e-de088db419fa tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2cbf8eb928484d54a5371365386e73f6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b2f9084931bc4b94b0ae85e3e4b0ba90', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 1823.506347] env[62824]: WARNING nova.virt.vmwareapi.vmops [None req-84c279f8-4acc-49b0-9f33-b648872f4030 tempest-FloatingIPsAssociationTestJSON-1164183782 tempest-FloatingIPsAssociationTestJSON-1164183782-project-member] [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 764d7b4f-49d3-4ddf-8152-c08721dbce59 could not be found. [ 1823.506588] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-84c279f8-4acc-49b0-9f33-b648872f4030 tempest-FloatingIPsAssociationTestJSON-1164183782 tempest-FloatingIPsAssociationTestJSON-1164183782-project-member] [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1823.506974] env[62824]: INFO nova.compute.manager [None req-84c279f8-4acc-49b0-9f33-b648872f4030 tempest-FloatingIPsAssociationTestJSON-1164183782 tempest-FloatingIPsAssociationTestJSON-1164183782-project-member] [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] Took 0.04 seconds to destroy the instance on the hypervisor. [ 1823.507243] env[62824]: DEBUG oslo.service.loopingcall [None req-84c279f8-4acc-49b0-9f33-b648872f4030 tempest-FloatingIPsAssociationTestJSON-1164183782 tempest-FloatingIPsAssociationTestJSON-1164183782-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1823.507465] env[62824]: DEBUG nova.compute.manager [-] [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1823.507553] env[62824]: DEBUG nova.network.neutron [-] [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1823.565032] env[62824]: DEBUG nova.network.neutron [-] [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1823.643534] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b73330de-7630-444b-9674-b1882b4c4da9 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Acquiring lock "refresh_cache-579e1ca1-a6b6-4fde-8368-3e635a303032" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1823.643724] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b73330de-7630-444b-9674-b1882b4c4da9 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Acquired lock "refresh_cache-579e1ca1-a6b6-4fde-8368-3e635a303032" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1823.643961] env[62824]: DEBUG nova.network.neutron [None req-b73330de-7630-444b-9674-b1882b4c4da9 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1823.726356] env[62824]: DEBUG nova.compute.manager [None req-250af368-8463-49f4-8a2e-de088db419fa tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1824.069433] env[62824]: DEBUG nova.network.neutron [-] [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1824.197711] env[62824]: DEBUG nova.network.neutron [None req-b73330de-7630-444b-9674-b1882b4c4da9 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1824.359772] env[62824]: DEBUG nova.network.neutron [None req-b73330de-7630-444b-9674-b1882b4c4da9 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1824.482284] env[62824]: DEBUG nova.network.neutron [None req-250af368-8463-49f4-8a2e-de088db419fa tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] Successfully created port: 2a8b04db-a377-45bb-9e8f-39327bbfffce {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1824.574497] env[62824]: INFO nova.compute.manager [-] [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] Took 1.07 seconds to deallocate network for instance. [ 1824.580799] env[62824]: DEBUG nova.compute.claims [None req-84c279f8-4acc-49b0-9f33-b648872f4030 tempest-FloatingIPsAssociationTestJSON-1164183782 tempest-FloatingIPsAssociationTestJSON-1164183782-project-member] [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] Aborting claim: {{(pid=62824) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 1824.580998] env[62824]: DEBUG oslo_concurrency.lockutils [None req-84c279f8-4acc-49b0-9f33-b648872f4030 tempest-FloatingIPsAssociationTestJSON-1164183782 tempest-FloatingIPsAssociationTestJSON-1164183782-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1824.581234] env[62824]: DEBUG oslo_concurrency.lockutils [None req-84c279f8-4acc-49b0-9f33-b648872f4030 tempest-FloatingIPsAssociationTestJSON-1164183782 tempest-FloatingIPsAssociationTestJSON-1164183782-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1824.740457] env[62824]: DEBUG nova.compute.manager [None req-250af368-8463-49f4-8a2e-de088db419fa tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1824.781328] env[62824]: DEBUG nova.virt.hardware [None req-250af368-8463-49f4-8a2e-de088db419fa tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1824.781328] env[62824]: DEBUG nova.virt.hardware [None req-250af368-8463-49f4-8a2e-de088db419fa tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1824.781493] env[62824]: DEBUG nova.virt.hardware [None req-250af368-8463-49f4-8a2e-de088db419fa tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1824.781548] env[62824]: DEBUG nova.virt.hardware [None req-250af368-8463-49f4-8a2e-de088db419fa tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1824.781765] env[62824]: DEBUG nova.virt.hardware [None req-250af368-8463-49f4-8a2e-de088db419fa tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1824.781983] env[62824]: DEBUG nova.virt.hardware [None req-250af368-8463-49f4-8a2e-de088db419fa tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1824.782087] env[62824]: DEBUG nova.virt.hardware [None req-250af368-8463-49f4-8a2e-de088db419fa tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1824.782253] env[62824]: DEBUG nova.virt.hardware [None req-250af368-8463-49f4-8a2e-de088db419fa tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1824.782414] env[62824]: DEBUG nova.virt.hardware [None req-250af368-8463-49f4-8a2e-de088db419fa tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1824.782573] env[62824]: DEBUG nova.virt.hardware [None req-250af368-8463-49f4-8a2e-de088db419fa tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1824.782849] env[62824]: DEBUG nova.virt.hardware [None req-250af368-8463-49f4-8a2e-de088db419fa tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1824.783747] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed80bf48-1446-44c0-9ffb-04a60bb7c983 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1824.793940] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57d71584-67fa-4499-9b8a-1351b4e89e30 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1824.841904] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2559566b-844a-4d61-b989-1f0a741aecd2 tempest-ServerDiagnosticsTest-1769488740 tempest-ServerDiagnosticsTest-1769488740-project-member] Acquiring lock "042c5f0f-bf1f-417e-be05-964cfcb20c7e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1824.842355] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2559566b-844a-4d61-b989-1f0a741aecd2 tempest-ServerDiagnosticsTest-1769488740 tempest-ServerDiagnosticsTest-1769488740-project-member] Lock "042c5f0f-bf1f-417e-be05-964cfcb20c7e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1824.863678] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b73330de-7630-444b-9674-b1882b4c4da9 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Releasing lock "refresh_cache-579e1ca1-a6b6-4fde-8368-3e635a303032" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1824.863863] env[62824]: DEBUG nova.compute.manager [None req-b73330de-7630-444b-9674-b1882b4c4da9 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1824.864034] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-b73330de-7630-444b-9674-b1882b4c4da9 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1824.865490] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ca5ad738-0a83-450e-9c6e-8846dabf4c90 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1824.877893] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bb090cd-933a-4d88-a2f0-d8bacba76730 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1824.907852] env[62824]: WARNING nova.virt.vmwareapi.vmops [None req-b73330de-7630-444b-9674-b1882b4c4da9 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 579e1ca1-a6b6-4fde-8368-3e635a303032 could not be found. [ 1824.907852] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-b73330de-7630-444b-9674-b1882b4c4da9 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1824.907852] env[62824]: INFO nova.compute.manager [None req-b73330de-7630-444b-9674-b1882b4c4da9 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] Took 0.04 seconds to destroy the instance on the hypervisor. [ 1824.907852] env[62824]: DEBUG oslo.service.loopingcall [None req-b73330de-7630-444b-9674-b1882b4c4da9 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1824.908541] env[62824]: DEBUG nova.compute.manager [-] [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1824.908791] env[62824]: DEBUG nova.network.neutron [-] [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1824.943605] env[62824]: DEBUG nova.network.neutron [-] [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1825.221696] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e67a4d41-7931-4eee-bc8a-6999fe1e9695 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1825.230626] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87769353-3105-40bd-b87f-9bd9fea72680 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1825.263996] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5f7dc25-a844-4182-84b4-2fe9fa7f4bc1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1825.272288] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9fa7e4b-f517-4755-875c-4f0d719bac8e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1825.286609] env[62824]: DEBUG nova.compute.provider_tree [None req-84c279f8-4acc-49b0-9f33-b648872f4030 tempest-FloatingIPsAssociationTestJSON-1164183782 tempest-FloatingIPsAssociationTestJSON-1164183782-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1825.345079] env[62824]: DEBUG nova.compute.manager [None req-2559566b-844a-4d61-b989-1f0a741aecd2 tempest-ServerDiagnosticsTest-1769488740 tempest-ServerDiagnosticsTest-1769488740-project-member] [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1825.450423] env[62824]: DEBUG nova.network.neutron [-] [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1825.792068] env[62824]: DEBUG nova.scheduler.client.report [None req-84c279f8-4acc-49b0-9f33-b648872f4030 tempest-FloatingIPsAssociationTestJSON-1164183782 tempest-FloatingIPsAssociationTestJSON-1164183782-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1825.874374] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2559566b-844a-4d61-b989-1f0a741aecd2 tempest-ServerDiagnosticsTest-1769488740 tempest-ServerDiagnosticsTest-1769488740-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1825.953349] env[62824]: INFO nova.compute.manager [-] [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] Took 1.04 seconds to deallocate network for instance. [ 1825.956389] env[62824]: DEBUG nova.compute.claims [None req-b73330de-7630-444b-9674-b1882b4c4da9 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] Aborting claim: {{(pid=62824) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 1825.956857] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b73330de-7630-444b-9674-b1882b4c4da9 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1826.255261] env[62824]: ERROR nova.compute.manager [None req-24599d6d-2f76-445f-8b64-070ec4ee0da9 tempest-TenantUsagesTestJSON-486947106 tempest-TenantUsagesTestJSON-486947106-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b42c0eea-348b-430a-8119-0a16fac3dc7f, please check neutron logs for more information. [ 1826.255261] env[62824]: ERROR nova.compute.manager Traceback (most recent call last): [ 1826.255261] env[62824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1996, in _allocate_network_async [ 1826.255261] env[62824]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 1826.255261] env[62824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 1826.255261] env[62824]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 1826.255261] env[62824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 1826.255261] env[62824]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 1826.255261] env[62824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1826.255261] env[62824]: ERROR nova.compute.manager self.force_reraise() [ 1826.255261] env[62824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1826.255261] env[62824]: ERROR nova.compute.manager raise self.value [ 1826.255261] env[62824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 1826.255261] env[62824]: ERROR nova.compute.manager updated_port = self._update_port( [ 1826.255261] env[62824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 1826.255261] env[62824]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 1826.255902] env[62824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 1826.255902] env[62824]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 1826.255902] env[62824]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b42c0eea-348b-430a-8119-0a16fac3dc7f, please check neutron logs for more information. [ 1826.255902] env[62824]: ERROR nova.compute.manager [ 1826.255902] env[62824]: Traceback (most recent call last): [ 1826.255902] env[62824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 1826.255902] env[62824]: listener.cb(fileno) [ 1826.255902] env[62824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 1826.255902] env[62824]: result = function(*args, **kwargs) [ 1826.255902] env[62824]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 1826.255902] env[62824]: return func(*args, **kwargs) [ 1826.255902] env[62824]: File "/opt/stack/nova/nova/compute/manager.py", line 2018, in _allocate_network_async [ 1826.255902] env[62824]: raise e [ 1826.255902] env[62824]: File "/opt/stack/nova/nova/compute/manager.py", line 1996, in _allocate_network_async [ 1826.255902] env[62824]: nwinfo = self.network_api.allocate_for_instance( [ 1826.255902] env[62824]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 1826.255902] env[62824]: created_port_ids = self._update_ports_for_instance( [ 1826.255902] env[62824]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 1826.255902] env[62824]: with excutils.save_and_reraise_exception(): [ 1826.255902] env[62824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1826.255902] env[62824]: self.force_reraise() [ 1826.255902] env[62824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1826.255902] env[62824]: raise self.value [ 1826.255902] env[62824]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 1826.255902] env[62824]: updated_port = self._update_port( [ 1826.255902] env[62824]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 1826.255902] env[62824]: _ensure_no_port_binding_failure(port) [ 1826.255902] env[62824]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 1826.255902] env[62824]: raise exception.PortBindingFailed(port_id=port['id']) [ 1826.257062] env[62824]: nova.exception.PortBindingFailed: Binding failed for port b42c0eea-348b-430a-8119-0a16fac3dc7f, please check neutron logs for more information. [ 1826.257062] env[62824]: Removing descriptor: 17 [ 1826.257062] env[62824]: ERROR nova.compute.manager [None req-24599d6d-2f76-445f-8b64-070ec4ee0da9 tempest-TenantUsagesTestJSON-486947106 tempest-TenantUsagesTestJSON-486947106-project-member] [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b42c0eea-348b-430a-8119-0a16fac3dc7f, please check neutron logs for more information. [ 1826.257062] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] Traceback (most recent call last): [ 1826.257062] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] File "/opt/stack/nova/nova/compute/manager.py", line 2901, in _build_resources [ 1826.257062] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] yield resources [ 1826.257062] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] File "/opt/stack/nova/nova/compute/manager.py", line 2648, in _build_and_run_instance [ 1826.257062] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] self.driver.spawn(context, instance, image_meta, [ 1826.257062] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 1826.257062] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] self._vmops.spawn(context, instance, image_meta, injected_files, [ 1826.257062] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 1826.257062] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] vm_ref = self.build_virtual_machine(instance, [ 1826.257491] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 1826.257491] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] vif_infos = vmwarevif.get_vif_info(self._session, [ 1826.257491] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 1826.257491] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] for vif in network_info: [ 1826.257491] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 1826.257491] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] return self._sync_wrapper(fn, *args, **kwargs) [ 1826.257491] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 1826.257491] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] self.wait() [ 1826.257491] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 1826.257491] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] self[:] = self._gt.wait() [ 1826.257491] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 1826.257491] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] return self._exit_event.wait() [ 1826.257491] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 1826.257812] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] result = hub.switch() [ 1826.257812] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 1826.257812] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] return self.greenlet.switch() [ 1826.257812] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 1826.257812] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] result = function(*args, **kwargs) [ 1826.257812] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 1826.257812] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] return func(*args, **kwargs) [ 1826.257812] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] File "/opt/stack/nova/nova/compute/manager.py", line 2018, in _allocate_network_async [ 1826.257812] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] raise e [ 1826.257812] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] File "/opt/stack/nova/nova/compute/manager.py", line 1996, in _allocate_network_async [ 1826.257812] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] nwinfo = self.network_api.allocate_for_instance( [ 1826.257812] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 1826.257812] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] created_port_ids = self._update_ports_for_instance( [ 1826.258157] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 1826.258157] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] with excutils.save_and_reraise_exception(): [ 1826.258157] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1826.258157] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] self.force_reraise() [ 1826.258157] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1826.258157] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] raise self.value [ 1826.258157] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 1826.258157] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] updated_port = self._update_port( [ 1826.258157] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 1826.258157] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] _ensure_no_port_binding_failure(port) [ 1826.258157] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 1826.258157] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] raise exception.PortBindingFailed(port_id=port['id']) [ 1826.258485] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] nova.exception.PortBindingFailed: Binding failed for port b42c0eea-348b-430a-8119-0a16fac3dc7f, please check neutron logs for more information. [ 1826.258485] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] [ 1826.258485] env[62824]: INFO nova.compute.manager [None req-24599d6d-2f76-445f-8b64-070ec4ee0da9 tempest-TenantUsagesTestJSON-486947106 tempest-TenantUsagesTestJSON-486947106-project-member] [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] Terminating instance [ 1826.298257] env[62824]: DEBUG oslo_concurrency.lockutils [None req-84c279f8-4acc-49b0-9f33-b648872f4030 tempest-FloatingIPsAssociationTestJSON-1164183782 tempest-FloatingIPsAssociationTestJSON-1164183782-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.717s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1826.299524] env[62824]: ERROR nova.compute.manager [None req-84c279f8-4acc-49b0-9f33-b648872f4030 tempest-FloatingIPsAssociationTestJSON-1164183782 tempest-FloatingIPsAssociationTestJSON-1164183782-project-member] [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5f952c6c-084a-4768-97e4-16b121390652, please check neutron logs for more information. [ 1826.299524] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] Traceback (most recent call last): [ 1826.299524] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] File "/opt/stack/nova/nova/compute/manager.py", line 2648, in _build_and_run_instance [ 1826.299524] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] self.driver.spawn(context, instance, image_meta, [ 1826.299524] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 1826.299524] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] self._vmops.spawn(context, instance, image_meta, injected_files, [ 1826.299524] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 1826.299524] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] vm_ref = self.build_virtual_machine(instance, [ 1826.299524] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 1826.299524] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] vif_infos = vmwarevif.get_vif_info(self._session, [ 1826.299524] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 1826.300341] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] for vif in network_info: [ 1826.300341] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 1826.300341] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] return self._sync_wrapper(fn, *args, **kwargs) [ 1826.300341] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 1826.300341] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] self.wait() [ 1826.300341] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 1826.300341] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] self[:] = self._gt.wait() [ 1826.300341] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 1826.300341] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] return self._exit_event.wait() [ 1826.300341] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 1826.300341] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] result = hub.switch() [ 1826.300341] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 1826.300341] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] return self.greenlet.switch() [ 1826.300771] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 1826.300771] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] result = function(*args, **kwargs) [ 1826.300771] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 1826.300771] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] return func(*args, **kwargs) [ 1826.300771] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] File "/opt/stack/nova/nova/compute/manager.py", line 2018, in _allocate_network_async [ 1826.300771] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] raise e [ 1826.300771] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] File "/opt/stack/nova/nova/compute/manager.py", line 1996, in _allocate_network_async [ 1826.300771] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] nwinfo = self.network_api.allocate_for_instance( [ 1826.300771] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 1826.300771] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] created_port_ids = self._update_ports_for_instance( [ 1826.300771] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 1826.300771] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] with excutils.save_and_reraise_exception(): [ 1826.300771] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1826.301141] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] self.force_reraise() [ 1826.301141] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1826.301141] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] raise self.value [ 1826.301141] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 1826.301141] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] updated_port = self._update_port( [ 1826.301141] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 1826.301141] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] _ensure_no_port_binding_failure(port) [ 1826.301141] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 1826.301141] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] raise exception.PortBindingFailed(port_id=port['id']) [ 1826.301141] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] nova.exception.PortBindingFailed: Binding failed for port 5f952c6c-084a-4768-97e4-16b121390652, please check neutron logs for more information. [ 1826.301141] env[62824]: ERROR nova.compute.manager [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] [ 1826.301409] env[62824]: DEBUG nova.compute.utils [None req-84c279f8-4acc-49b0-9f33-b648872f4030 tempest-FloatingIPsAssociationTestJSON-1164183782 tempest-FloatingIPsAssociationTestJSON-1164183782-project-member] [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] Binding failed for port 5f952c6c-084a-4768-97e4-16b121390652, please check neutron logs for more information. {{(pid=62824) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 1826.301409] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2559566b-844a-4d61-b989-1f0a741aecd2 tempest-ServerDiagnosticsTest-1769488740 tempest-ServerDiagnosticsTest-1769488740-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.427s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1826.303209] env[62824]: INFO nova.compute.claims [None req-2559566b-844a-4d61-b989-1f0a741aecd2 tempest-ServerDiagnosticsTest-1769488740 tempest-ServerDiagnosticsTest-1769488740-project-member] [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1826.309767] env[62824]: DEBUG nova.compute.manager [None req-84c279f8-4acc-49b0-9f33-b648872f4030 tempest-FloatingIPsAssociationTestJSON-1164183782 tempest-FloatingIPsAssociationTestJSON-1164183782-project-member] [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] Build of instance 764d7b4f-49d3-4ddf-8152-c08721dbce59 was re-scheduled: Binding failed for port 5f952c6c-084a-4768-97e4-16b121390652, please check neutron logs for more information. {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2487}} [ 1826.310315] env[62824]: DEBUG nova.compute.manager [None req-84c279f8-4acc-49b0-9f33-b648872f4030 tempest-FloatingIPsAssociationTestJSON-1164183782 tempest-FloatingIPsAssociationTestJSON-1164183782-project-member] [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] Unplugging VIFs for instance {{(pid=62824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3013}} [ 1826.310456] env[62824]: DEBUG oslo_concurrency.lockutils [None req-84c279f8-4acc-49b0-9f33-b648872f4030 tempest-FloatingIPsAssociationTestJSON-1164183782 tempest-FloatingIPsAssociationTestJSON-1164183782-project-member] Acquiring lock "refresh_cache-764d7b4f-49d3-4ddf-8152-c08721dbce59" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1826.310596] env[62824]: DEBUG oslo_concurrency.lockutils [None req-84c279f8-4acc-49b0-9f33-b648872f4030 tempest-FloatingIPsAssociationTestJSON-1164183782 tempest-FloatingIPsAssociationTestJSON-1164183782-project-member] Acquired lock "refresh_cache-764d7b4f-49d3-4ddf-8152-c08721dbce59" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1826.310760] env[62824]: DEBUG nova.network.neutron [None req-84c279f8-4acc-49b0-9f33-b648872f4030 tempest-FloatingIPsAssociationTestJSON-1164183782 tempest-FloatingIPsAssociationTestJSON-1164183782-project-member] [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1826.447551] env[62824]: DEBUG oslo_concurrency.lockutils [None req-61fb5137-5f87-475f-a5d2-0722cddc68b4 tempest-ServersTestJSON-928648138 tempest-ServersTestJSON-928648138-project-member] Acquiring lock "0b845d80-c8b2-4b16-bda2-08fc78fb4337" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1826.447815] env[62824]: DEBUG oslo_concurrency.lockutils [None req-61fb5137-5f87-475f-a5d2-0722cddc68b4 tempest-ServersTestJSON-928648138 tempest-ServersTestJSON-928648138-project-member] Lock "0b845d80-c8b2-4b16-bda2-08fc78fb4337" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1826.762791] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7cd5dacb-b71b-429c-bfda-96302f774c53 tempest-ServersTestFqdnHostnames-528120407 tempest-ServersTestFqdnHostnames-528120407-project-member] Acquiring lock "27e6ff4d-c4f2-41e7-92c2-a0e2445830bf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1826.763267] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7cd5dacb-b71b-429c-bfda-96302f774c53 tempest-ServersTestFqdnHostnames-528120407 tempest-ServersTestFqdnHostnames-528120407-project-member] Lock "27e6ff4d-c4f2-41e7-92c2-a0e2445830bf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1826.765615] env[62824]: DEBUG oslo_concurrency.lockutils [None req-24599d6d-2f76-445f-8b64-070ec4ee0da9 tempest-TenantUsagesTestJSON-486947106 tempest-TenantUsagesTestJSON-486947106-project-member] Acquiring lock "refresh_cache-9f016121-09f1-4da7-bbd0-7769d5104d27" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1826.765821] env[62824]: DEBUG oslo_concurrency.lockutils [None req-24599d6d-2f76-445f-8b64-070ec4ee0da9 tempest-TenantUsagesTestJSON-486947106 tempest-TenantUsagesTestJSON-486947106-project-member] Acquired lock "refresh_cache-9f016121-09f1-4da7-bbd0-7769d5104d27" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1826.766402] env[62824]: DEBUG nova.network.neutron [None req-24599d6d-2f76-445f-8b64-070ec4ee0da9 tempest-TenantUsagesTestJSON-486947106 tempest-TenantUsagesTestJSON-486947106-project-member] [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1826.874063] env[62824]: DEBUG nova.network.neutron [None req-84c279f8-4acc-49b0-9f33-b648872f4030 tempest-FloatingIPsAssociationTestJSON-1164183782 tempest-FloatingIPsAssociationTestJSON-1164183782-project-member] [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1826.916688] env[62824]: ERROR nova.compute.manager [None req-250af368-8463-49f4-8a2e-de088db419fa tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 2a8b04db-a377-45bb-9e8f-39327bbfffce, please check neutron logs for more information. [ 1826.916688] env[62824]: ERROR nova.compute.manager Traceback (most recent call last): [ 1826.916688] env[62824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1996, in _allocate_network_async [ 1826.916688] env[62824]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 1826.916688] env[62824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 1826.916688] env[62824]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 1826.916688] env[62824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 1826.916688] env[62824]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 1826.916688] env[62824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1826.916688] env[62824]: ERROR nova.compute.manager self.force_reraise() [ 1826.916688] env[62824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1826.916688] env[62824]: ERROR nova.compute.manager raise self.value [ 1826.916688] env[62824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 1826.916688] env[62824]: ERROR nova.compute.manager updated_port = self._update_port( [ 1826.916688] env[62824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 1826.916688] env[62824]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 1826.917152] env[62824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 1826.917152] env[62824]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 1826.917152] env[62824]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 2a8b04db-a377-45bb-9e8f-39327bbfffce, please check neutron logs for more information. [ 1826.917152] env[62824]: ERROR nova.compute.manager [ 1826.917152] env[62824]: Traceback (most recent call last): [ 1826.917152] env[62824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 1826.917152] env[62824]: listener.cb(fileno) [ 1826.917152] env[62824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 1826.917152] env[62824]: result = function(*args, **kwargs) [ 1826.917152] env[62824]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 1826.917152] env[62824]: return func(*args, **kwargs) [ 1826.917152] env[62824]: File "/opt/stack/nova/nova/compute/manager.py", line 2018, in _allocate_network_async [ 1826.917152] env[62824]: raise e [ 1826.917152] env[62824]: File "/opt/stack/nova/nova/compute/manager.py", line 1996, in _allocate_network_async [ 1826.917152] env[62824]: nwinfo = self.network_api.allocate_for_instance( [ 1826.917152] env[62824]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 1826.917152] env[62824]: created_port_ids = self._update_ports_for_instance( [ 1826.917152] env[62824]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 1826.917152] env[62824]: with excutils.save_and_reraise_exception(): [ 1826.917152] env[62824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1826.917152] env[62824]: self.force_reraise() [ 1826.917152] env[62824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1826.917152] env[62824]: raise self.value [ 1826.917152] env[62824]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 1826.917152] env[62824]: updated_port = self._update_port( [ 1826.917152] env[62824]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 1826.917152] env[62824]: _ensure_no_port_binding_failure(port) [ 1826.917152] env[62824]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 1826.917152] env[62824]: raise exception.PortBindingFailed(port_id=port['id']) [ 1826.917895] env[62824]: nova.exception.PortBindingFailed: Binding failed for port 2a8b04db-a377-45bb-9e8f-39327bbfffce, please check neutron logs for more information. [ 1826.917895] env[62824]: Removing descriptor: 15 [ 1826.917895] env[62824]: ERROR nova.compute.manager [None req-250af368-8463-49f4-8a2e-de088db419fa tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 2a8b04db-a377-45bb-9e8f-39327bbfffce, please check neutron logs for more information. [ 1826.917895] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] Traceback (most recent call last): [ 1826.917895] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] File "/opt/stack/nova/nova/compute/manager.py", line 2901, in _build_resources [ 1826.917895] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] yield resources [ 1826.917895] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] File "/opt/stack/nova/nova/compute/manager.py", line 2648, in _build_and_run_instance [ 1826.917895] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] self.driver.spawn(context, instance, image_meta, [ 1826.917895] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 1826.917895] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 1826.917895] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 1826.917895] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] vm_ref = self.build_virtual_machine(instance, [ 1826.918216] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 1826.918216] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] vif_infos = vmwarevif.get_vif_info(self._session, [ 1826.918216] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 1826.918216] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] for vif in network_info: [ 1826.918216] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 1826.918216] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] return self._sync_wrapper(fn, *args, **kwargs) [ 1826.918216] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 1826.918216] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] self.wait() [ 1826.918216] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 1826.918216] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] self[:] = self._gt.wait() [ 1826.918216] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 1826.918216] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] return self._exit_event.wait() [ 1826.918216] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 1826.918555] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] result = hub.switch() [ 1826.918555] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 1826.918555] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] return self.greenlet.switch() [ 1826.918555] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 1826.918555] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] result = function(*args, **kwargs) [ 1826.918555] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 1826.918555] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] return func(*args, **kwargs) [ 1826.918555] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] File "/opt/stack/nova/nova/compute/manager.py", line 2018, in _allocate_network_async [ 1826.918555] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] raise e [ 1826.918555] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] File "/opt/stack/nova/nova/compute/manager.py", line 1996, in _allocate_network_async [ 1826.918555] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] nwinfo = self.network_api.allocate_for_instance( [ 1826.918555] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 1826.918555] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] created_port_ids = self._update_ports_for_instance( [ 1826.918902] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 1826.918902] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] with excutils.save_and_reraise_exception(): [ 1826.918902] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1826.918902] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] self.force_reraise() [ 1826.918902] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1826.918902] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] raise self.value [ 1826.918902] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 1826.918902] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] updated_port = self._update_port( [ 1826.918902] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 1826.918902] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] _ensure_no_port_binding_failure(port) [ 1826.918902] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 1826.918902] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] raise exception.PortBindingFailed(port_id=port['id']) [ 1826.919220] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] nova.exception.PortBindingFailed: Binding failed for port 2a8b04db-a377-45bb-9e8f-39327bbfffce, please check neutron logs for more information. [ 1826.919220] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] [ 1826.919220] env[62824]: INFO nova.compute.manager [None req-250af368-8463-49f4-8a2e-de088db419fa tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] Terminating instance [ 1826.950685] env[62824]: DEBUG nova.compute.manager [None req-61fb5137-5f87-475f-a5d2-0722cddc68b4 tempest-ServersTestJSON-928648138 tempest-ServersTestJSON-928648138-project-member] [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1827.080205] env[62824]: ERROR nova.compute.manager [None req-6070f146-69b5-4aa1-b292-f106b5e9564b tempest-ServerAddressesNegativeTestJSON-1575558569 tempest-ServerAddressesNegativeTestJSON-1575558569-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 59ec24da-a159-4775-a06b-bf6a02869bed, please check neutron logs for more information. [ 1827.080205] env[62824]: ERROR nova.compute.manager Traceback (most recent call last): [ 1827.080205] env[62824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1996, in _allocate_network_async [ 1827.080205] env[62824]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 1827.080205] env[62824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 1827.080205] env[62824]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 1827.080205] env[62824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 1827.080205] env[62824]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 1827.080205] env[62824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1827.080205] env[62824]: ERROR nova.compute.manager self.force_reraise() [ 1827.080205] env[62824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1827.080205] env[62824]: ERROR nova.compute.manager raise self.value [ 1827.080205] env[62824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 1827.080205] env[62824]: ERROR nova.compute.manager updated_port = self._update_port( [ 1827.080205] env[62824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 1827.080205] env[62824]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 1827.081025] env[62824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 1827.081025] env[62824]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 1827.081025] env[62824]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 59ec24da-a159-4775-a06b-bf6a02869bed, please check neutron logs for more information. [ 1827.081025] env[62824]: ERROR nova.compute.manager [ 1827.081025] env[62824]: Traceback (most recent call last): [ 1827.081025] env[62824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 1827.081025] env[62824]: listener.cb(fileno) [ 1827.081025] env[62824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 1827.081025] env[62824]: result = function(*args, **kwargs) [ 1827.081025] env[62824]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 1827.081025] env[62824]: return func(*args, **kwargs) [ 1827.081025] env[62824]: File "/opt/stack/nova/nova/compute/manager.py", line 2018, in _allocate_network_async [ 1827.081025] env[62824]: raise e [ 1827.081025] env[62824]: File "/opt/stack/nova/nova/compute/manager.py", line 1996, in _allocate_network_async [ 1827.081025] env[62824]: nwinfo = self.network_api.allocate_for_instance( [ 1827.081025] env[62824]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 1827.081025] env[62824]: created_port_ids = self._update_ports_for_instance( [ 1827.081025] env[62824]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 1827.081025] env[62824]: with excutils.save_and_reraise_exception(): [ 1827.081025] env[62824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1827.081025] env[62824]: self.force_reraise() [ 1827.081025] env[62824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1827.081025] env[62824]: raise self.value [ 1827.081025] env[62824]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 1827.081025] env[62824]: updated_port = self._update_port( [ 1827.081025] env[62824]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 1827.081025] env[62824]: _ensure_no_port_binding_failure(port) [ 1827.081025] env[62824]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 1827.081025] env[62824]: raise exception.PortBindingFailed(port_id=port['id']) [ 1827.082032] env[62824]: nova.exception.PortBindingFailed: Binding failed for port 59ec24da-a159-4775-a06b-bf6a02869bed, please check neutron logs for more information. [ 1827.082032] env[62824]: Removing descriptor: 18 [ 1827.082032] env[62824]: ERROR nova.compute.manager [None req-6070f146-69b5-4aa1-b292-f106b5e9564b tempest-ServerAddressesNegativeTestJSON-1575558569 tempest-ServerAddressesNegativeTestJSON-1575558569-project-member] [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 59ec24da-a159-4775-a06b-bf6a02869bed, please check neutron logs for more information. [ 1827.082032] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] Traceback (most recent call last): [ 1827.082032] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] File "/opt/stack/nova/nova/compute/manager.py", line 2901, in _build_resources [ 1827.082032] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] yield resources [ 1827.082032] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] File "/opt/stack/nova/nova/compute/manager.py", line 2648, in _build_and_run_instance [ 1827.082032] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] self.driver.spawn(context, instance, image_meta, [ 1827.082032] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 1827.082032] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 1827.082032] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 1827.082032] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] vm_ref = self.build_virtual_machine(instance, [ 1827.082417] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 1827.082417] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] vif_infos = vmwarevif.get_vif_info(self._session, [ 1827.082417] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 1827.082417] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] for vif in network_info: [ 1827.082417] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 1827.082417] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] return self._sync_wrapper(fn, *args, **kwargs) [ 1827.082417] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 1827.082417] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] self.wait() [ 1827.082417] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 1827.082417] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] self[:] = self._gt.wait() [ 1827.082417] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 1827.082417] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] return self._exit_event.wait() [ 1827.082417] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 1827.082730] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] result = hub.switch() [ 1827.082730] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 1827.082730] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] return self.greenlet.switch() [ 1827.082730] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 1827.082730] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] result = function(*args, **kwargs) [ 1827.082730] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 1827.082730] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] return func(*args, **kwargs) [ 1827.082730] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] File "/opt/stack/nova/nova/compute/manager.py", line 2018, in _allocate_network_async [ 1827.082730] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] raise e [ 1827.082730] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] File "/opt/stack/nova/nova/compute/manager.py", line 1996, in _allocate_network_async [ 1827.082730] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] nwinfo = self.network_api.allocate_for_instance( [ 1827.082730] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 1827.082730] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] created_port_ids = self._update_ports_for_instance( [ 1827.083116] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 1827.083116] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] with excutils.save_and_reraise_exception(): [ 1827.083116] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1827.083116] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] self.force_reraise() [ 1827.083116] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1827.083116] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] raise self.value [ 1827.083116] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 1827.083116] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] updated_port = self._update_port( [ 1827.083116] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 1827.083116] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] _ensure_no_port_binding_failure(port) [ 1827.083116] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 1827.083116] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] raise exception.PortBindingFailed(port_id=port['id']) [ 1827.083462] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] nova.exception.PortBindingFailed: Binding failed for port 59ec24da-a159-4775-a06b-bf6a02869bed, please check neutron logs for more information. [ 1827.083462] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] [ 1827.083462] env[62824]: INFO nova.compute.manager [None req-6070f146-69b5-4aa1-b292-f106b5e9564b tempest-ServerAddressesNegativeTestJSON-1575558569 tempest-ServerAddressesNegativeTestJSON-1575558569-project-member] [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] Terminating instance [ 1827.153526] env[62824]: DEBUG nova.network.neutron [None req-84c279f8-4acc-49b0-9f33-b648872f4030 tempest-FloatingIPsAssociationTestJSON-1164183782 tempest-FloatingIPsAssociationTestJSON-1164183782-project-member] [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1827.269539] env[62824]: DEBUG nova.compute.manager [None req-7cd5dacb-b71b-429c-bfda-96302f774c53 tempest-ServersTestFqdnHostnames-528120407 tempest-ServersTestFqdnHostnames-528120407-project-member] [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1827.299731] env[62824]: DEBUG nova.network.neutron [None req-24599d6d-2f76-445f-8b64-070ec4ee0da9 tempest-TenantUsagesTestJSON-486947106 tempest-TenantUsagesTestJSON-486947106-project-member] [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1827.395243] env[62824]: DEBUG nova.network.neutron [None req-24599d6d-2f76-445f-8b64-070ec4ee0da9 tempest-TenantUsagesTestJSON-486947106 tempest-TenantUsagesTestJSON-486947106-project-member] [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1827.429265] env[62824]: DEBUG oslo_concurrency.lockutils [None req-250af368-8463-49f4-8a2e-de088db419fa tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Acquiring lock "refresh_cache-8b138cc7-8bde-48c8-b1bb-cec6ead877d3" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1827.429265] env[62824]: DEBUG oslo_concurrency.lockutils [None req-250af368-8463-49f4-8a2e-de088db419fa tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Acquired lock "refresh_cache-8b138cc7-8bde-48c8-b1bb-cec6ead877d3" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1827.429401] env[62824]: DEBUG nova.network.neutron [None req-250af368-8463-49f4-8a2e-de088db419fa tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1827.474462] env[62824]: DEBUG oslo_concurrency.lockutils [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Acquiring lock "f4157385-43ea-4e8c-887c-0985af37abd1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1827.474694] env[62824]: DEBUG oslo_concurrency.lockutils [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Lock "f4157385-43ea-4e8c-887c-0985af37abd1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1827.484800] env[62824]: DEBUG oslo_concurrency.lockutils [None req-61fb5137-5f87-475f-a5d2-0722cddc68b4 tempest-ServersTestJSON-928648138 tempest-ServersTestJSON-928648138-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1827.497329] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f2cf231-ff55-43b1-83ec-86ae8733a81d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1827.506821] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71a11f02-a395-4b55-8027-09f99ad87eb6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1827.540680] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-247e18fe-15a2-4f75-bd0b-f886e1c33f86 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1827.550444] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9328592f-ea62-44ee-8bbc-8189d8e4d7db {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1827.565750] env[62824]: DEBUG nova.compute.provider_tree [None req-2559566b-844a-4d61-b989-1f0a741aecd2 tempest-ServerDiagnosticsTest-1769488740 tempest-ServerDiagnosticsTest-1769488740-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1827.586872] env[62824]: DEBUG oslo_concurrency.lockutils [None req-6070f146-69b5-4aa1-b292-f106b5e9564b tempest-ServerAddressesNegativeTestJSON-1575558569 tempest-ServerAddressesNegativeTestJSON-1575558569-project-member] Acquiring lock "refresh_cache-6cc60e74-0474-47a5-ba57-c1c646ce0e1d" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1827.587056] env[62824]: DEBUG oslo_concurrency.lockutils [None req-6070f146-69b5-4aa1-b292-f106b5e9564b tempest-ServerAddressesNegativeTestJSON-1575558569 tempest-ServerAddressesNegativeTestJSON-1575558569-project-member] Acquired lock "refresh_cache-6cc60e74-0474-47a5-ba57-c1c646ce0e1d" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1827.587281] env[62824]: DEBUG nova.network.neutron [None req-6070f146-69b5-4aa1-b292-f106b5e9564b tempest-ServerAddressesNegativeTestJSON-1575558569 tempest-ServerAddressesNegativeTestJSON-1575558569-project-member] [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1827.656393] env[62824]: DEBUG oslo_concurrency.lockutils [None req-84c279f8-4acc-49b0-9f33-b648872f4030 tempest-FloatingIPsAssociationTestJSON-1164183782 tempest-FloatingIPsAssociationTestJSON-1164183782-project-member] Releasing lock "refresh_cache-764d7b4f-49d3-4ddf-8152-c08721dbce59" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1827.656644] env[62824]: DEBUG nova.compute.manager [None req-84c279f8-4acc-49b0-9f33-b648872f4030 tempest-FloatingIPsAssociationTestJSON-1164183782 tempest-FloatingIPsAssociationTestJSON-1164183782-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3036}} [ 1827.656825] env[62824]: DEBUG nova.compute.manager [None req-84c279f8-4acc-49b0-9f33-b648872f4030 tempest-FloatingIPsAssociationTestJSON-1164183782 tempest-FloatingIPsAssociationTestJSON-1164183782-project-member] [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1827.656989] env[62824]: DEBUG nova.network.neutron [None req-84c279f8-4acc-49b0-9f33-b648872f4030 tempest-FloatingIPsAssociationTestJSON-1164183782 tempest-FloatingIPsAssociationTestJSON-1164183782-project-member] [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1827.674998] env[62824]: DEBUG nova.network.neutron [None req-84c279f8-4acc-49b0-9f33-b648872f4030 tempest-FloatingIPsAssociationTestJSON-1164183782 tempest-FloatingIPsAssociationTestJSON-1164183782-project-member] [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1827.774140] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1827.807225] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7cd5dacb-b71b-429c-bfda-96302f774c53 tempest-ServersTestFqdnHostnames-528120407 tempest-ServersTestFqdnHostnames-528120407-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1827.898932] env[62824]: DEBUG oslo_concurrency.lockutils [None req-24599d6d-2f76-445f-8b64-070ec4ee0da9 tempest-TenantUsagesTestJSON-486947106 tempest-TenantUsagesTestJSON-486947106-project-member] Releasing lock "refresh_cache-9f016121-09f1-4da7-bbd0-7769d5104d27" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1827.899544] env[62824]: DEBUG nova.compute.manager [None req-24599d6d-2f76-445f-8b64-070ec4ee0da9 tempest-TenantUsagesTestJSON-486947106 tempest-TenantUsagesTestJSON-486947106-project-member] [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1827.903018] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-24599d6d-2f76-445f-8b64-070ec4ee0da9 tempest-TenantUsagesTestJSON-486947106 tempest-TenantUsagesTestJSON-486947106-project-member] [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1827.903018] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4fac0893-b20f-4989-afe5-b2ef87392e40 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1827.917799] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70ee2ff6-0d12-4fae-897b-05c6e9b031da {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1827.956933] env[62824]: WARNING nova.virt.vmwareapi.vmops [None req-24599d6d-2f76-445f-8b64-070ec4ee0da9 tempest-TenantUsagesTestJSON-486947106 tempest-TenantUsagesTestJSON-486947106-project-member] [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9f016121-09f1-4da7-bbd0-7769d5104d27 could not be found. [ 1827.956933] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-24599d6d-2f76-445f-8b64-070ec4ee0da9 tempest-TenantUsagesTestJSON-486947106 tempest-TenantUsagesTestJSON-486947106-project-member] [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1827.956933] env[62824]: INFO nova.compute.manager [None req-24599d6d-2f76-445f-8b64-070ec4ee0da9 tempest-TenantUsagesTestJSON-486947106 tempest-TenantUsagesTestJSON-486947106-project-member] [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] Took 0.06 seconds to destroy the instance on the hypervisor. [ 1827.956933] env[62824]: DEBUG oslo.service.loopingcall [None req-24599d6d-2f76-445f-8b64-070ec4ee0da9 tempest-TenantUsagesTestJSON-486947106 tempest-TenantUsagesTestJSON-486947106-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1827.956933] env[62824]: DEBUG nova.compute.manager [-] [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1827.956933] env[62824]: DEBUG nova.network.neutron [-] [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1827.977792] env[62824]: DEBUG nova.compute.manager [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] [instance: f4157385-43ea-4e8c-887c-0985af37abd1] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1828.030843] env[62824]: DEBUG nova.network.neutron [None req-250af368-8463-49f4-8a2e-de088db419fa tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1828.033055] env[62824]: DEBUG nova.network.neutron [-] [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1828.043011] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9f666b55-9bba-45f0-98a3-b9d8b6261703 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Acquiring lock "00fca8eb-c209-4632-abe6-5f092b04ac56" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1828.043011] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9f666b55-9bba-45f0-98a3-b9d8b6261703 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Lock "00fca8eb-c209-4632-abe6-5f092b04ac56" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1828.069594] env[62824]: DEBUG nova.scheduler.client.report [None req-2559566b-844a-4d61-b989-1f0a741aecd2 tempest-ServerDiagnosticsTest-1769488740 tempest-ServerDiagnosticsTest-1769488740-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1828.099537] env[62824]: DEBUG nova.network.neutron [None req-250af368-8463-49f4-8a2e-de088db419fa tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1828.120104] env[62824]: DEBUG nova.network.neutron [None req-6070f146-69b5-4aa1-b292-f106b5e9564b tempest-ServerAddressesNegativeTestJSON-1575558569 tempest-ServerAddressesNegativeTestJSON-1575558569-project-member] [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1828.134956] env[62824]: DEBUG nova.compute.manager [req-fcbdfaee-0206-4477-a705-8f0758525b87 req-d3b2000c-58be-4a87-b8b7-a107628c9a7d service nova] [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] Received event network-changed-b42c0eea-348b-430a-8119-0a16fac3dc7f {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1828.135317] env[62824]: DEBUG nova.compute.manager [req-fcbdfaee-0206-4477-a705-8f0758525b87 req-d3b2000c-58be-4a87-b8b7-a107628c9a7d service nova] [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] Refreshing instance network info cache due to event network-changed-b42c0eea-348b-430a-8119-0a16fac3dc7f. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1828.135543] env[62824]: DEBUG oslo_concurrency.lockutils [req-fcbdfaee-0206-4477-a705-8f0758525b87 req-d3b2000c-58be-4a87-b8b7-a107628c9a7d service nova] Acquiring lock "refresh_cache-9f016121-09f1-4da7-bbd0-7769d5104d27" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1828.135672] env[62824]: DEBUG oslo_concurrency.lockutils [req-fcbdfaee-0206-4477-a705-8f0758525b87 req-d3b2000c-58be-4a87-b8b7-a107628c9a7d service nova] Acquired lock "refresh_cache-9f016121-09f1-4da7-bbd0-7769d5104d27" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1828.135935] env[62824]: DEBUG nova.network.neutron [req-fcbdfaee-0206-4477-a705-8f0758525b87 req-d3b2000c-58be-4a87-b8b7-a107628c9a7d service nova] [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] Refreshing network info cache for port b42c0eea-348b-430a-8119-0a16fac3dc7f {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1828.178014] env[62824]: DEBUG nova.network.neutron [None req-84c279f8-4acc-49b0-9f33-b648872f4030 tempest-FloatingIPsAssociationTestJSON-1164183782 tempest-FloatingIPsAssociationTestJSON-1164183782-project-member] [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1828.238377] env[62824]: DEBUG nova.network.neutron [None req-6070f146-69b5-4aa1-b292-f106b5e9564b tempest-ServerAddressesNegativeTestJSON-1575558569 tempest-ServerAddressesNegativeTestJSON-1575558569-project-member] [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1828.506439] env[62824]: DEBUG oslo_concurrency.lockutils [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1828.543766] env[62824]: DEBUG nova.network.neutron [-] [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1828.553216] env[62824]: DEBUG nova.compute.manager [None req-9f666b55-9bba-45f0-98a3-b9d8b6261703 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1828.580042] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2559566b-844a-4d61-b989-1f0a741aecd2 tempest-ServerDiagnosticsTest-1769488740 tempest-ServerDiagnosticsTest-1769488740-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.275s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1828.580042] env[62824]: DEBUG nova.compute.manager [None req-2559566b-844a-4d61-b989-1f0a741aecd2 tempest-ServerDiagnosticsTest-1769488740 tempest-ServerDiagnosticsTest-1769488740-project-member] [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1828.581736] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b73330de-7630-444b-9674-b1882b4c4da9 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 2.625s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1828.605645] env[62824]: DEBUG oslo_concurrency.lockutils [None req-250af368-8463-49f4-8a2e-de088db419fa tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Releasing lock "refresh_cache-8b138cc7-8bde-48c8-b1bb-cec6ead877d3" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1828.606070] env[62824]: DEBUG nova.compute.manager [None req-250af368-8463-49f4-8a2e-de088db419fa tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1828.606307] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-250af368-8463-49f4-8a2e-de088db419fa tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1828.608120] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6b6007f7-baa5-42c6-a58a-c99282e06955 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1828.626019] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5f944c9-7338-45cc-91db-0b26eeec4310 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1828.661280] env[62824]: WARNING nova.virt.vmwareapi.vmops [None req-250af368-8463-49f4-8a2e-de088db419fa tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8b138cc7-8bde-48c8-b1bb-cec6ead877d3 could not be found. [ 1828.661869] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-250af368-8463-49f4-8a2e-de088db419fa tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1828.661869] env[62824]: INFO nova.compute.manager [None req-250af368-8463-49f4-8a2e-de088db419fa tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] Took 0.06 seconds to destroy the instance on the hypervisor. [ 1828.662034] env[62824]: DEBUG oslo.service.loopingcall [None req-250af368-8463-49f4-8a2e-de088db419fa tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1828.662277] env[62824]: DEBUG nova.compute.manager [-] [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1828.662396] env[62824]: DEBUG nova.network.neutron [-] [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1828.671276] env[62824]: DEBUG oslo_concurrency.lockutils [None req-34e3c185-2b6d-4b05-bb0f-3253c603dd1c tempest-InstanceActionsV221TestJSON-1308314300 tempest-InstanceActionsV221TestJSON-1308314300-project-member] Acquiring lock "bec6441f-72b8-4664-b125-b0aa5f21636a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1828.671276] env[62824]: DEBUG oslo_concurrency.lockutils [None req-34e3c185-2b6d-4b05-bb0f-3253c603dd1c tempest-InstanceActionsV221TestJSON-1308314300 tempest-InstanceActionsV221TestJSON-1308314300-project-member] Lock "bec6441f-72b8-4664-b125-b0aa5f21636a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1828.684116] env[62824]: INFO nova.compute.manager [None req-84c279f8-4acc-49b0-9f33-b648872f4030 tempest-FloatingIPsAssociationTestJSON-1164183782 tempest-FloatingIPsAssociationTestJSON-1164183782-project-member] [instance: 764d7b4f-49d3-4ddf-8152-c08721dbce59] Took 1.03 seconds to deallocate network for instance. [ 1828.711576] env[62824]: DEBUG nova.network.neutron [req-fcbdfaee-0206-4477-a705-8f0758525b87 req-d3b2000c-58be-4a87-b8b7-a107628c9a7d service nova] [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1828.741359] env[62824]: DEBUG oslo_concurrency.lockutils [None req-6070f146-69b5-4aa1-b292-f106b5e9564b tempest-ServerAddressesNegativeTestJSON-1575558569 tempest-ServerAddressesNegativeTestJSON-1575558569-project-member] Releasing lock "refresh_cache-6cc60e74-0474-47a5-ba57-c1c646ce0e1d" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1828.741858] env[62824]: DEBUG nova.compute.manager [None req-6070f146-69b5-4aa1-b292-f106b5e9564b tempest-ServerAddressesNegativeTestJSON-1575558569 tempest-ServerAddressesNegativeTestJSON-1575558569-project-member] [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1828.742098] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-6070f146-69b5-4aa1-b292-f106b5e9564b tempest-ServerAddressesNegativeTestJSON-1575558569 tempest-ServerAddressesNegativeTestJSON-1575558569-project-member] [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1828.742750] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-44f677fe-fc05-4756-b84e-b91ecb1f74eb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1828.755029] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-412bdd5c-b0a0-4400-a3bb-eb2da639f1e3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1828.768362] env[62824]: DEBUG nova.network.neutron [-] [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1828.777865] env[62824]: WARNING nova.virt.vmwareapi.vmops [None req-6070f146-69b5-4aa1-b292-f106b5e9564b tempest-ServerAddressesNegativeTestJSON-1575558569 tempest-ServerAddressesNegativeTestJSON-1575558569-project-member] [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6cc60e74-0474-47a5-ba57-c1c646ce0e1d could not be found. [ 1828.778135] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-6070f146-69b5-4aa1-b292-f106b5e9564b tempest-ServerAddressesNegativeTestJSON-1575558569 tempest-ServerAddressesNegativeTestJSON-1575558569-project-member] [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1828.778328] env[62824]: INFO nova.compute.manager [None req-6070f146-69b5-4aa1-b292-f106b5e9564b tempest-ServerAddressesNegativeTestJSON-1575558569 tempest-ServerAddressesNegativeTestJSON-1575558569-project-member] [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] Took 0.04 seconds to destroy the instance on the hypervisor. [ 1828.778573] env[62824]: DEBUG oslo.service.loopingcall [None req-6070f146-69b5-4aa1-b292-f106b5e9564b tempest-ServerAddressesNegativeTestJSON-1575558569 tempest-ServerAddressesNegativeTestJSON-1575558569-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1828.779505] env[62824]: DEBUG nova.compute.manager [-] [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1828.779505] env[62824]: DEBUG nova.network.neutron [-] [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1828.830439] env[62824]: DEBUG nova.network.neutron [-] [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1829.024724] env[62824]: DEBUG nova.network.neutron [req-fcbdfaee-0206-4477-a705-8f0758525b87 req-d3b2000c-58be-4a87-b8b7-a107628c9a7d service nova] [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1829.054274] env[62824]: INFO nova.compute.manager [-] [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] Took 1.10 seconds to deallocate network for instance. [ 1829.055966] env[62824]: DEBUG nova.compute.claims [None req-24599d6d-2f76-445f-8b64-070ec4ee0da9 tempest-TenantUsagesTestJSON-486947106 tempest-TenantUsagesTestJSON-486947106-project-member] [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] Aborting claim: {{(pid=62824) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 1829.056269] env[62824]: DEBUG oslo_concurrency.lockutils [None req-24599d6d-2f76-445f-8b64-070ec4ee0da9 tempest-TenantUsagesTestJSON-486947106 tempest-TenantUsagesTestJSON-486947106-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1829.081744] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9f666b55-9bba-45f0-98a3-b9d8b6261703 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1829.088281] env[62824]: DEBUG nova.compute.utils [None req-2559566b-844a-4d61-b989-1f0a741aecd2 tempest-ServerDiagnosticsTest-1769488740 tempest-ServerDiagnosticsTest-1769488740-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1829.092454] env[62824]: DEBUG nova.compute.manager [None req-2559566b-844a-4d61-b989-1f0a741aecd2 tempest-ServerDiagnosticsTest-1769488740 tempest-ServerDiagnosticsTest-1769488740-project-member] [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1829.092630] env[62824]: DEBUG nova.network.neutron [None req-2559566b-844a-4d61-b989-1f0a741aecd2 tempest-ServerDiagnosticsTest-1769488740 tempest-ServerDiagnosticsTest-1769488740-project-member] [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1829.271546] env[62824]: DEBUG nova.network.neutron [-] [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1829.313914] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27a56c87-87b1-41d9-ae76-5da027d6b8dc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1829.325291] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4729313a-dd9a-4762-9392-deed4470dd02 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1829.371567] env[62824]: DEBUG nova.network.neutron [-] [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1829.376830] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c7bbab7-9ff3-4fe7-8ec3-de47e88f377d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1829.386915] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a383058-6067-4887-9d61-8488c8171c41 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1829.402181] env[62824]: DEBUG nova.compute.provider_tree [None req-b73330de-7630-444b-9674-b1882b4c4da9 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1829.475507] env[62824]: DEBUG nova.policy [None req-2559566b-844a-4d61-b989-1f0a741aecd2 tempest-ServerDiagnosticsTest-1769488740 tempest-ServerDiagnosticsTest-1769488740-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4d7dbca4a8ec47498f16a7749291648d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd0b2f2d66c574f3faf82666260df78db', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 1829.529372] env[62824]: DEBUG oslo_concurrency.lockutils [req-fcbdfaee-0206-4477-a705-8f0758525b87 req-d3b2000c-58be-4a87-b8b7-a107628c9a7d service nova] Releasing lock "refresh_cache-9f016121-09f1-4da7-bbd0-7769d5104d27" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1829.595732] env[62824]: DEBUG nova.compute.manager [None req-2559566b-844a-4d61-b989-1f0a741aecd2 tempest-ServerDiagnosticsTest-1769488740 tempest-ServerDiagnosticsTest-1769488740-project-member] [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1829.745877] env[62824]: INFO nova.scheduler.client.report [None req-84c279f8-4acc-49b0-9f33-b648872f4030 tempest-FloatingIPsAssociationTestJSON-1164183782 tempest-FloatingIPsAssociationTestJSON-1164183782-project-member] Deleted allocations for instance 764d7b4f-49d3-4ddf-8152-c08721dbce59 [ 1829.768955] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1829.772800] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager.update_available_resource {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1829.774085] env[62824]: INFO nova.compute.manager [-] [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] Took 1.11 seconds to deallocate network for instance. [ 1829.777601] env[62824]: DEBUG nova.compute.claims [None req-250af368-8463-49f4-8a2e-de088db419fa tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] Aborting claim: {{(pid=62824) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 1829.777601] env[62824]: DEBUG oslo_concurrency.lockutils [None req-250af368-8463-49f4-8a2e-de088db419fa tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1829.877565] env[62824]: INFO nova.compute.manager [-] [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] Took 1.10 seconds to deallocate network for instance. [ 1829.880719] env[62824]: DEBUG nova.compute.claims [None req-6070f146-69b5-4aa1-b292-f106b5e9564b tempest-ServerAddressesNegativeTestJSON-1575558569 tempest-ServerAddressesNegativeTestJSON-1575558569-project-member] [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] Aborting claim: {{(pid=62824) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 1829.880888] env[62824]: DEBUG oslo_concurrency.lockutils [None req-6070f146-69b5-4aa1-b292-f106b5e9564b tempest-ServerAddressesNegativeTestJSON-1575558569 tempest-ServerAddressesNegativeTestJSON-1575558569-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1829.907175] env[62824]: DEBUG nova.scheduler.client.report [None req-b73330de-7630-444b-9674-b1882b4c4da9 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1830.085091] env[62824]: DEBUG nova.compute.manager [req-4ee55ea5-84a2-4234-ab0e-bdc8c73790bb req-7e7c06e7-6bff-46c0-b4d6-1ab5938f2728 service nova] [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] Received event network-changed-2a8b04db-a377-45bb-9e8f-39327bbfffce {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1830.085281] env[62824]: DEBUG nova.compute.manager [req-4ee55ea5-84a2-4234-ab0e-bdc8c73790bb req-7e7c06e7-6bff-46c0-b4d6-1ab5938f2728 service nova] [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] Refreshing instance network info cache due to event network-changed-2a8b04db-a377-45bb-9e8f-39327bbfffce. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1830.086889] env[62824]: DEBUG oslo_concurrency.lockutils [req-4ee55ea5-84a2-4234-ab0e-bdc8c73790bb req-7e7c06e7-6bff-46c0-b4d6-1ab5938f2728 service nova] Acquiring lock "refresh_cache-8b138cc7-8bde-48c8-b1bb-cec6ead877d3" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1830.086889] env[62824]: DEBUG oslo_concurrency.lockutils [req-4ee55ea5-84a2-4234-ab0e-bdc8c73790bb req-7e7c06e7-6bff-46c0-b4d6-1ab5938f2728 service nova] Acquired lock "refresh_cache-8b138cc7-8bde-48c8-b1bb-cec6ead877d3" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1830.086889] env[62824]: DEBUG nova.network.neutron [req-4ee55ea5-84a2-4234-ab0e-bdc8c73790bb req-7e7c06e7-6bff-46c0-b4d6-1ab5938f2728 service nova] [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] Refreshing network info cache for port 2a8b04db-a377-45bb-9e8f-39327bbfffce {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1830.256352] env[62824]: DEBUG oslo_concurrency.lockutils [None req-84c279f8-4acc-49b0-9f33-b648872f4030 tempest-FloatingIPsAssociationTestJSON-1164183782 tempest-FloatingIPsAssociationTestJSON-1164183782-project-member] Lock "764d7b4f-49d3-4ddf-8152-c08721dbce59" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.855s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1830.275415] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1830.414254] env[62824]: DEBUG nova.network.neutron [None req-2559566b-844a-4d61-b989-1f0a741aecd2 tempest-ServerDiagnosticsTest-1769488740 tempest-ServerDiagnosticsTest-1769488740-project-member] [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] Successfully created port: 3fe39149-045f-4e9f-8ab4-72da2377c013 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1830.416900] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b73330de-7630-444b-9674-b1882b4c4da9 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.835s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1830.417597] env[62824]: ERROR nova.compute.manager [None req-b73330de-7630-444b-9674-b1882b4c4da9 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3af5eeb4-c5d9-469e-9f53-3b9cc122490c, please check neutron logs for more information. [ 1830.417597] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] Traceback (most recent call last): [ 1830.417597] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] File "/opt/stack/nova/nova/compute/manager.py", line 2648, in _build_and_run_instance [ 1830.417597] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] self.driver.spawn(context, instance, image_meta, [ 1830.417597] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 1830.417597] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] self._vmops.spawn(context, instance, image_meta, injected_files, [ 1830.417597] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 1830.417597] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] vm_ref = self.build_virtual_machine(instance, [ 1830.417597] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 1830.417597] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] vif_infos = vmwarevif.get_vif_info(self._session, [ 1830.417597] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 1830.417941] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] for vif in network_info: [ 1830.417941] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 1830.417941] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] return self._sync_wrapper(fn, *args, **kwargs) [ 1830.417941] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 1830.417941] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] self.wait() [ 1830.417941] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 1830.417941] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] self[:] = self._gt.wait() [ 1830.417941] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 1830.417941] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] return self._exit_event.wait() [ 1830.417941] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 1830.417941] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] result = hub.switch() [ 1830.417941] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 1830.417941] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] return self.greenlet.switch() [ 1830.418343] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 1830.418343] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] result = function(*args, **kwargs) [ 1830.418343] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 1830.418343] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] return func(*args, **kwargs) [ 1830.418343] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] File "/opt/stack/nova/nova/compute/manager.py", line 2018, in _allocate_network_async [ 1830.418343] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] raise e [ 1830.418343] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] File "/opt/stack/nova/nova/compute/manager.py", line 1996, in _allocate_network_async [ 1830.418343] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] nwinfo = self.network_api.allocate_for_instance( [ 1830.418343] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 1830.418343] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] created_port_ids = self._update_ports_for_instance( [ 1830.418343] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 1830.418343] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] with excutils.save_and_reraise_exception(): [ 1830.418343] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1830.418657] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] self.force_reraise() [ 1830.418657] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1830.418657] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] raise self.value [ 1830.418657] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 1830.418657] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] updated_port = self._update_port( [ 1830.418657] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 1830.418657] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] _ensure_no_port_binding_failure(port) [ 1830.418657] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 1830.418657] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] raise exception.PortBindingFailed(port_id=port['id']) [ 1830.418657] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] nova.exception.PortBindingFailed: Binding failed for port 3af5eeb4-c5d9-469e-9f53-3b9cc122490c, please check neutron logs for more information. [ 1830.418657] env[62824]: ERROR nova.compute.manager [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] [ 1830.418932] env[62824]: DEBUG nova.compute.utils [None req-b73330de-7630-444b-9674-b1882b4c4da9 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] Binding failed for port 3af5eeb4-c5d9-469e-9f53-3b9cc122490c, please check neutron logs for more information. {{(pid=62824) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 1830.419695] env[62824]: DEBUG oslo_concurrency.lockutils [None req-61fb5137-5f87-475f-a5d2-0722cddc68b4 tempest-ServersTestJSON-928648138 tempest-ServersTestJSON-928648138-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.935s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1830.420944] env[62824]: INFO nova.compute.claims [None req-61fb5137-5f87-475f-a5d2-0722cddc68b4 tempest-ServersTestJSON-928648138 tempest-ServersTestJSON-928648138-project-member] [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1830.425591] env[62824]: DEBUG nova.compute.manager [None req-b73330de-7630-444b-9674-b1882b4c4da9 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] Build of instance 579e1ca1-a6b6-4fde-8368-3e635a303032 was re-scheduled: Binding failed for port 3af5eeb4-c5d9-469e-9f53-3b9cc122490c, please check neutron logs for more information. {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2487}} [ 1830.425837] env[62824]: DEBUG nova.compute.manager [None req-b73330de-7630-444b-9674-b1882b4c4da9 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] Unplugging VIFs for instance {{(pid=62824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3013}} [ 1830.426065] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b73330de-7630-444b-9674-b1882b4c4da9 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Acquiring lock "refresh_cache-579e1ca1-a6b6-4fde-8368-3e635a303032" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1830.426256] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b73330de-7630-444b-9674-b1882b4c4da9 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Acquired lock "refresh_cache-579e1ca1-a6b6-4fde-8368-3e635a303032" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1830.426362] env[62824]: DEBUG nova.network.neutron [None req-b73330de-7630-444b-9674-b1882b4c4da9 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1830.443732] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5cdb74c7-a27b-45c4-a89a-19c7714f04b5 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694-project-member] Acquiring lock "2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1830.444087] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5cdb74c7-a27b-45c4-a89a-19c7714f04b5 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694-project-member] Lock "2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1830.612967] env[62824]: DEBUG nova.compute.manager [None req-2559566b-844a-4d61-b989-1f0a741aecd2 tempest-ServerDiagnosticsTest-1769488740 tempest-ServerDiagnosticsTest-1769488740-project-member] [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1830.618899] env[62824]: DEBUG nova.network.neutron [req-4ee55ea5-84a2-4234-ab0e-bdc8c73790bb req-7e7c06e7-6bff-46c0-b4d6-1ab5938f2728 service nova] [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1830.641466] env[62824]: DEBUG nova.virt.hardware [None req-2559566b-844a-4d61-b989-1f0a741aecd2 tempest-ServerDiagnosticsTest-1769488740 tempest-ServerDiagnosticsTest-1769488740-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1830.641918] env[62824]: DEBUG nova.virt.hardware [None req-2559566b-844a-4d61-b989-1f0a741aecd2 tempest-ServerDiagnosticsTest-1769488740 tempest-ServerDiagnosticsTest-1769488740-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1830.643056] env[62824]: DEBUG nova.virt.hardware [None req-2559566b-844a-4d61-b989-1f0a741aecd2 tempest-ServerDiagnosticsTest-1769488740 tempest-ServerDiagnosticsTest-1769488740-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1830.643056] env[62824]: DEBUG nova.virt.hardware [None req-2559566b-844a-4d61-b989-1f0a741aecd2 tempest-ServerDiagnosticsTest-1769488740 tempest-ServerDiagnosticsTest-1769488740-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1830.643056] env[62824]: DEBUG nova.virt.hardware [None req-2559566b-844a-4d61-b989-1f0a741aecd2 tempest-ServerDiagnosticsTest-1769488740 tempest-ServerDiagnosticsTest-1769488740-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1830.643056] env[62824]: DEBUG nova.virt.hardware [None req-2559566b-844a-4d61-b989-1f0a741aecd2 tempest-ServerDiagnosticsTest-1769488740 tempest-ServerDiagnosticsTest-1769488740-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1830.643056] env[62824]: DEBUG nova.virt.hardware [None req-2559566b-844a-4d61-b989-1f0a741aecd2 tempest-ServerDiagnosticsTest-1769488740 tempest-ServerDiagnosticsTest-1769488740-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1830.643420] env[62824]: DEBUG nova.virt.hardware [None req-2559566b-844a-4d61-b989-1f0a741aecd2 tempest-ServerDiagnosticsTest-1769488740 tempest-ServerDiagnosticsTest-1769488740-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1830.644071] env[62824]: DEBUG nova.virt.hardware [None req-2559566b-844a-4d61-b989-1f0a741aecd2 tempest-ServerDiagnosticsTest-1769488740 tempest-ServerDiagnosticsTest-1769488740-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1830.644071] env[62824]: DEBUG nova.virt.hardware [None req-2559566b-844a-4d61-b989-1f0a741aecd2 tempest-ServerDiagnosticsTest-1769488740 tempest-ServerDiagnosticsTest-1769488740-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1830.644202] env[62824]: DEBUG nova.virt.hardware [None req-2559566b-844a-4d61-b989-1f0a741aecd2 tempest-ServerDiagnosticsTest-1769488740 tempest-ServerDiagnosticsTest-1769488740-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1830.645214] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c9e4a35-5814-4438-863a-18903d5535a2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1830.654843] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3214e2db-2468-47b2-a0d2-e386bc43ea50 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1830.760599] env[62824]: DEBUG nova.compute.manager [None req-34e3c185-2b6d-4b05-bb0f-3253c603dd1c tempest-InstanceActionsV221TestJSON-1308314300 tempest-InstanceActionsV221TestJSON-1308314300-project-member] [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1830.766488] env[62824]: DEBUG nova.network.neutron [req-4ee55ea5-84a2-4234-ab0e-bdc8c73790bb req-7e7c06e7-6bff-46c0-b4d6-1ab5938f2728 service nova] [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1830.959749] env[62824]: DEBUG nova.network.neutron [None req-b73330de-7630-444b-9674-b1882b4c4da9 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1831.052359] env[62824]: DEBUG nova.network.neutron [None req-b73330de-7630-444b-9674-b1882b4c4da9 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1831.162021] env[62824]: DEBUG nova.compute.manager [req-0aa7108e-0ecc-4938-a22e-30c82142d1c5 req-69a8de9f-6c35-4652-8842-d9b528460c2a service nova] [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] Received event network-vif-deleted-b42c0eea-348b-430a-8119-0a16fac3dc7f {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1831.162021] env[62824]: DEBUG nova.compute.manager [req-0aa7108e-0ecc-4938-a22e-30c82142d1c5 req-69a8de9f-6c35-4652-8842-d9b528460c2a service nova] [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] Received event network-changed-59ec24da-a159-4775-a06b-bf6a02869bed {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1831.162021] env[62824]: DEBUG nova.compute.manager [req-0aa7108e-0ecc-4938-a22e-30c82142d1c5 req-69a8de9f-6c35-4652-8842-d9b528460c2a service nova] [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] Refreshing instance network info cache due to event network-changed-59ec24da-a159-4775-a06b-bf6a02869bed. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1831.162021] env[62824]: DEBUG oslo_concurrency.lockutils [req-0aa7108e-0ecc-4938-a22e-30c82142d1c5 req-69a8de9f-6c35-4652-8842-d9b528460c2a service nova] Acquiring lock "refresh_cache-6cc60e74-0474-47a5-ba57-c1c646ce0e1d" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1831.162021] env[62824]: DEBUG oslo_concurrency.lockutils [req-0aa7108e-0ecc-4938-a22e-30c82142d1c5 req-69a8de9f-6c35-4652-8842-d9b528460c2a service nova] Acquired lock "refresh_cache-6cc60e74-0474-47a5-ba57-c1c646ce0e1d" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1831.162359] env[62824]: DEBUG nova.network.neutron [req-0aa7108e-0ecc-4938-a22e-30c82142d1c5 req-69a8de9f-6c35-4652-8842-d9b528460c2a service nova] [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] Refreshing network info cache for port 59ec24da-a159-4775-a06b-bf6a02869bed {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1831.272368] env[62824]: DEBUG oslo_concurrency.lockutils [req-4ee55ea5-84a2-4234-ab0e-bdc8c73790bb req-7e7c06e7-6bff-46c0-b4d6-1ab5938f2728 service nova] Releasing lock "refresh_cache-8b138cc7-8bde-48c8-b1bb-cec6ead877d3" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1831.272619] env[62824]: DEBUG nova.compute.manager [req-4ee55ea5-84a2-4234-ab0e-bdc8c73790bb req-7e7c06e7-6bff-46c0-b4d6-1ab5938f2728 service nova] [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] Received event network-vif-deleted-2a8b04db-a377-45bb-9e8f-39327bbfffce {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1831.291246] env[62824]: DEBUG oslo_concurrency.lockutils [None req-34e3c185-2b6d-4b05-bb0f-3253c603dd1c tempest-InstanceActionsV221TestJSON-1308314300 tempest-InstanceActionsV221TestJSON-1308314300-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1831.554236] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b73330de-7630-444b-9674-b1882b4c4da9 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Releasing lock "refresh_cache-579e1ca1-a6b6-4fde-8368-3e635a303032" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1831.554477] env[62824]: DEBUG nova.compute.manager [None req-b73330de-7630-444b-9674-b1882b4c4da9 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3036}} [ 1831.554731] env[62824]: DEBUG nova.compute.manager [None req-b73330de-7630-444b-9674-b1882b4c4da9 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1831.554813] env[62824]: DEBUG nova.network.neutron [None req-b73330de-7630-444b-9674-b1882b4c4da9 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1831.588792] env[62824]: DEBUG nova.network.neutron [None req-b73330de-7630-444b-9674-b1882b4c4da9 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1831.635996] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54fbc504-f00f-4dde-953a-42a250797f51 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1831.645064] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ff9643b-2a1a-4711-ac1a-cc799d0e5c43 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1831.683888] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-061ab47a-09cd-4f1a-8fde-6393b167a9d9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1831.693774] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42f5ee1c-f639-4cfb-a90a-46eef90dd06b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1831.710140] env[62824]: DEBUG nova.compute.provider_tree [None req-61fb5137-5f87-475f-a5d2-0722cddc68b4 tempest-ServersTestJSON-928648138 tempest-ServersTestJSON-928648138-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1831.719753] env[62824]: DEBUG nova.network.neutron [req-0aa7108e-0ecc-4938-a22e-30c82142d1c5 req-69a8de9f-6c35-4652-8842-d9b528460c2a service nova] [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1831.943354] env[62824]: DEBUG nova.network.neutron [req-0aa7108e-0ecc-4938-a22e-30c82142d1c5 req-69a8de9f-6c35-4652-8842-d9b528460c2a service nova] [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1832.096667] env[62824]: DEBUG nova.network.neutron [None req-b73330de-7630-444b-9674-b1882b4c4da9 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1832.213755] env[62824]: DEBUG nova.scheduler.client.report [None req-61fb5137-5f87-475f-a5d2-0722cddc68b4 tempest-ServersTestJSON-928648138 tempest-ServersTestJSON-928648138-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1832.374232] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Acquiring lock "29b8dd5f-1855-490c-a01b-54840073a753" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1832.374232] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Lock "29b8dd5f-1855-490c-a01b-54840073a753" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1832.446859] env[62824]: DEBUG oslo_concurrency.lockutils [req-0aa7108e-0ecc-4938-a22e-30c82142d1c5 req-69a8de9f-6c35-4652-8842-d9b528460c2a service nova] Releasing lock "refresh_cache-6cc60e74-0474-47a5-ba57-c1c646ce0e1d" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1832.448019] env[62824]: DEBUG nova.compute.manager [req-0aa7108e-0ecc-4938-a22e-30c82142d1c5 req-69a8de9f-6c35-4652-8842-d9b528460c2a service nova] [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] Received event network-vif-deleted-59ec24da-a159-4775-a06b-bf6a02869bed {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1832.599078] env[62824]: INFO nova.compute.manager [None req-b73330de-7630-444b-9674-b1882b4c4da9 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] [instance: 579e1ca1-a6b6-4fde-8368-3e635a303032] Took 1.04 seconds to deallocate network for instance. [ 1832.725079] env[62824]: DEBUG oslo_concurrency.lockutils [None req-61fb5137-5f87-475f-a5d2-0722cddc68b4 tempest-ServersTestJSON-928648138 tempest-ServersTestJSON-928648138-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.305s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1832.726213] env[62824]: DEBUG nova.compute.manager [None req-61fb5137-5f87-475f-a5d2-0722cddc68b4 tempest-ServersTestJSON-928648138 tempest-ServersTestJSON-928648138-project-member] [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1832.735018] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7cd5dacb-b71b-429c-bfda-96302f774c53 tempest-ServersTestFqdnHostnames-528120407 tempest-ServersTestFqdnHostnames-528120407-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.925s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1832.736438] env[62824]: INFO nova.compute.claims [None req-7cd5dacb-b71b-429c-bfda-96302f774c53 tempest-ServersTestFqdnHostnames-528120407 tempest-ServersTestFqdnHostnames-528120407-project-member] [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1832.993834] env[62824]: ERROR nova.compute.manager [None req-2559566b-844a-4d61-b989-1f0a741aecd2 tempest-ServerDiagnosticsTest-1769488740 tempest-ServerDiagnosticsTest-1769488740-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3fe39149-045f-4e9f-8ab4-72da2377c013, please check neutron logs for more information. [ 1832.993834] env[62824]: ERROR nova.compute.manager Traceback (most recent call last): [ 1832.993834] env[62824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1996, in _allocate_network_async [ 1832.993834] env[62824]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 1832.993834] env[62824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 1832.993834] env[62824]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 1832.993834] env[62824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 1832.993834] env[62824]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 1832.993834] env[62824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1832.993834] env[62824]: ERROR nova.compute.manager self.force_reraise() [ 1832.993834] env[62824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1832.993834] env[62824]: ERROR nova.compute.manager raise self.value [ 1832.993834] env[62824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 1832.993834] env[62824]: ERROR nova.compute.manager updated_port = self._update_port( [ 1832.993834] env[62824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 1832.993834] env[62824]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 1832.994320] env[62824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 1832.994320] env[62824]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 1832.994320] env[62824]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3fe39149-045f-4e9f-8ab4-72da2377c013, please check neutron logs for more information. [ 1832.994320] env[62824]: ERROR nova.compute.manager [ 1832.994320] env[62824]: Traceback (most recent call last): [ 1832.994320] env[62824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 1832.994320] env[62824]: listener.cb(fileno) [ 1832.994320] env[62824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 1832.994320] env[62824]: result = function(*args, **kwargs) [ 1832.994320] env[62824]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 1832.994320] env[62824]: return func(*args, **kwargs) [ 1832.994320] env[62824]: File "/opt/stack/nova/nova/compute/manager.py", line 2018, in _allocate_network_async [ 1832.994320] env[62824]: raise e [ 1832.994320] env[62824]: File "/opt/stack/nova/nova/compute/manager.py", line 1996, in _allocate_network_async [ 1832.994320] env[62824]: nwinfo = self.network_api.allocate_for_instance( [ 1832.994320] env[62824]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 1832.994320] env[62824]: created_port_ids = self._update_ports_for_instance( [ 1832.994320] env[62824]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 1832.994320] env[62824]: with excutils.save_and_reraise_exception(): [ 1832.994320] env[62824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1832.994320] env[62824]: self.force_reraise() [ 1832.994320] env[62824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1832.994320] env[62824]: raise self.value [ 1832.994320] env[62824]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 1832.994320] env[62824]: updated_port = self._update_port( [ 1832.994320] env[62824]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 1832.994320] env[62824]: _ensure_no_port_binding_failure(port) [ 1832.994320] env[62824]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 1832.994320] env[62824]: raise exception.PortBindingFailed(port_id=port['id']) [ 1832.995090] env[62824]: nova.exception.PortBindingFailed: Binding failed for port 3fe39149-045f-4e9f-8ab4-72da2377c013, please check neutron logs for more information. [ 1832.995090] env[62824]: Removing descriptor: 18 [ 1832.995090] env[62824]: ERROR nova.compute.manager [None req-2559566b-844a-4d61-b989-1f0a741aecd2 tempest-ServerDiagnosticsTest-1769488740 tempest-ServerDiagnosticsTest-1769488740-project-member] [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3fe39149-045f-4e9f-8ab4-72da2377c013, please check neutron logs for more information. [ 1832.995090] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] Traceback (most recent call last): [ 1832.995090] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] File "/opt/stack/nova/nova/compute/manager.py", line 2901, in _build_resources [ 1832.995090] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] yield resources [ 1832.995090] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] File "/opt/stack/nova/nova/compute/manager.py", line 2648, in _build_and_run_instance [ 1832.995090] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] self.driver.spawn(context, instance, image_meta, [ 1832.995090] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 1832.995090] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 1832.995090] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 1832.995090] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] vm_ref = self.build_virtual_machine(instance, [ 1832.995473] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 1832.995473] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] vif_infos = vmwarevif.get_vif_info(self._session, [ 1832.995473] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 1832.995473] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] for vif in network_info: [ 1832.995473] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 1832.995473] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] return self._sync_wrapper(fn, *args, **kwargs) [ 1832.995473] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 1832.995473] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] self.wait() [ 1832.995473] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 1832.995473] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] self[:] = self._gt.wait() [ 1832.995473] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 1832.995473] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] return self._exit_event.wait() [ 1832.995473] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 1832.995865] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] result = hub.switch() [ 1832.995865] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 1832.995865] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] return self.greenlet.switch() [ 1832.995865] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 1832.995865] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] result = function(*args, **kwargs) [ 1832.995865] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 1832.995865] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] return func(*args, **kwargs) [ 1832.995865] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] File "/opt/stack/nova/nova/compute/manager.py", line 2018, in _allocate_network_async [ 1832.995865] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] raise e [ 1832.995865] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] File "/opt/stack/nova/nova/compute/manager.py", line 1996, in _allocate_network_async [ 1832.995865] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] nwinfo = self.network_api.allocate_for_instance( [ 1832.995865] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 1832.995865] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] created_port_ids = self._update_ports_for_instance( [ 1832.996215] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 1832.996215] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] with excutils.save_and_reraise_exception(): [ 1832.996215] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1832.996215] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] self.force_reraise() [ 1832.996215] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1832.996215] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] raise self.value [ 1832.996215] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 1832.996215] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] updated_port = self._update_port( [ 1832.996215] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 1832.996215] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] _ensure_no_port_binding_failure(port) [ 1832.996215] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 1832.996215] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] raise exception.PortBindingFailed(port_id=port['id']) [ 1832.996522] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] nova.exception.PortBindingFailed: Binding failed for port 3fe39149-045f-4e9f-8ab4-72da2377c013, please check neutron logs for more information. [ 1832.996522] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] [ 1832.996522] env[62824]: INFO nova.compute.manager [None req-2559566b-844a-4d61-b989-1f0a741aecd2 tempest-ServerDiagnosticsTest-1769488740 tempest-ServerDiagnosticsTest-1769488740-project-member] [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] Terminating instance [ 1833.244363] env[62824]: DEBUG nova.compute.utils [None req-61fb5137-5f87-475f-a5d2-0722cddc68b4 tempest-ServersTestJSON-928648138 tempest-ServersTestJSON-928648138-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1833.254012] env[62824]: DEBUG nova.compute.manager [None req-61fb5137-5f87-475f-a5d2-0722cddc68b4 tempest-ServersTestJSON-928648138 tempest-ServersTestJSON-928648138-project-member] [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1833.254219] env[62824]: DEBUG nova.network.neutron [None req-61fb5137-5f87-475f-a5d2-0722cddc68b4 tempest-ServersTestJSON-928648138 tempest-ServersTestJSON-928648138-project-member] [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1833.356925] env[62824]: DEBUG nova.policy [None req-61fb5137-5f87-475f-a5d2-0722cddc68b4 tempest-ServersTestJSON-928648138 tempest-ServersTestJSON-928648138-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '577ad1d1fd2a442587174ba32d84a4d0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bd76edbee43b4729890238bad67a6f11', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 1833.502036] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2559566b-844a-4d61-b989-1f0a741aecd2 tempest-ServerDiagnosticsTest-1769488740 tempest-ServerDiagnosticsTest-1769488740-project-member] Acquiring lock "refresh_cache-042c5f0f-bf1f-417e-be05-964cfcb20c7e" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1833.502036] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2559566b-844a-4d61-b989-1f0a741aecd2 tempest-ServerDiagnosticsTest-1769488740 tempest-ServerDiagnosticsTest-1769488740-project-member] Acquired lock "refresh_cache-042c5f0f-bf1f-417e-be05-964cfcb20c7e" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1833.502036] env[62824]: DEBUG nova.network.neutron [None req-2559566b-844a-4d61-b989-1f0a741aecd2 tempest-ServerDiagnosticsTest-1769488740 tempest-ServerDiagnosticsTest-1769488740-project-member] [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1833.641176] env[62824]: INFO nova.scheduler.client.report [None req-b73330de-7630-444b-9674-b1882b4c4da9 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Deleted allocations for instance 579e1ca1-a6b6-4fde-8368-3e635a303032 [ 1833.754506] env[62824]: DEBUG nova.compute.manager [None req-61fb5137-5f87-475f-a5d2-0722cddc68b4 tempest-ServersTestJSON-928648138 tempest-ServersTestJSON-928648138-project-member] [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1833.880327] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Acquiring lock "04ac4ade-0bc8-4469-9b1b-f288b2f0a367" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1833.880327] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Lock "04ac4ade-0bc8-4469-9b1b-f288b2f0a367" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1834.011794] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f18aa27-a75d-404b-a42e-2de6f0b608d9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1834.017738] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9376d205-fa04-40b0-b74c-2becae960624 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1834.050960] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5cfe87b-8151-4cb0-a682-1406b57b736e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1834.058973] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02f05729-45af-4882-b55b-2cf36dec7a08 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1834.077077] env[62824]: DEBUG oslo_concurrency.lockutils [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Acquiring lock "82382932-7302-4441-a6f8-9aa2300ec0f6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1834.077451] env[62824]: DEBUG oslo_concurrency.lockutils [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Lock "82382932-7302-4441-a6f8-9aa2300ec0f6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1834.077783] env[62824]: DEBUG nova.compute.provider_tree [None req-7cd5dacb-b71b-429c-bfda-96302f774c53 tempest-ServersTestFqdnHostnames-528120407 tempest-ServersTestFqdnHostnames-528120407-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1834.101462] env[62824]: DEBUG nova.network.neutron [None req-2559566b-844a-4d61-b989-1f0a741aecd2 tempest-ServerDiagnosticsTest-1769488740 tempest-ServerDiagnosticsTest-1769488740-project-member] [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1834.152915] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b73330de-7630-444b-9674-b1882b4c4da9 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Lock "579e1ca1-a6b6-4fde-8368-3e635a303032" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.429s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1834.305885] env[62824]: DEBUG nova.network.neutron [None req-2559566b-844a-4d61-b989-1f0a741aecd2 tempest-ServerDiagnosticsTest-1769488740 tempest-ServerDiagnosticsTest-1769488740-project-member] [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1834.318812] env[62824]: DEBUG nova.network.neutron [None req-61fb5137-5f87-475f-a5d2-0722cddc68b4 tempest-ServersTestJSON-928648138 tempest-ServersTestJSON-928648138-project-member] [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] Successfully created port: e427511e-b0f8-42bd-b24f-0037a1ac6e76 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1834.584374] env[62824]: DEBUG nova.scheduler.client.report [None req-7cd5dacb-b71b-429c-bfda-96302f774c53 tempest-ServersTestFqdnHostnames-528120407 tempest-ServersTestFqdnHostnames-528120407-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1834.656060] env[62824]: DEBUG nova.compute.manager [None req-5cdb74c7-a27b-45c4-a89a-19c7714f04b5 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694-project-member] [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1834.772647] env[62824]: DEBUG nova.compute.manager [None req-61fb5137-5f87-475f-a5d2-0722cddc68b4 tempest-ServersTestJSON-928648138 tempest-ServersTestJSON-928648138-project-member] [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1834.811450] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2559566b-844a-4d61-b989-1f0a741aecd2 tempest-ServerDiagnosticsTest-1769488740 tempest-ServerDiagnosticsTest-1769488740-project-member] Releasing lock "refresh_cache-042c5f0f-bf1f-417e-be05-964cfcb20c7e" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1834.811984] env[62824]: DEBUG nova.compute.manager [None req-2559566b-844a-4d61-b989-1f0a741aecd2 tempest-ServerDiagnosticsTest-1769488740 tempest-ServerDiagnosticsTest-1769488740-project-member] [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1834.812235] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-2559566b-844a-4d61-b989-1f0a741aecd2 tempest-ServerDiagnosticsTest-1769488740 tempest-ServerDiagnosticsTest-1769488740-project-member] [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1834.814475] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-83b693a0-28f5-4152-b88f-dbfa93ee6464 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1834.824154] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-984d4fa0-d1f0-4adf-b570-e61ca9a6e778 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1834.840759] env[62824]: DEBUG nova.virt.hardware [None req-61fb5137-5f87-475f-a5d2-0722cddc68b4 tempest-ServersTestJSON-928648138 tempest-ServersTestJSON-928648138-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1834.841435] env[62824]: DEBUG nova.virt.hardware [None req-61fb5137-5f87-475f-a5d2-0722cddc68b4 tempest-ServersTestJSON-928648138 tempest-ServersTestJSON-928648138-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1834.841435] env[62824]: DEBUG nova.virt.hardware [None req-61fb5137-5f87-475f-a5d2-0722cddc68b4 tempest-ServersTestJSON-928648138 tempest-ServersTestJSON-928648138-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1834.841624] env[62824]: DEBUG nova.virt.hardware [None req-61fb5137-5f87-475f-a5d2-0722cddc68b4 tempest-ServersTestJSON-928648138 tempest-ServersTestJSON-928648138-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1834.841804] env[62824]: DEBUG nova.virt.hardware [None req-61fb5137-5f87-475f-a5d2-0722cddc68b4 tempest-ServersTestJSON-928648138 tempest-ServersTestJSON-928648138-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1834.841981] env[62824]: DEBUG nova.virt.hardware [None req-61fb5137-5f87-475f-a5d2-0722cddc68b4 tempest-ServersTestJSON-928648138 tempest-ServersTestJSON-928648138-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1834.842168] env[62824]: DEBUG nova.virt.hardware [None req-61fb5137-5f87-475f-a5d2-0722cddc68b4 tempest-ServersTestJSON-928648138 tempest-ServersTestJSON-928648138-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1834.842249] env[62824]: DEBUG nova.virt.hardware [None req-61fb5137-5f87-475f-a5d2-0722cddc68b4 tempest-ServersTestJSON-928648138 tempest-ServersTestJSON-928648138-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1834.842731] env[62824]: DEBUG nova.virt.hardware [None req-61fb5137-5f87-475f-a5d2-0722cddc68b4 tempest-ServersTestJSON-928648138 tempest-ServersTestJSON-928648138-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1834.842731] env[62824]: DEBUG nova.virt.hardware [None req-61fb5137-5f87-475f-a5d2-0722cddc68b4 tempest-ServersTestJSON-928648138 tempest-ServersTestJSON-928648138-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1834.842731] env[62824]: DEBUG nova.virt.hardware [None req-61fb5137-5f87-475f-a5d2-0722cddc68b4 tempest-ServersTestJSON-928648138 tempest-ServersTestJSON-928648138-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1834.844033] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69b54a79-b11b-4a32-ad92-24ab3f0ac21f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1834.860081] env[62824]: WARNING nova.virt.vmwareapi.vmops [None req-2559566b-844a-4d61-b989-1f0a741aecd2 tempest-ServerDiagnosticsTest-1769488740 tempest-ServerDiagnosticsTest-1769488740-project-member] [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 042c5f0f-bf1f-417e-be05-964cfcb20c7e could not be found. [ 1834.860290] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-2559566b-844a-4d61-b989-1f0a741aecd2 tempest-ServerDiagnosticsTest-1769488740 tempest-ServerDiagnosticsTest-1769488740-project-member] [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1834.860557] env[62824]: INFO nova.compute.manager [None req-2559566b-844a-4d61-b989-1f0a741aecd2 tempest-ServerDiagnosticsTest-1769488740 tempest-ServerDiagnosticsTest-1769488740-project-member] [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] Took 0.05 seconds to destroy the instance on the hypervisor. [ 1834.860860] env[62824]: DEBUG oslo.service.loopingcall [None req-2559566b-844a-4d61-b989-1f0a741aecd2 tempest-ServerDiagnosticsTest-1769488740 tempest-ServerDiagnosticsTest-1769488740-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1834.861516] env[62824]: DEBUG nova.compute.manager [-] [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1834.861673] env[62824]: DEBUG nova.network.neutron [-] [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1834.867128] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abaa8575-7166-4bec-b6de-a9509c2b944b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1834.916414] env[62824]: DEBUG nova.network.neutron [-] [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1835.089928] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7cd5dacb-b71b-429c-bfda-96302f774c53 tempest-ServersTestFqdnHostnames-528120407 tempest-ServersTestFqdnHostnames-528120407-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.358s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1835.091030] env[62824]: DEBUG nova.compute.manager [None req-7cd5dacb-b71b-429c-bfda-96302f774c53 tempest-ServersTestFqdnHostnames-528120407 tempest-ServersTestFqdnHostnames-528120407-project-member] [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1835.095386] env[62824]: DEBUG oslo_concurrency.lockutils [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.590s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1835.098476] env[62824]: INFO nova.compute.claims [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] [instance: f4157385-43ea-4e8c-887c-0985af37abd1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1835.181811] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5cdb74c7-a27b-45c4-a89a-19c7714f04b5 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1835.423357] env[62824]: DEBUG nova.network.neutron [-] [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1835.603456] env[62824]: DEBUG nova.compute.utils [None req-7cd5dacb-b71b-429c-bfda-96302f774c53 tempest-ServersTestFqdnHostnames-528120407 tempest-ServersTestFqdnHostnames-528120407-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1835.604877] env[62824]: DEBUG nova.compute.manager [None req-7cd5dacb-b71b-429c-bfda-96302f774c53 tempest-ServersTestFqdnHostnames-528120407 tempest-ServersTestFqdnHostnames-528120407-project-member] [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1835.606308] env[62824]: DEBUG nova.network.neutron [None req-7cd5dacb-b71b-429c-bfda-96302f774c53 tempest-ServersTestFqdnHostnames-528120407 tempest-ServersTestFqdnHostnames-528120407-project-member] [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1835.867098] env[62824]: DEBUG nova.policy [None req-7cd5dacb-b71b-429c-bfda-96302f774c53 tempest-ServersTestFqdnHostnames-528120407 tempest-ServersTestFqdnHostnames-528120407-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6acaf4ae557b4d55a2f8fca8e92d7ad3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6be7f0748ee0430496af5c6252c4f7fc', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 1835.884501] env[62824]: DEBUG nova.compute.manager [req-3c9e4fc2-f11b-475d-8cfa-d072daf32279 req-4b3ee11b-9315-4dd6-9b10-634e112b1c47 service nova] [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] Received event network-changed-3fe39149-045f-4e9f-8ab4-72da2377c013 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1835.884696] env[62824]: DEBUG nova.compute.manager [req-3c9e4fc2-f11b-475d-8cfa-d072daf32279 req-4b3ee11b-9315-4dd6-9b10-634e112b1c47 service nova] [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] Refreshing instance network info cache due to event network-changed-3fe39149-045f-4e9f-8ab4-72da2377c013. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1835.884922] env[62824]: DEBUG oslo_concurrency.lockutils [req-3c9e4fc2-f11b-475d-8cfa-d072daf32279 req-4b3ee11b-9315-4dd6-9b10-634e112b1c47 service nova] Acquiring lock "refresh_cache-042c5f0f-bf1f-417e-be05-964cfcb20c7e" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1835.886353] env[62824]: DEBUG oslo_concurrency.lockutils [req-3c9e4fc2-f11b-475d-8cfa-d072daf32279 req-4b3ee11b-9315-4dd6-9b10-634e112b1c47 service nova] Acquired lock "refresh_cache-042c5f0f-bf1f-417e-be05-964cfcb20c7e" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1835.886353] env[62824]: DEBUG nova.network.neutron [req-3c9e4fc2-f11b-475d-8cfa-d072daf32279 req-4b3ee11b-9315-4dd6-9b10-634e112b1c47 service nova] [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] Refreshing network info cache for port 3fe39149-045f-4e9f-8ab4-72da2377c013 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1835.926546] env[62824]: INFO nova.compute.manager [-] [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] Took 1.06 seconds to deallocate network for instance. [ 1835.928331] env[62824]: DEBUG nova.compute.claims [None req-2559566b-844a-4d61-b989-1f0a741aecd2 tempest-ServerDiagnosticsTest-1769488740 tempest-ServerDiagnosticsTest-1769488740-project-member] [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] Aborting claim: {{(pid=62824) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 1835.928699] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2559566b-844a-4d61-b989-1f0a741aecd2 tempest-ServerDiagnosticsTest-1769488740 tempest-ServerDiagnosticsTest-1769488740-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1836.070141] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquiring lock "f4d63a93-23af-470c-b36a-662af81dc386" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1836.070141] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lock "f4d63a93-23af-470c-b36a-662af81dc386" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1836.111573] env[62824]: DEBUG nova.compute.manager [None req-7cd5dacb-b71b-429c-bfda-96302f774c53 tempest-ServersTestFqdnHostnames-528120407 tempest-ServersTestFqdnHostnames-528120407-project-member] [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1836.368710] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6f25755-6996-4c14-9fa3-400028772ce3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1836.378270] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-630ae11a-f9f7-4f3f-b2bb-a329b22902e6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1836.415606] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45bfebc7-96aa-40fb-aa1d-7f2ea3950e28 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1836.424708] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19e5cc6d-fb40-498c-9aaf-2f6ded849191 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1836.440016] env[62824]: DEBUG nova.compute.provider_tree [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1836.485875] env[62824]: DEBUG nova.network.neutron [req-3c9e4fc2-f11b-475d-8cfa-d072daf32279 req-4b3ee11b-9315-4dd6-9b10-634e112b1c47 service nova] [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1836.700640] env[62824]: DEBUG nova.network.neutron [req-3c9e4fc2-f11b-475d-8cfa-d072daf32279 req-4b3ee11b-9315-4dd6-9b10-634e112b1c47 service nova] [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1836.944174] env[62824]: DEBUG nova.scheduler.client.report [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1837.079808] env[62824]: DEBUG nova.network.neutron [None req-7cd5dacb-b71b-429c-bfda-96302f774c53 tempest-ServersTestFqdnHostnames-528120407 tempest-ServersTestFqdnHostnames-528120407-project-member] [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] Successfully created port: 3a529f42-989d-46e2-ae3f-1e1a4fa18860 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1837.130745] env[62824]: DEBUG nova.compute.manager [None req-7cd5dacb-b71b-429c-bfda-96302f774c53 tempest-ServersTestFqdnHostnames-528120407 tempest-ServersTestFqdnHostnames-528120407-project-member] [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1837.160695] env[62824]: DEBUG nova.virt.hardware [None req-7cd5dacb-b71b-429c-bfda-96302f774c53 tempest-ServersTestFqdnHostnames-528120407 tempest-ServersTestFqdnHostnames-528120407-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1837.160991] env[62824]: DEBUG nova.virt.hardware [None req-7cd5dacb-b71b-429c-bfda-96302f774c53 tempest-ServersTestFqdnHostnames-528120407 tempest-ServersTestFqdnHostnames-528120407-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1837.161237] env[62824]: DEBUG nova.virt.hardware [None req-7cd5dacb-b71b-429c-bfda-96302f774c53 tempest-ServersTestFqdnHostnames-528120407 tempest-ServersTestFqdnHostnames-528120407-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1837.161566] env[62824]: DEBUG nova.virt.hardware [None req-7cd5dacb-b71b-429c-bfda-96302f774c53 tempest-ServersTestFqdnHostnames-528120407 tempest-ServersTestFqdnHostnames-528120407-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1837.161856] env[62824]: DEBUG nova.virt.hardware [None req-7cd5dacb-b71b-429c-bfda-96302f774c53 tempest-ServersTestFqdnHostnames-528120407 tempest-ServersTestFqdnHostnames-528120407-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1837.162181] env[62824]: DEBUG nova.virt.hardware [None req-7cd5dacb-b71b-429c-bfda-96302f774c53 tempest-ServersTestFqdnHostnames-528120407 tempest-ServersTestFqdnHostnames-528120407-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1837.162595] env[62824]: DEBUG nova.virt.hardware [None req-7cd5dacb-b71b-429c-bfda-96302f774c53 tempest-ServersTestFqdnHostnames-528120407 tempest-ServersTestFqdnHostnames-528120407-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1837.162942] env[62824]: DEBUG nova.virt.hardware [None req-7cd5dacb-b71b-429c-bfda-96302f774c53 tempest-ServersTestFqdnHostnames-528120407 tempest-ServersTestFqdnHostnames-528120407-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1837.163123] env[62824]: DEBUG nova.virt.hardware [None req-7cd5dacb-b71b-429c-bfda-96302f774c53 tempest-ServersTestFqdnHostnames-528120407 tempest-ServersTestFqdnHostnames-528120407-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1837.163340] env[62824]: DEBUG nova.virt.hardware [None req-7cd5dacb-b71b-429c-bfda-96302f774c53 tempest-ServersTestFqdnHostnames-528120407 tempest-ServersTestFqdnHostnames-528120407-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1837.163738] env[62824]: DEBUG nova.virt.hardware [None req-7cd5dacb-b71b-429c-bfda-96302f774c53 tempest-ServersTestFqdnHostnames-528120407 tempest-ServersTestFqdnHostnames-528120407-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1837.164870] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5767137f-79d1-4c58-a511-88e9c8c853a3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1837.174089] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-190d9c8b-5f52-4ee7-9bc9-90571dda90be {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1837.204268] env[62824]: DEBUG oslo_concurrency.lockutils [req-3c9e4fc2-f11b-475d-8cfa-d072daf32279 req-4b3ee11b-9315-4dd6-9b10-634e112b1c47 service nova] Releasing lock "refresh_cache-042c5f0f-bf1f-417e-be05-964cfcb20c7e" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1837.204530] env[62824]: DEBUG nova.compute.manager [req-3c9e4fc2-f11b-475d-8cfa-d072daf32279 req-4b3ee11b-9315-4dd6-9b10-634e112b1c47 service nova] [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] Received event network-vif-deleted-3fe39149-045f-4e9f-8ab4-72da2377c013 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1837.452353] env[62824]: DEBUG oslo_concurrency.lockutils [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.356s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1837.452353] env[62824]: DEBUG nova.compute.manager [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] [instance: f4157385-43ea-4e8c-887c-0985af37abd1] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1837.456389] env[62824]: DEBUG oslo_concurrency.lockutils [None req-24599d6d-2f76-445f-8b64-070ec4ee0da9 tempest-TenantUsagesTestJSON-486947106 tempest-TenantUsagesTestJSON-486947106-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 8.400s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1837.963026] env[62824]: DEBUG nova.compute.utils [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1837.966602] env[62824]: DEBUG nova.compute.manager [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] [instance: f4157385-43ea-4e8c-887c-0985af37abd1] Not allocating networking since 'none' was specified. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1984}} [ 1838.210743] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-794160c1-5758-466d-8e85-16f28ed598c1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1838.221349] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-887a6b22-62bb-4281-9e82-5e842bf76828 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1838.259225] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d050076-1902-492c-9c37-7f7a560fa695 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1838.269095] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e64b070-7f9c-4159-bfad-0c1bab14e820 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1838.288538] env[62824]: DEBUG nova.compute.provider_tree [None req-24599d6d-2f76-445f-8b64-070ec4ee0da9 tempest-TenantUsagesTestJSON-486947106 tempest-TenantUsagesTestJSON-486947106-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1838.468204] env[62824]: DEBUG nova.compute.manager [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] [instance: f4157385-43ea-4e8c-887c-0985af37abd1] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1838.792805] env[62824]: DEBUG nova.scheduler.client.report [None req-24599d6d-2f76-445f-8b64-070ec4ee0da9 tempest-TenantUsagesTestJSON-486947106 tempest-TenantUsagesTestJSON-486947106-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1838.850490] env[62824]: ERROR nova.compute.manager [None req-61fb5137-5f87-475f-a5d2-0722cddc68b4 tempest-ServersTestJSON-928648138 tempest-ServersTestJSON-928648138-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e427511e-b0f8-42bd-b24f-0037a1ac6e76, please check neutron logs for more information. [ 1838.850490] env[62824]: ERROR nova.compute.manager Traceback (most recent call last): [ 1838.850490] env[62824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1996, in _allocate_network_async [ 1838.850490] env[62824]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 1838.850490] env[62824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 1838.850490] env[62824]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 1838.850490] env[62824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 1838.850490] env[62824]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 1838.850490] env[62824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1838.850490] env[62824]: ERROR nova.compute.manager self.force_reraise() [ 1838.850490] env[62824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1838.850490] env[62824]: ERROR nova.compute.manager raise self.value [ 1838.850490] env[62824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 1838.850490] env[62824]: ERROR nova.compute.manager updated_port = self._update_port( [ 1838.850490] env[62824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 1838.850490] env[62824]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 1838.850979] env[62824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 1838.850979] env[62824]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 1838.850979] env[62824]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e427511e-b0f8-42bd-b24f-0037a1ac6e76, please check neutron logs for more information. [ 1838.850979] env[62824]: ERROR nova.compute.manager [ 1838.850979] env[62824]: Traceback (most recent call last): [ 1838.850979] env[62824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 1838.850979] env[62824]: listener.cb(fileno) [ 1838.850979] env[62824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 1838.850979] env[62824]: result = function(*args, **kwargs) [ 1838.850979] env[62824]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 1838.850979] env[62824]: return func(*args, **kwargs) [ 1838.850979] env[62824]: File "/opt/stack/nova/nova/compute/manager.py", line 2018, in _allocate_network_async [ 1838.850979] env[62824]: raise e [ 1838.850979] env[62824]: File "/opt/stack/nova/nova/compute/manager.py", line 1996, in _allocate_network_async [ 1838.850979] env[62824]: nwinfo = self.network_api.allocate_for_instance( [ 1838.850979] env[62824]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 1838.850979] env[62824]: created_port_ids = self._update_ports_for_instance( [ 1838.850979] env[62824]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 1838.850979] env[62824]: with excutils.save_and_reraise_exception(): [ 1838.850979] env[62824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1838.850979] env[62824]: self.force_reraise() [ 1838.850979] env[62824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1838.850979] env[62824]: raise self.value [ 1838.850979] env[62824]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 1838.850979] env[62824]: updated_port = self._update_port( [ 1838.850979] env[62824]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 1838.850979] env[62824]: _ensure_no_port_binding_failure(port) [ 1838.850979] env[62824]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 1838.850979] env[62824]: raise exception.PortBindingFailed(port_id=port['id']) [ 1838.851818] env[62824]: nova.exception.PortBindingFailed: Binding failed for port e427511e-b0f8-42bd-b24f-0037a1ac6e76, please check neutron logs for more information. [ 1838.851818] env[62824]: Removing descriptor: 18 [ 1838.851818] env[62824]: ERROR nova.compute.manager [None req-61fb5137-5f87-475f-a5d2-0722cddc68b4 tempest-ServersTestJSON-928648138 tempest-ServersTestJSON-928648138-project-member] [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e427511e-b0f8-42bd-b24f-0037a1ac6e76, please check neutron logs for more information. [ 1838.851818] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] Traceback (most recent call last): [ 1838.851818] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] File "/opt/stack/nova/nova/compute/manager.py", line 2901, in _build_resources [ 1838.851818] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] yield resources [ 1838.851818] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] File "/opt/stack/nova/nova/compute/manager.py", line 2648, in _build_and_run_instance [ 1838.851818] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] self.driver.spawn(context, instance, image_meta, [ 1838.851818] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 1838.851818] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] self._vmops.spawn(context, instance, image_meta, injected_files, [ 1838.851818] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 1838.851818] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] vm_ref = self.build_virtual_machine(instance, [ 1838.852338] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 1838.852338] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] vif_infos = vmwarevif.get_vif_info(self._session, [ 1838.852338] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 1838.852338] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] for vif in network_info: [ 1838.852338] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 1838.852338] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] return self._sync_wrapper(fn, *args, **kwargs) [ 1838.852338] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 1838.852338] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] self.wait() [ 1838.852338] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 1838.852338] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] self[:] = self._gt.wait() [ 1838.852338] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 1838.852338] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] return self._exit_event.wait() [ 1838.852338] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 1838.852734] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] result = hub.switch() [ 1838.852734] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 1838.852734] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] return self.greenlet.switch() [ 1838.852734] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 1838.852734] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] result = function(*args, **kwargs) [ 1838.852734] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 1838.852734] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] return func(*args, **kwargs) [ 1838.852734] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] File "/opt/stack/nova/nova/compute/manager.py", line 2018, in _allocate_network_async [ 1838.852734] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] raise e [ 1838.852734] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] File "/opt/stack/nova/nova/compute/manager.py", line 1996, in _allocate_network_async [ 1838.852734] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] nwinfo = self.network_api.allocate_for_instance( [ 1838.852734] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 1838.852734] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] created_port_ids = self._update_ports_for_instance( [ 1838.853400] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 1838.853400] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] with excutils.save_and_reraise_exception(): [ 1838.853400] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1838.853400] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] self.force_reraise() [ 1838.853400] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1838.853400] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] raise self.value [ 1838.853400] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 1838.853400] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] updated_port = self._update_port( [ 1838.853400] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 1838.853400] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] _ensure_no_port_binding_failure(port) [ 1838.853400] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 1838.853400] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] raise exception.PortBindingFailed(port_id=port['id']) [ 1838.854543] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] nova.exception.PortBindingFailed: Binding failed for port e427511e-b0f8-42bd-b24f-0037a1ac6e76, please check neutron logs for more information. [ 1838.854543] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] [ 1838.854543] env[62824]: INFO nova.compute.manager [None req-61fb5137-5f87-475f-a5d2-0722cddc68b4 tempest-ServersTestJSON-928648138 tempest-ServersTestJSON-928648138-project-member] [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] Terminating instance [ 1839.303922] env[62824]: DEBUG oslo_concurrency.lockutils [None req-24599d6d-2f76-445f-8b64-070ec4ee0da9 tempest-TenantUsagesTestJSON-486947106 tempest-TenantUsagesTestJSON-486947106-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.847s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1839.304812] env[62824]: ERROR nova.compute.manager [None req-24599d6d-2f76-445f-8b64-070ec4ee0da9 tempest-TenantUsagesTestJSON-486947106 tempest-TenantUsagesTestJSON-486947106-project-member] [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b42c0eea-348b-430a-8119-0a16fac3dc7f, please check neutron logs for more information. [ 1839.304812] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] Traceback (most recent call last): [ 1839.304812] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] File "/opt/stack/nova/nova/compute/manager.py", line 2648, in _build_and_run_instance [ 1839.304812] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] self.driver.spawn(context, instance, image_meta, [ 1839.304812] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 1839.304812] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] self._vmops.spawn(context, instance, image_meta, injected_files, [ 1839.304812] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 1839.304812] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] vm_ref = self.build_virtual_machine(instance, [ 1839.304812] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 1839.304812] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] vif_infos = vmwarevif.get_vif_info(self._session, [ 1839.304812] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 1839.305276] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] for vif in network_info: [ 1839.305276] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 1839.305276] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] return self._sync_wrapper(fn, *args, **kwargs) [ 1839.305276] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 1839.305276] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] self.wait() [ 1839.305276] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 1839.305276] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] self[:] = self._gt.wait() [ 1839.305276] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 1839.305276] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] return self._exit_event.wait() [ 1839.305276] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 1839.305276] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] result = hub.switch() [ 1839.305276] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 1839.305276] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] return self.greenlet.switch() [ 1839.305615] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 1839.305615] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] result = function(*args, **kwargs) [ 1839.305615] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 1839.305615] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] return func(*args, **kwargs) [ 1839.305615] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] File "/opt/stack/nova/nova/compute/manager.py", line 2018, in _allocate_network_async [ 1839.305615] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] raise e [ 1839.305615] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] File "/opt/stack/nova/nova/compute/manager.py", line 1996, in _allocate_network_async [ 1839.305615] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] nwinfo = self.network_api.allocate_for_instance( [ 1839.305615] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 1839.305615] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] created_port_ids = self._update_ports_for_instance( [ 1839.305615] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 1839.305615] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] with excutils.save_and_reraise_exception(): [ 1839.305615] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1839.305957] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] self.force_reraise() [ 1839.305957] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1839.305957] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] raise self.value [ 1839.305957] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 1839.305957] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] updated_port = self._update_port( [ 1839.305957] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 1839.305957] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] _ensure_no_port_binding_failure(port) [ 1839.305957] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 1839.305957] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] raise exception.PortBindingFailed(port_id=port['id']) [ 1839.305957] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] nova.exception.PortBindingFailed: Binding failed for port b42c0eea-348b-430a-8119-0a16fac3dc7f, please check neutron logs for more information. [ 1839.305957] env[62824]: ERROR nova.compute.manager [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] [ 1839.307382] env[62824]: DEBUG nova.compute.utils [None req-24599d6d-2f76-445f-8b64-070ec4ee0da9 tempest-TenantUsagesTestJSON-486947106 tempest-TenantUsagesTestJSON-486947106-project-member] [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] Binding failed for port b42c0eea-348b-430a-8119-0a16fac3dc7f, please check neutron logs for more information. {{(pid=62824) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 1839.307382] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9f666b55-9bba-45f0-98a3-b9d8b6261703 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.226s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1839.308809] env[62824]: INFO nova.compute.claims [None req-9f666b55-9bba-45f0-98a3-b9d8b6261703 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1839.313212] env[62824]: DEBUG nova.compute.manager [None req-24599d6d-2f76-445f-8b64-070ec4ee0da9 tempest-TenantUsagesTestJSON-486947106 tempest-TenantUsagesTestJSON-486947106-project-member] [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] Build of instance 9f016121-09f1-4da7-bbd0-7769d5104d27 was re-scheduled: Binding failed for port b42c0eea-348b-430a-8119-0a16fac3dc7f, please check neutron logs for more information. {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2487}} [ 1839.313658] env[62824]: DEBUG nova.compute.manager [None req-24599d6d-2f76-445f-8b64-070ec4ee0da9 tempest-TenantUsagesTestJSON-486947106 tempest-TenantUsagesTestJSON-486947106-project-member] [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] Unplugging VIFs for instance {{(pid=62824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3013}} [ 1839.315055] env[62824]: DEBUG oslo_concurrency.lockutils [None req-24599d6d-2f76-445f-8b64-070ec4ee0da9 tempest-TenantUsagesTestJSON-486947106 tempest-TenantUsagesTestJSON-486947106-project-member] Acquiring lock "refresh_cache-9f016121-09f1-4da7-bbd0-7769d5104d27" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1839.315055] env[62824]: DEBUG oslo_concurrency.lockutils [None req-24599d6d-2f76-445f-8b64-070ec4ee0da9 tempest-TenantUsagesTestJSON-486947106 tempest-TenantUsagesTestJSON-486947106-project-member] Acquired lock "refresh_cache-9f016121-09f1-4da7-bbd0-7769d5104d27" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1839.315055] env[62824]: DEBUG nova.network.neutron [None req-24599d6d-2f76-445f-8b64-070ec4ee0da9 tempest-TenantUsagesTestJSON-486947106 tempest-TenantUsagesTestJSON-486947106-project-member] [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1839.361529] env[62824]: DEBUG oslo_concurrency.lockutils [None req-61fb5137-5f87-475f-a5d2-0722cddc68b4 tempest-ServersTestJSON-928648138 tempest-ServersTestJSON-928648138-project-member] Acquiring lock "refresh_cache-0b845d80-c8b2-4b16-bda2-08fc78fb4337" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1839.361529] env[62824]: DEBUG oslo_concurrency.lockutils [None req-61fb5137-5f87-475f-a5d2-0722cddc68b4 tempest-ServersTestJSON-928648138 tempest-ServersTestJSON-928648138-project-member] Acquired lock "refresh_cache-0b845d80-c8b2-4b16-bda2-08fc78fb4337" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1839.361529] env[62824]: DEBUG nova.network.neutron [None req-61fb5137-5f87-475f-a5d2-0722cddc68b4 tempest-ServersTestJSON-928648138 tempest-ServersTestJSON-928648138-project-member] [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1839.412872] env[62824]: DEBUG oslo_concurrency.lockutils [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Acquiring lock "9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1839.412872] env[62824]: DEBUG oslo_concurrency.lockutils [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Lock "9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1839.486898] env[62824]: DEBUG nova.compute.manager [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] [instance: f4157385-43ea-4e8c-887c-0985af37abd1] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1839.519466] env[62824]: DEBUG nova.virt.hardware [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1839.519913] env[62824]: DEBUG nova.virt.hardware [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1839.520241] env[62824]: DEBUG nova.virt.hardware [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1839.520661] env[62824]: DEBUG nova.virt.hardware [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1839.520928] env[62824]: DEBUG nova.virt.hardware [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1839.521201] env[62824]: DEBUG nova.virt.hardware [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1839.521751] env[62824]: DEBUG nova.virt.hardware [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1839.522077] env[62824]: DEBUG nova.virt.hardware [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1839.522377] env[62824]: DEBUG nova.virt.hardware [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1839.522644] env[62824]: DEBUG nova.virt.hardware [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1839.522939] env[62824]: DEBUG nova.virt.hardware [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1839.523927] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-428bf847-c42c-4e1d-ad95-3fb8199e7067 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1839.534567] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-911bea37-612c-4560-b6d7-67095e899783 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1839.550965] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] [instance: f4157385-43ea-4e8c-887c-0985af37abd1] Instance VIF info [] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1839.568663] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Creating folder: OpenStack. Parent ref: group-v4. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1839.569451] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d3e55814-9e60-4769-ba5c-5764908c19d5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1839.583341] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Created folder: OpenStack in parent group-v4. [ 1839.583341] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Creating folder: Project (ccaed923f1e34905a6af6b75e0cb4676). Parent ref: group-v438503. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1839.583341] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f3f8ec05-99d8-418b-9144-6b8ee3d0c9da {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1839.595307] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Created folder: Project (ccaed923f1e34905a6af6b75e0cb4676) in parent group-v438503. [ 1839.595307] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Creating folder: Instances. Parent ref: group-v438504. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1839.595307] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fbca72ab-c489-4c9d-9187-71bf640fe784 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1839.609477] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Created folder: Instances in parent group-v438504. [ 1839.609477] env[62824]: DEBUG oslo.service.loopingcall [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1839.609477] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f4157385-43ea-4e8c-887c-0985af37abd1] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1839.609477] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-93620ddb-419a-4ac8-b535-ead222d61004 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1839.625521] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1839.625521] env[62824]: value = "task-2145614" [ 1839.625521] env[62824]: _type = "Task" [ 1839.625521] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1839.634619] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145614, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1839.871210] env[62824]: DEBUG nova.network.neutron [None req-24599d6d-2f76-445f-8b64-070ec4ee0da9 tempest-TenantUsagesTestJSON-486947106 tempest-TenantUsagesTestJSON-486947106-project-member] [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1839.902805] env[62824]: DEBUG nova.network.neutron [None req-61fb5137-5f87-475f-a5d2-0722cddc68b4 tempest-ServersTestJSON-928648138 tempest-ServersTestJSON-928648138-project-member] [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1840.067777] env[62824]: DEBUG nova.network.neutron [None req-24599d6d-2f76-445f-8b64-070ec4ee0da9 tempest-TenantUsagesTestJSON-486947106 tempest-TenantUsagesTestJSON-486947106-project-member] [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1840.141097] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145614, 'name': CreateVM_Task, 'duration_secs': 0.355954} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1840.143018] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f4157385-43ea-4e8c-887c-0985af37abd1] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1840.143018] env[62824]: DEBUG oslo_vmware.service [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-612ad671-9f6f-478f-8ddc-cb75cc718d1d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1840.152691] env[62824]: DEBUG oslo_concurrency.lockutils [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1840.152691] env[62824]: DEBUG oslo_concurrency.lockutils [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1840.152691] env[62824]: DEBUG oslo_concurrency.lockutils [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1840.152691] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e68d8ebd-0855-441e-a373-4c2525608589 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1840.158517] env[62824]: DEBUG nova.compute.manager [req-98cef897-bddd-46f3-96c4-2d7c4623e992 req-ca88693f-6e14-42db-bceb-fc49aed65f68 service nova] [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] Received event network-changed-e427511e-b0f8-42bd-b24f-0037a1ac6e76 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1840.159710] env[62824]: DEBUG nova.compute.manager [req-98cef897-bddd-46f3-96c4-2d7c4623e992 req-ca88693f-6e14-42db-bceb-fc49aed65f68 service nova] [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] Refreshing instance network info cache due to event network-changed-e427511e-b0f8-42bd-b24f-0037a1ac6e76. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1840.159710] env[62824]: DEBUG oslo_concurrency.lockutils [req-98cef897-bddd-46f3-96c4-2d7c4623e992 req-ca88693f-6e14-42db-bceb-fc49aed65f68 service nova] Acquiring lock "refresh_cache-0b845d80-c8b2-4b16-bda2-08fc78fb4337" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1840.164277] env[62824]: DEBUG oslo_vmware.api [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Waiting for the task: (returnval){ [ 1840.164277] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52f0f801-0cb7-5d49-bd9b-793bbed69d48" [ 1840.164277] env[62824]: _type = "Task" [ 1840.164277] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1840.172209] env[62824]: DEBUG oslo_vmware.api [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52f0f801-0cb7-5d49-bd9b-793bbed69d48, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1840.211228] env[62824]: DEBUG nova.network.neutron [None req-61fb5137-5f87-475f-a5d2-0722cddc68b4 tempest-ServersTestJSON-928648138 tempest-ServersTestJSON-928648138-project-member] [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1840.573740] env[62824]: DEBUG oslo_concurrency.lockutils [None req-24599d6d-2f76-445f-8b64-070ec4ee0da9 tempest-TenantUsagesTestJSON-486947106 tempest-TenantUsagesTestJSON-486947106-project-member] Releasing lock "refresh_cache-9f016121-09f1-4da7-bbd0-7769d5104d27" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1840.573740] env[62824]: DEBUG nova.compute.manager [None req-24599d6d-2f76-445f-8b64-070ec4ee0da9 tempest-TenantUsagesTestJSON-486947106 tempest-TenantUsagesTestJSON-486947106-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3036}} [ 1840.573740] env[62824]: DEBUG nova.compute.manager [None req-24599d6d-2f76-445f-8b64-070ec4ee0da9 tempest-TenantUsagesTestJSON-486947106 tempest-TenantUsagesTestJSON-486947106-project-member] [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1840.574242] env[62824]: DEBUG nova.network.neutron [None req-24599d6d-2f76-445f-8b64-070ec4ee0da9 tempest-TenantUsagesTestJSON-486947106 tempest-TenantUsagesTestJSON-486947106-project-member] [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1840.613836] env[62824]: DEBUG nova.network.neutron [None req-24599d6d-2f76-445f-8b64-070ec4ee0da9 tempest-TenantUsagesTestJSON-486947106 tempest-TenantUsagesTestJSON-486947106-project-member] [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1840.657025] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89887b6a-ad63-433f-9c51-5c53f3baa5d7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1840.674021] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-725266dc-806a-459c-93cd-678f24171200 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1840.679768] env[62824]: DEBUG oslo_concurrency.lockutils [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1840.680145] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] [instance: f4157385-43ea-4e8c-887c-0985af37abd1] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1840.680260] env[62824]: DEBUG oslo_concurrency.lockutils [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1840.680399] env[62824]: DEBUG oslo_concurrency.lockutils [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1840.680811] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1840.681769] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-22c3f214-a81a-4b6d-a980-c29d8ae23783 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1840.713705] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86b33cf4-2061-4cde-a2a2-aa799281d901 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1840.717143] env[62824]: DEBUG oslo_concurrency.lockutils [None req-61fb5137-5f87-475f-a5d2-0722cddc68b4 tempest-ServersTestJSON-928648138 tempest-ServersTestJSON-928648138-project-member] Releasing lock "refresh_cache-0b845d80-c8b2-4b16-bda2-08fc78fb4337" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1840.717574] env[62824]: DEBUG nova.compute.manager [None req-61fb5137-5f87-475f-a5d2-0722cddc68b4 tempest-ServersTestJSON-928648138 tempest-ServersTestJSON-928648138-project-member] [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1840.718091] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-61fb5137-5f87-475f-a5d2-0722cddc68b4 tempest-ServersTestJSON-928648138 tempest-ServersTestJSON-928648138-project-member] [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1840.718091] env[62824]: DEBUG oslo_concurrency.lockutils [req-98cef897-bddd-46f3-96c4-2d7c4623e992 req-ca88693f-6e14-42db-bceb-fc49aed65f68 service nova] Acquired lock "refresh_cache-0b845d80-c8b2-4b16-bda2-08fc78fb4337" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1840.718210] env[62824]: DEBUG nova.network.neutron [req-98cef897-bddd-46f3-96c4-2d7c4623e992 req-ca88693f-6e14-42db-bceb-fc49aed65f68 service nova] [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] Refreshing network info cache for port e427511e-b0f8-42bd-b24f-0037a1ac6e76 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1840.719299] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-db39ce50-fe17-4109-9cd5-471b9665d8b0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1840.729195] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d59de945-7f7f-4c8f-941b-be26b3796ebe {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1840.733435] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1840.733612] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1840.737233] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aca78549-798d-4fda-96ac-0e7a4a8ee823 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1840.754935] env[62824]: DEBUG nova.compute.provider_tree [None req-9f666b55-9bba-45f0-98a3-b9d8b6261703 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1840.759398] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-acc399b1-1ad2-4c63-8e3e-6a6cfb382753 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1840.764682] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69e95bcc-5b68-4663-9d0e-5c95f949a95e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1840.780904] env[62824]: DEBUG oslo_vmware.api [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Waiting for the task: (returnval){ [ 1840.780904] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52f1e59f-2366-4ec4-fb2e-d5a527d7cb9d" [ 1840.780904] env[62824]: _type = "Task" [ 1840.780904] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1840.793798] env[62824]: WARNING nova.virt.vmwareapi.vmops [None req-61fb5137-5f87-475f-a5d2-0722cddc68b4 tempest-ServersTestJSON-928648138 tempest-ServersTestJSON-928648138-project-member] [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0b845d80-c8b2-4b16-bda2-08fc78fb4337 could not be found. [ 1840.793798] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-61fb5137-5f87-475f-a5d2-0722cddc68b4 tempest-ServersTestJSON-928648138 tempest-ServersTestJSON-928648138-project-member] [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1840.793798] env[62824]: INFO nova.compute.manager [None req-61fb5137-5f87-475f-a5d2-0722cddc68b4 tempest-ServersTestJSON-928648138 tempest-ServersTestJSON-928648138-project-member] [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] Took 0.08 seconds to destroy the instance on the hypervisor. [ 1840.794154] env[62824]: DEBUG oslo.service.loopingcall [None req-61fb5137-5f87-475f-a5d2-0722cddc68b4 tempest-ServersTestJSON-928648138 tempest-ServersTestJSON-928648138-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1840.794865] env[62824]: DEBUG nova.compute.manager [-] [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1840.795020] env[62824]: DEBUG nova.network.neutron [-] [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1840.800881] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] [instance: f4157385-43ea-4e8c-887c-0985af37abd1] Preparing fetch location {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1840.801147] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Creating directory with path [datastore2] vmware_temp/d4eedb2e-bde4-4c73-9b6b-6eff5761f7ab/9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1840.801307] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-72027c9a-1c8c-437a-b372-a760024ec592 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1840.825864] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Created directory with path [datastore2] vmware_temp/d4eedb2e-bde4-4c73-9b6b-6eff5761f7ab/9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1840.826089] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] [instance: f4157385-43ea-4e8c-887c-0985af37abd1] Fetch image to [datastore2] vmware_temp/d4eedb2e-bde4-4c73-9b6b-6eff5761f7ab/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/tmp-sparse.vmdk {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1840.826358] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] [instance: f4157385-43ea-4e8c-887c-0985af37abd1] Downloading image file data 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 to [datastore2] vmware_temp/d4eedb2e-bde4-4c73-9b6b-6eff5761f7ab/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/tmp-sparse.vmdk on the data store datastore2 {{(pid=62824) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 1840.831637] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f97300ec-4c58-4e31-af49-fbcc72286f19 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1840.839293] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-896ed8b0-3a3b-4cb2-a958-ec62d9f4e086 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1840.850175] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a5e6e03-ea1a-45a1-890a-406b61a8ca80 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1840.885970] env[62824]: DEBUG nova.network.neutron [-] [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1840.891023] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94ae3416-ad51-4705-b9d1-586bdfbe13a9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1840.896585] env[62824]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-4ddec5de-4c15-429e-82f4-2f0e652e3d09 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1840.990515] env[62824]: DEBUG nova.virt.vmwareapi.images [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] [instance: f4157385-43ea-4e8c-887c-0985af37abd1] Downloading image file data 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 to the data store datastore2 {{(pid=62824) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 1841.079187] env[62824]: DEBUG oslo_vmware.rw_handles [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c2n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/d4eedb2e-bde4-4c73-9b6b-6eff5761f7ab/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=62824) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1841.146720] env[62824]: DEBUG nova.network.neutron [None req-24599d6d-2f76-445f-8b64-070ec4ee0da9 tempest-TenantUsagesTestJSON-486947106 tempest-TenantUsagesTestJSON-486947106-project-member] [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1841.260551] env[62824]: DEBUG nova.scheduler.client.report [None req-9f666b55-9bba-45f0-98a3-b9d8b6261703 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1841.280145] env[62824]: DEBUG nova.network.neutron [req-98cef897-bddd-46f3-96c4-2d7c4623e992 req-ca88693f-6e14-42db-bceb-fc49aed65f68 service nova] [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1841.391067] env[62824]: ERROR nova.compute.manager [None req-7cd5dacb-b71b-429c-bfda-96302f774c53 tempest-ServersTestFqdnHostnames-528120407 tempest-ServersTestFqdnHostnames-528120407-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3a529f42-989d-46e2-ae3f-1e1a4fa18860, please check neutron logs for more information. [ 1841.391067] env[62824]: ERROR nova.compute.manager Traceback (most recent call last): [ 1841.391067] env[62824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1996, in _allocate_network_async [ 1841.391067] env[62824]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 1841.391067] env[62824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 1841.391067] env[62824]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 1841.391067] env[62824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 1841.391067] env[62824]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 1841.391067] env[62824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1841.391067] env[62824]: ERROR nova.compute.manager self.force_reraise() [ 1841.391067] env[62824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1841.391067] env[62824]: ERROR nova.compute.manager raise self.value [ 1841.391067] env[62824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 1841.391067] env[62824]: ERROR nova.compute.manager updated_port = self._update_port( [ 1841.391067] env[62824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 1841.391067] env[62824]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 1841.391742] env[62824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 1841.391742] env[62824]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 1841.391742] env[62824]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3a529f42-989d-46e2-ae3f-1e1a4fa18860, please check neutron logs for more information. [ 1841.391742] env[62824]: ERROR nova.compute.manager [ 1841.391742] env[62824]: Traceback (most recent call last): [ 1841.391742] env[62824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 1841.391742] env[62824]: listener.cb(fileno) [ 1841.391742] env[62824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 1841.391742] env[62824]: result = function(*args, **kwargs) [ 1841.391742] env[62824]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 1841.391742] env[62824]: return func(*args, **kwargs) [ 1841.391742] env[62824]: File "/opt/stack/nova/nova/compute/manager.py", line 2018, in _allocate_network_async [ 1841.391742] env[62824]: raise e [ 1841.391742] env[62824]: File "/opt/stack/nova/nova/compute/manager.py", line 1996, in _allocate_network_async [ 1841.391742] env[62824]: nwinfo = self.network_api.allocate_for_instance( [ 1841.391742] env[62824]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 1841.391742] env[62824]: created_port_ids = self._update_ports_for_instance( [ 1841.391742] env[62824]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 1841.391742] env[62824]: with excutils.save_and_reraise_exception(): [ 1841.391742] env[62824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1841.391742] env[62824]: self.force_reraise() [ 1841.391742] env[62824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1841.391742] env[62824]: raise self.value [ 1841.391742] env[62824]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 1841.391742] env[62824]: updated_port = self._update_port( [ 1841.391742] env[62824]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 1841.391742] env[62824]: _ensure_no_port_binding_failure(port) [ 1841.391742] env[62824]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 1841.391742] env[62824]: raise exception.PortBindingFailed(port_id=port['id']) [ 1841.394232] env[62824]: nova.exception.PortBindingFailed: Binding failed for port 3a529f42-989d-46e2-ae3f-1e1a4fa18860, please check neutron logs for more information. [ 1841.394232] env[62824]: Removing descriptor: 17 [ 1841.394232] env[62824]: ERROR nova.compute.manager [None req-7cd5dacb-b71b-429c-bfda-96302f774c53 tempest-ServersTestFqdnHostnames-528120407 tempest-ServersTestFqdnHostnames-528120407-project-member] [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3a529f42-989d-46e2-ae3f-1e1a4fa18860, please check neutron logs for more information. [ 1841.394232] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] Traceback (most recent call last): [ 1841.394232] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] File "/opt/stack/nova/nova/compute/manager.py", line 2901, in _build_resources [ 1841.394232] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] yield resources [ 1841.394232] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] File "/opt/stack/nova/nova/compute/manager.py", line 2648, in _build_and_run_instance [ 1841.394232] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] self.driver.spawn(context, instance, image_meta, [ 1841.394232] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 1841.394232] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] self._vmops.spawn(context, instance, image_meta, injected_files, [ 1841.394232] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 1841.394232] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] vm_ref = self.build_virtual_machine(instance, [ 1841.394774] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 1841.394774] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] vif_infos = vmwarevif.get_vif_info(self._session, [ 1841.394774] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 1841.394774] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] for vif in network_info: [ 1841.394774] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 1841.394774] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] return self._sync_wrapper(fn, *args, **kwargs) [ 1841.394774] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 1841.394774] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] self.wait() [ 1841.394774] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 1841.394774] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] self[:] = self._gt.wait() [ 1841.394774] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 1841.394774] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] return self._exit_event.wait() [ 1841.394774] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 1841.395184] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] result = hub.switch() [ 1841.395184] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 1841.395184] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] return self.greenlet.switch() [ 1841.395184] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 1841.395184] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] result = function(*args, **kwargs) [ 1841.395184] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 1841.395184] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] return func(*args, **kwargs) [ 1841.395184] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] File "/opt/stack/nova/nova/compute/manager.py", line 2018, in _allocate_network_async [ 1841.395184] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] raise e [ 1841.395184] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] File "/opt/stack/nova/nova/compute/manager.py", line 1996, in _allocate_network_async [ 1841.395184] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] nwinfo = self.network_api.allocate_for_instance( [ 1841.395184] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 1841.395184] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] created_port_ids = self._update_ports_for_instance( [ 1841.395635] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 1841.395635] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] with excutils.save_and_reraise_exception(): [ 1841.395635] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1841.395635] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] self.force_reraise() [ 1841.395635] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1841.395635] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] raise self.value [ 1841.395635] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 1841.395635] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] updated_port = self._update_port( [ 1841.395635] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 1841.395635] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] _ensure_no_port_binding_failure(port) [ 1841.395635] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 1841.395635] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] raise exception.PortBindingFailed(port_id=port['id']) [ 1841.395979] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] nova.exception.PortBindingFailed: Binding failed for port 3a529f42-989d-46e2-ae3f-1e1a4fa18860, please check neutron logs for more information. [ 1841.395979] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] [ 1841.395979] env[62824]: INFO nova.compute.manager [None req-7cd5dacb-b71b-429c-bfda-96302f774c53 tempest-ServersTestFqdnHostnames-528120407 tempest-ServersTestFqdnHostnames-528120407-project-member] [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] Terminating instance [ 1841.399785] env[62824]: DEBUG nova.network.neutron [-] [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1841.586221] env[62824]: DEBUG nova.network.neutron [req-98cef897-bddd-46f3-96c4-2d7c4623e992 req-ca88693f-6e14-42db-bceb-fc49aed65f68 service nova] [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1841.655250] env[62824]: INFO nova.compute.manager [None req-24599d6d-2f76-445f-8b64-070ec4ee0da9 tempest-TenantUsagesTestJSON-486947106 tempest-TenantUsagesTestJSON-486947106-project-member] [instance: 9f016121-09f1-4da7-bbd0-7769d5104d27] Took 1.08 seconds to deallocate network for instance. [ 1841.755467] env[62824]: DEBUG oslo_vmware.rw_handles [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Completed reading data from the image iterator. {{(pid=62824) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1841.755678] env[62824]: DEBUG oslo_vmware.rw_handles [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Closing write handle for https://esx7c2n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/d4eedb2e-bde4-4c73-9b6b-6eff5761f7ab/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=62824) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 1841.767906] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9f666b55-9bba-45f0-98a3-b9d8b6261703 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.460s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1841.768499] env[62824]: DEBUG nova.compute.manager [None req-9f666b55-9bba-45f0-98a3-b9d8b6261703 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1841.770964] env[62824]: DEBUG oslo_concurrency.lockutils [None req-250af368-8463-49f4-8a2e-de088db419fa tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 11.994s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1841.903260] env[62824]: DEBUG nova.virt.vmwareapi.images [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] [instance: f4157385-43ea-4e8c-887c-0985af37abd1] Downloaded image file data 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 to vmware_temp/d4eedb2e-bde4-4c73-9b6b-6eff5761f7ab/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/tmp-sparse.vmdk on the data store datastore2 {{(pid=62824) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 1841.904805] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] [instance: f4157385-43ea-4e8c-887c-0985af37abd1] Caching image {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1841.905114] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Copying Virtual Disk [datastore2] vmware_temp/d4eedb2e-bde4-4c73-9b6b-6eff5761f7ab/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/tmp-sparse.vmdk to [datastore2] vmware_temp/d4eedb2e-bde4-4c73-9b6b-6eff5761f7ab/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1841.905408] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-93498406-ff8e-4211-a686-bfbd0b9ffac1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1841.908893] env[62824]: INFO nova.compute.manager [-] [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] Took 1.11 seconds to deallocate network for instance. [ 1841.911254] env[62824]: DEBUG nova.compute.claims [None req-61fb5137-5f87-475f-a5d2-0722cddc68b4 tempest-ServersTestJSON-928648138 tempest-ServersTestJSON-928648138-project-member] [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] Aborting claim: {{(pid=62824) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 1841.911454] env[62824]: DEBUG oslo_concurrency.lockutils [None req-61fb5137-5f87-475f-a5d2-0722cddc68b4 tempest-ServersTestJSON-928648138 tempest-ServersTestJSON-928648138-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1841.911926] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7cd5dacb-b71b-429c-bfda-96302f774c53 tempest-ServersTestFqdnHostnames-528120407 tempest-ServersTestFqdnHostnames-528120407-project-member] Acquiring lock "refresh_cache-27e6ff4d-c4f2-41e7-92c2-a0e2445830bf" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1841.912083] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7cd5dacb-b71b-429c-bfda-96302f774c53 tempest-ServersTestFqdnHostnames-528120407 tempest-ServersTestFqdnHostnames-528120407-project-member] Acquired lock "refresh_cache-27e6ff4d-c4f2-41e7-92c2-a0e2445830bf" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1841.912269] env[62824]: DEBUG nova.network.neutron [None req-7cd5dacb-b71b-429c-bfda-96302f774c53 tempest-ServersTestFqdnHostnames-528120407 tempest-ServersTestFqdnHostnames-528120407-project-member] [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1841.914610] env[62824]: DEBUG oslo_vmware.api [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Waiting for the task: (returnval){ [ 1841.914610] env[62824]: value = "task-2145615" [ 1841.914610] env[62824]: _type = "Task" [ 1841.914610] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1841.924322] env[62824]: DEBUG oslo_vmware.api [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Task: {'id': task-2145615, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1842.089534] env[62824]: DEBUG oslo_concurrency.lockutils [req-98cef897-bddd-46f3-96c4-2d7c4623e992 req-ca88693f-6e14-42db-bceb-fc49aed65f68 service nova] Releasing lock "refresh_cache-0b845d80-c8b2-4b16-bda2-08fc78fb4337" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1842.151295] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Acquiring lock "517bf2cf-e142-4f36-bf53-79a8d2d8a530" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1842.151582] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Lock "517bf2cf-e142-4f36-bf53-79a8d2d8a530" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1842.281826] env[62824]: DEBUG nova.compute.utils [None req-9f666b55-9bba-45f0-98a3-b9d8b6261703 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1842.287773] env[62824]: DEBUG nova.compute.manager [None req-9f666b55-9bba-45f0-98a3-b9d8b6261703 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1842.288850] env[62824]: DEBUG nova.network.neutron [None req-9f666b55-9bba-45f0-98a3-b9d8b6261703 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1842.431725] env[62824]: DEBUG oslo_vmware.api [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Task: {'id': task-2145615, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1842.506413] env[62824]: DEBUG nova.policy [None req-9f666b55-9bba-45f0-98a3-b9d8b6261703 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c1978c29e9d54f73bf6fb9505ddf76d0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4cb963b19c1549ab8a11e5a77e4f202d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 1842.597023] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27bff184-f110-41e4-8b53-1a17671b4a7d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1842.605088] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46911700-4476-42b6-b1fe-3315b45ba75e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1842.643571] env[62824]: DEBUG nova.network.neutron [None req-7cd5dacb-b71b-429c-bfda-96302f774c53 tempest-ServersTestFqdnHostnames-528120407 tempest-ServersTestFqdnHostnames-528120407-project-member] [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1842.647026] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a996bf73-5ced-42ab-b518-7dae318cee62 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1842.654826] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf931dd5-240b-4a82-9dfd-8a068848e4ec {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1842.676244] env[62824]: DEBUG nova.compute.provider_tree [None req-250af368-8463-49f4-8a2e-de088db419fa tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1842.707160] env[62824]: INFO nova.scheduler.client.report [None req-24599d6d-2f76-445f-8b64-070ec4ee0da9 tempest-TenantUsagesTestJSON-486947106 tempest-TenantUsagesTestJSON-486947106-project-member] Deleted allocations for instance 9f016121-09f1-4da7-bbd0-7769d5104d27 [ 1842.788017] env[62824]: DEBUG nova.compute.manager [None req-9f666b55-9bba-45f0-98a3-b9d8b6261703 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1842.929328] env[62824]: DEBUG oslo_vmware.api [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Task: {'id': task-2145615, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.670995} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1842.929676] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Copied Virtual Disk [datastore2] vmware_temp/d4eedb2e-bde4-4c73-9b6b-6eff5761f7ab/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/tmp-sparse.vmdk to [datastore2] vmware_temp/d4eedb2e-bde4-4c73-9b6b-6eff5761f7ab/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1842.930126] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Deleting the datastore file [datastore2] vmware_temp/d4eedb2e-bde4-4c73-9b6b-6eff5761f7ab/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/tmp-sparse.vmdk {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1842.930461] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-95c06516-45da-4d04-ab31-f791c4f0dc84 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1842.943614] env[62824]: DEBUG oslo_vmware.api [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Waiting for the task: (returnval){ [ 1842.943614] env[62824]: value = "task-2145616" [ 1842.943614] env[62824]: _type = "Task" [ 1842.943614] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1842.952141] env[62824]: DEBUG oslo_vmware.api [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Task: {'id': task-2145616, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1843.089815] env[62824]: DEBUG nova.network.neutron [None req-7cd5dacb-b71b-429c-bfda-96302f774c53 tempest-ServersTestFqdnHostnames-528120407 tempest-ServersTestFqdnHostnames-528120407-project-member] [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1843.182717] env[62824]: DEBUG nova.scheduler.client.report [None req-250af368-8463-49f4-8a2e-de088db419fa tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1843.216838] env[62824]: DEBUG oslo_concurrency.lockutils [None req-24599d6d-2f76-445f-8b64-070ec4ee0da9 tempest-TenantUsagesTestJSON-486947106 tempest-TenantUsagesTestJSON-486947106-project-member] Lock "9f016121-09f1-4da7-bbd0-7769d5104d27" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.702s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1843.274377] env[62824]: DEBUG nova.network.neutron [None req-9f666b55-9bba-45f0-98a3-b9d8b6261703 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] Successfully created port: 2bf4c281-010d-4311-be12-2ec99b845218 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1843.465437] env[62824]: DEBUG oslo_vmware.api [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Task: {'id': task-2145616, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.02455} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1843.465733] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1843.465941] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Moving file from [datastore2] vmware_temp/d4eedb2e-bde4-4c73-9b6b-6eff5761f7ab/9e2a7d30-212d-4ab8-9606-c5c6d52629e8 to [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8. {{(pid=62824) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 1843.466226] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-88304d95-1391-41d1-8594-1e6614f85d68 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1843.479318] env[62824]: DEBUG oslo_vmware.api [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Waiting for the task: (returnval){ [ 1843.479318] env[62824]: value = "task-2145617" [ 1843.479318] env[62824]: _type = "Task" [ 1843.479318] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1843.488873] env[62824]: DEBUG oslo_vmware.api [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Task: {'id': task-2145617, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1843.594745] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7cd5dacb-b71b-429c-bfda-96302f774c53 tempest-ServersTestFqdnHostnames-528120407 tempest-ServersTestFqdnHostnames-528120407-project-member] Releasing lock "refresh_cache-27e6ff4d-c4f2-41e7-92c2-a0e2445830bf" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1843.597597] env[62824]: DEBUG nova.compute.manager [None req-7cd5dacb-b71b-429c-bfda-96302f774c53 tempest-ServersTestFqdnHostnames-528120407 tempest-ServersTestFqdnHostnames-528120407-project-member] [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1843.597597] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-7cd5dacb-b71b-429c-bfda-96302f774c53 tempest-ServersTestFqdnHostnames-528120407 tempest-ServersTestFqdnHostnames-528120407-project-member] [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1843.597597] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-308a31f0-5807-40fd-bd94-7e63edc593ce {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1843.610849] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e44a570-ed82-491d-8f0e-63d23008e223 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1843.645052] env[62824]: WARNING nova.virt.vmwareapi.vmops [None req-7cd5dacb-b71b-429c-bfda-96302f774c53 tempest-ServersTestFqdnHostnames-528120407 tempest-ServersTestFqdnHostnames-528120407-project-member] [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf could not be found. [ 1843.645464] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-7cd5dacb-b71b-429c-bfda-96302f774c53 tempest-ServersTestFqdnHostnames-528120407 tempest-ServersTestFqdnHostnames-528120407-project-member] [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1843.645775] env[62824]: INFO nova.compute.manager [None req-7cd5dacb-b71b-429c-bfda-96302f774c53 tempest-ServersTestFqdnHostnames-528120407 tempest-ServersTestFqdnHostnames-528120407-project-member] [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] Took 0.05 seconds to destroy the instance on the hypervisor. [ 1843.646147] env[62824]: DEBUG oslo.service.loopingcall [None req-7cd5dacb-b71b-429c-bfda-96302f774c53 tempest-ServersTestFqdnHostnames-528120407 tempest-ServersTestFqdnHostnames-528120407-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1843.646489] env[62824]: DEBUG nova.compute.manager [-] [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1843.646875] env[62824]: DEBUG nova.network.neutron [-] [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1843.688504] env[62824]: DEBUG oslo_concurrency.lockutils [None req-250af368-8463-49f4-8a2e-de088db419fa tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.917s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1843.691248] env[62824]: ERROR nova.compute.manager [None req-250af368-8463-49f4-8a2e-de088db419fa tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 2a8b04db-a377-45bb-9e8f-39327bbfffce, please check neutron logs for more information. [ 1843.691248] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] Traceback (most recent call last): [ 1843.691248] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] File "/opt/stack/nova/nova/compute/manager.py", line 2648, in _build_and_run_instance [ 1843.691248] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] self.driver.spawn(context, instance, image_meta, [ 1843.691248] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 1843.691248] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 1843.691248] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 1843.691248] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] vm_ref = self.build_virtual_machine(instance, [ 1843.691248] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 1843.691248] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] vif_infos = vmwarevif.get_vif_info(self._session, [ 1843.691248] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 1843.691616] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] for vif in network_info: [ 1843.691616] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 1843.691616] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] return self._sync_wrapper(fn, *args, **kwargs) [ 1843.691616] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 1843.691616] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] self.wait() [ 1843.691616] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 1843.691616] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] self[:] = self._gt.wait() [ 1843.691616] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 1843.691616] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] return self._exit_event.wait() [ 1843.691616] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 1843.691616] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] result = hub.switch() [ 1843.691616] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 1843.691616] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] return self.greenlet.switch() [ 1843.693331] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 1843.693331] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] result = function(*args, **kwargs) [ 1843.693331] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 1843.693331] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] return func(*args, **kwargs) [ 1843.693331] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] File "/opt/stack/nova/nova/compute/manager.py", line 2018, in _allocate_network_async [ 1843.693331] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] raise e [ 1843.693331] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] File "/opt/stack/nova/nova/compute/manager.py", line 1996, in _allocate_network_async [ 1843.693331] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] nwinfo = self.network_api.allocate_for_instance( [ 1843.693331] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 1843.693331] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] created_port_ids = self._update_ports_for_instance( [ 1843.693331] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 1843.693331] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] with excutils.save_and_reraise_exception(): [ 1843.693331] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1843.693854] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] self.force_reraise() [ 1843.693854] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1843.693854] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] raise self.value [ 1843.693854] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 1843.693854] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] updated_port = self._update_port( [ 1843.693854] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 1843.693854] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] _ensure_no_port_binding_failure(port) [ 1843.693854] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 1843.693854] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] raise exception.PortBindingFailed(port_id=port['id']) [ 1843.693854] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] nova.exception.PortBindingFailed: Binding failed for port 2a8b04db-a377-45bb-9e8f-39327bbfffce, please check neutron logs for more information. [ 1843.693854] env[62824]: ERROR nova.compute.manager [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] [ 1843.694240] env[62824]: DEBUG nova.compute.utils [None req-250af368-8463-49f4-8a2e-de088db419fa tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] Binding failed for port 2a8b04db-a377-45bb-9e8f-39327bbfffce, please check neutron logs for more information. {{(pid=62824) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 1843.694463] env[62824]: DEBUG nova.compute.manager [None req-250af368-8463-49f4-8a2e-de088db419fa tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] Build of instance 8b138cc7-8bde-48c8-b1bb-cec6ead877d3 was re-scheduled: Binding failed for port 2a8b04db-a377-45bb-9e8f-39327bbfffce, please check neutron logs for more information. {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2487}} [ 1843.697379] env[62824]: DEBUG nova.compute.manager [None req-250af368-8463-49f4-8a2e-de088db419fa tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] Unplugging VIFs for instance {{(pid=62824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3013}} [ 1843.697825] env[62824]: DEBUG oslo_concurrency.lockutils [None req-250af368-8463-49f4-8a2e-de088db419fa tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Acquiring lock "refresh_cache-8b138cc7-8bde-48c8-b1bb-cec6ead877d3" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1843.697948] env[62824]: DEBUG oslo_concurrency.lockutils [None req-250af368-8463-49f4-8a2e-de088db419fa tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Acquired lock "refresh_cache-8b138cc7-8bde-48c8-b1bb-cec6ead877d3" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1843.698586] env[62824]: DEBUG nova.network.neutron [None req-250af368-8463-49f4-8a2e-de088db419fa tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1843.700273] env[62824]: DEBUG oslo_concurrency.lockutils [None req-6070f146-69b5-4aa1-b292-f106b5e9564b tempest-ServerAddressesNegativeTestJSON-1575558569 tempest-ServerAddressesNegativeTestJSON-1575558569-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.819s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1843.720482] env[62824]: DEBUG nova.compute.manager [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1843.743517] env[62824]: DEBUG nova.network.neutron [-] [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1843.801090] env[62824]: DEBUG nova.compute.manager [None req-9f666b55-9bba-45f0-98a3-b9d8b6261703 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1843.847062] env[62824]: DEBUG nova.compute.manager [req-cfbd09cd-0850-4454-8f01-220740c7e437 req-378d4b71-357c-4800-b068-4002c4d1ac50 service nova] [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] Received event network-vif-deleted-e427511e-b0f8-42bd-b24f-0037a1ac6e76 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1843.847278] env[62824]: DEBUG nova.compute.manager [req-cfbd09cd-0850-4454-8f01-220740c7e437 req-378d4b71-357c-4800-b068-4002c4d1ac50 service nova] [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] Received event network-changed-3a529f42-989d-46e2-ae3f-1e1a4fa18860 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1843.847566] env[62824]: DEBUG nova.compute.manager [req-cfbd09cd-0850-4454-8f01-220740c7e437 req-378d4b71-357c-4800-b068-4002c4d1ac50 service nova] [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] Refreshing instance network info cache due to event network-changed-3a529f42-989d-46e2-ae3f-1e1a4fa18860. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1843.847627] env[62824]: DEBUG oslo_concurrency.lockutils [req-cfbd09cd-0850-4454-8f01-220740c7e437 req-378d4b71-357c-4800-b068-4002c4d1ac50 service nova] Acquiring lock "refresh_cache-27e6ff4d-c4f2-41e7-92c2-a0e2445830bf" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1843.847766] env[62824]: DEBUG oslo_concurrency.lockutils [req-cfbd09cd-0850-4454-8f01-220740c7e437 req-378d4b71-357c-4800-b068-4002c4d1ac50 service nova] Acquired lock "refresh_cache-27e6ff4d-c4f2-41e7-92c2-a0e2445830bf" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1843.847920] env[62824]: DEBUG nova.network.neutron [req-cfbd09cd-0850-4454-8f01-220740c7e437 req-378d4b71-357c-4800-b068-4002c4d1ac50 service nova] [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] Refreshing network info cache for port 3a529f42-989d-46e2-ae3f-1e1a4fa18860 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1843.855776] env[62824]: DEBUG nova.virt.hardware [None req-9f666b55-9bba-45f0-98a3-b9d8b6261703 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1843.856019] env[62824]: DEBUG nova.virt.hardware [None req-9f666b55-9bba-45f0-98a3-b9d8b6261703 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1843.856491] env[62824]: DEBUG nova.virt.hardware [None req-9f666b55-9bba-45f0-98a3-b9d8b6261703 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1843.856630] env[62824]: DEBUG nova.virt.hardware [None req-9f666b55-9bba-45f0-98a3-b9d8b6261703 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1843.856855] env[62824]: DEBUG nova.virt.hardware [None req-9f666b55-9bba-45f0-98a3-b9d8b6261703 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1843.856855] env[62824]: DEBUG nova.virt.hardware [None req-9f666b55-9bba-45f0-98a3-b9d8b6261703 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1843.857189] env[62824]: DEBUG nova.virt.hardware [None req-9f666b55-9bba-45f0-98a3-b9d8b6261703 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1843.857229] env[62824]: DEBUG nova.virt.hardware [None req-9f666b55-9bba-45f0-98a3-b9d8b6261703 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1843.857787] env[62824]: DEBUG nova.virt.hardware [None req-9f666b55-9bba-45f0-98a3-b9d8b6261703 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1843.857787] env[62824]: DEBUG nova.virt.hardware [None req-9f666b55-9bba-45f0-98a3-b9d8b6261703 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1843.857787] env[62824]: DEBUG nova.virt.hardware [None req-9f666b55-9bba-45f0-98a3-b9d8b6261703 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1843.860575] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fb93acf-ea34-4a7c-a4d8-ca00273f7445 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1843.874693] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ac035f8-a3b4-4348-8a7f-1a0b56828274 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1843.975948] env[62824]: DEBUG oslo_concurrency.lockutils [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Acquiring lock "9062e606-511f-4fe7-9621-90b3c0e51eec" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1843.976202] env[62824]: DEBUG oslo_concurrency.lockutils [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Lock "9062e606-511f-4fe7-9621-90b3c0e51eec" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1843.986977] env[62824]: DEBUG oslo_vmware.api [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Task: {'id': task-2145617, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.023839} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1843.987546] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] File moved {{(pid=62824) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 1843.987744] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] [instance: f4157385-43ea-4e8c-887c-0985af37abd1] Cleaning up location [datastore2] vmware_temp/d4eedb2e-bde4-4c73-9b6b-6eff5761f7ab {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1843.987904] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Deleting the datastore file [datastore2] vmware_temp/d4eedb2e-bde4-4c73-9b6b-6eff5761f7ab {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1843.988231] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1ed6e35f-4a6d-47c6-84e6-0ec289e30d0d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1843.995754] env[62824]: DEBUG oslo_vmware.api [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Waiting for the task: (returnval){ [ 1843.995754] env[62824]: value = "task-2145618" [ 1843.995754] env[62824]: _type = "Task" [ 1843.995754] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1844.009999] env[62824]: DEBUG oslo_vmware.api [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Task: {'id': task-2145618, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1844.091965] env[62824]: DEBUG nova.network.neutron [None req-9f666b55-9bba-45f0-98a3-b9d8b6261703 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] Successfully created port: 688abfdf-8f0c-42ff-b158-a6c0793892c2 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1844.253891] env[62824]: DEBUG nova.network.neutron [-] [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1844.256096] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1844.268280] env[62824]: DEBUG nova.network.neutron [None req-250af368-8463-49f4-8a2e-de088db419fa tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1844.486886] env[62824]: DEBUG nova.network.neutron [req-cfbd09cd-0850-4454-8f01-220740c7e437 req-378d4b71-357c-4800-b068-4002c4d1ac50 service nova] [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1844.513417] env[62824]: DEBUG oslo_vmware.api [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Task: {'id': task-2145618, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.025632} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1844.514350] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1844.514639] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b233ce98-0e82-4974-88f4-1bea990228ac {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1844.520365] env[62824]: DEBUG oslo_vmware.api [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Waiting for the task: (returnval){ [ 1844.520365] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52230ce8-aaba-e47d-c766-fe086921d3f4" [ 1844.520365] env[62824]: _type = "Task" [ 1844.520365] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1844.533209] env[62824]: DEBUG oslo_vmware.api [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52230ce8-aaba-e47d-c766-fe086921d3f4, 'name': SearchDatastore_Task, 'duration_secs': 0.009808} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1844.533584] env[62824]: DEBUG oslo_concurrency.lockutils [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1844.533786] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] f4157385-43ea-4e8c-887c-0985af37abd1/f4157385-43ea-4e8c-887c-0985af37abd1.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1844.534127] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3c96d151-49a2-4944-a512-d65898a08f40 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1844.544686] env[62824]: DEBUG oslo_vmware.api [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Waiting for the task: (returnval){ [ 1844.544686] env[62824]: value = "task-2145623" [ 1844.544686] env[62824]: _type = "Task" [ 1844.544686] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1844.554968] env[62824]: DEBUG oslo_vmware.api [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Task: {'id': task-2145623, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1844.556671] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b7147a2-063a-4f94-9f2b-5d0efb4e6f47 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1844.564703] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d561813-8607-4903-981a-645d179d5a79 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1844.595664] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ece61c48-57e0-42a2-ac94-b7401005ba12 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1844.603755] env[62824]: DEBUG nova.network.neutron [None req-250af368-8463-49f4-8a2e-de088db419fa tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1844.606226] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2facf136-a376-4890-9c05-2d4ba2c76c8e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1844.628437] env[62824]: DEBUG nova.compute.provider_tree [None req-6070f146-69b5-4aa1-b292-f106b5e9564b tempest-ServerAddressesNegativeTestJSON-1575558569 tempest-ServerAddressesNegativeTestJSON-1575558569-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1844.728543] env[62824]: DEBUG nova.network.neutron [None req-9f666b55-9bba-45f0-98a3-b9d8b6261703 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] Successfully created port: 48731347-3ac3-4e3a-98b4-b8f5520584e9 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1844.757519] env[62824]: INFO nova.compute.manager [-] [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] Took 1.11 seconds to deallocate network for instance. [ 1844.760224] env[62824]: DEBUG nova.compute.claims [None req-7cd5dacb-b71b-429c-bfda-96302f774c53 tempest-ServersTestFqdnHostnames-528120407 tempest-ServersTestFqdnHostnames-528120407-project-member] [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] Aborting claim: {{(pid=62824) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 1844.760420] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7cd5dacb-b71b-429c-bfda-96302f774c53 tempest-ServersTestFqdnHostnames-528120407 tempest-ServersTestFqdnHostnames-528120407-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1844.968478] env[62824]: DEBUG nova.network.neutron [req-cfbd09cd-0850-4454-8f01-220740c7e437 req-378d4b71-357c-4800-b068-4002c4d1ac50 service nova] [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1845.058635] env[62824]: DEBUG oslo_vmware.api [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Task: {'id': task-2145623, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1845.116016] env[62824]: DEBUG oslo_concurrency.lockutils [None req-250af368-8463-49f4-8a2e-de088db419fa tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Releasing lock "refresh_cache-8b138cc7-8bde-48c8-b1bb-cec6ead877d3" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1845.116016] env[62824]: DEBUG nova.compute.manager [None req-250af368-8463-49f4-8a2e-de088db419fa tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3036}} [ 1845.116016] env[62824]: DEBUG nova.compute.manager [None req-250af368-8463-49f4-8a2e-de088db419fa tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1845.116016] env[62824]: DEBUG nova.network.neutron [None req-250af368-8463-49f4-8a2e-de088db419fa tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1845.135033] env[62824]: DEBUG nova.scheduler.client.report [None req-6070f146-69b5-4aa1-b292-f106b5e9564b tempest-ServerAddressesNegativeTestJSON-1575558569 tempest-ServerAddressesNegativeTestJSON-1575558569-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1845.148210] env[62824]: DEBUG nova.network.neutron [None req-250af368-8463-49f4-8a2e-de088db419fa tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1845.474374] env[62824]: DEBUG oslo_concurrency.lockutils [req-cfbd09cd-0850-4454-8f01-220740c7e437 req-378d4b71-357c-4800-b068-4002c4d1ac50 service nova] Releasing lock "refresh_cache-27e6ff4d-c4f2-41e7-92c2-a0e2445830bf" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1845.474374] env[62824]: DEBUG nova.compute.manager [req-cfbd09cd-0850-4454-8f01-220740c7e437 req-378d4b71-357c-4800-b068-4002c4d1ac50 service nova] [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] Received event network-vif-deleted-3a529f42-989d-46e2-ae3f-1e1a4fa18860 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1845.556753] env[62824]: DEBUG oslo_vmware.api [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Task: {'id': task-2145623, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1845.641631] env[62824]: DEBUG oslo_concurrency.lockutils [None req-6070f146-69b5-4aa1-b292-f106b5e9564b tempest-ServerAddressesNegativeTestJSON-1575558569 tempest-ServerAddressesNegativeTestJSON-1575558569-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.941s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1845.643034] env[62824]: ERROR nova.compute.manager [None req-6070f146-69b5-4aa1-b292-f106b5e9564b tempest-ServerAddressesNegativeTestJSON-1575558569 tempest-ServerAddressesNegativeTestJSON-1575558569-project-member] [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 59ec24da-a159-4775-a06b-bf6a02869bed, please check neutron logs for more information. [ 1845.643034] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] Traceback (most recent call last): [ 1845.643034] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] File "/opt/stack/nova/nova/compute/manager.py", line 2648, in _build_and_run_instance [ 1845.643034] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] self.driver.spawn(context, instance, image_meta, [ 1845.643034] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 1845.643034] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 1845.643034] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 1845.643034] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] vm_ref = self.build_virtual_machine(instance, [ 1845.643034] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 1845.643034] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] vif_infos = vmwarevif.get_vif_info(self._session, [ 1845.643034] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 1845.643729] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] for vif in network_info: [ 1845.643729] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 1845.643729] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] return self._sync_wrapper(fn, *args, **kwargs) [ 1845.643729] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 1845.643729] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] self.wait() [ 1845.643729] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 1845.643729] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] self[:] = self._gt.wait() [ 1845.643729] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 1845.643729] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] return self._exit_event.wait() [ 1845.643729] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 1845.643729] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] result = hub.switch() [ 1845.643729] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 1845.643729] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] return self.greenlet.switch() [ 1845.644133] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 1845.644133] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] result = function(*args, **kwargs) [ 1845.644133] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 1845.644133] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] return func(*args, **kwargs) [ 1845.644133] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] File "/opt/stack/nova/nova/compute/manager.py", line 2018, in _allocate_network_async [ 1845.644133] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] raise e [ 1845.644133] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] File "/opt/stack/nova/nova/compute/manager.py", line 1996, in _allocate_network_async [ 1845.644133] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] nwinfo = self.network_api.allocate_for_instance( [ 1845.644133] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 1845.644133] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] created_port_ids = self._update_ports_for_instance( [ 1845.644133] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 1845.644133] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] with excutils.save_and_reraise_exception(): [ 1845.644133] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1845.645375] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] self.force_reraise() [ 1845.645375] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1845.645375] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] raise self.value [ 1845.645375] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 1845.645375] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] updated_port = self._update_port( [ 1845.645375] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 1845.645375] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] _ensure_no_port_binding_failure(port) [ 1845.645375] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 1845.645375] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] raise exception.PortBindingFailed(port_id=port['id']) [ 1845.645375] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] nova.exception.PortBindingFailed: Binding failed for port 59ec24da-a159-4775-a06b-bf6a02869bed, please check neutron logs for more information. [ 1845.645375] env[62824]: ERROR nova.compute.manager [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] [ 1845.645911] env[62824]: DEBUG nova.compute.utils [None req-6070f146-69b5-4aa1-b292-f106b5e9564b tempest-ServerAddressesNegativeTestJSON-1575558569 tempest-ServerAddressesNegativeTestJSON-1575558569-project-member] [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] Binding failed for port 59ec24da-a159-4775-a06b-bf6a02869bed, please check neutron logs for more information. {{(pid=62824) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 1845.645911] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 15.370s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1845.645911] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1845.645911] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62824) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1845.645911] env[62824]: DEBUG oslo_concurrency.lockutils [None req-34e3c185-2b6d-4b05-bb0f-3253c603dd1c tempest-InstanceActionsV221TestJSON-1308314300 tempest-InstanceActionsV221TestJSON-1308314300-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.355s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1845.647228] env[62824]: INFO nova.compute.claims [None req-34e3c185-2b6d-4b05-bb0f-3253c603dd1c tempest-InstanceActionsV221TestJSON-1308314300 tempest-InstanceActionsV221TestJSON-1308314300-project-member] [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1845.654887] env[62824]: DEBUG nova.compute.manager [None req-6070f146-69b5-4aa1-b292-f106b5e9564b tempest-ServerAddressesNegativeTestJSON-1575558569 tempest-ServerAddressesNegativeTestJSON-1575558569-project-member] [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] Build of instance 6cc60e74-0474-47a5-ba57-c1c646ce0e1d was re-scheduled: Binding failed for port 59ec24da-a159-4775-a06b-bf6a02869bed, please check neutron logs for more information. {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2487}} [ 1845.655452] env[62824]: DEBUG nova.compute.manager [None req-6070f146-69b5-4aa1-b292-f106b5e9564b tempest-ServerAddressesNegativeTestJSON-1575558569 tempest-ServerAddressesNegativeTestJSON-1575558569-project-member] [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] Unplugging VIFs for instance {{(pid=62824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3013}} [ 1845.656251] env[62824]: DEBUG oslo_concurrency.lockutils [None req-6070f146-69b5-4aa1-b292-f106b5e9564b tempest-ServerAddressesNegativeTestJSON-1575558569 tempest-ServerAddressesNegativeTestJSON-1575558569-project-member] Acquiring lock "refresh_cache-6cc60e74-0474-47a5-ba57-c1c646ce0e1d" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1845.656251] env[62824]: DEBUG oslo_concurrency.lockutils [None req-6070f146-69b5-4aa1-b292-f106b5e9564b tempest-ServerAddressesNegativeTestJSON-1575558569 tempest-ServerAddressesNegativeTestJSON-1575558569-project-member] Acquired lock "refresh_cache-6cc60e74-0474-47a5-ba57-c1c646ce0e1d" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1845.656251] env[62824]: DEBUG nova.network.neutron [None req-6070f146-69b5-4aa1-b292-f106b5e9564b tempest-ServerAddressesNegativeTestJSON-1575558569 tempest-ServerAddressesNegativeTestJSON-1575558569-project-member] [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1845.658644] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-093000c1-9aa4-4517-ac8e-732700753833 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1845.662806] env[62824]: DEBUG nova.network.neutron [None req-250af368-8463-49f4-8a2e-de088db419fa tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1845.669906] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8777b403-4546-4859-9d75-c50007e6c413 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1845.697266] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-931ad023-0d73-46ba-a07c-bd8527913ac3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1845.706814] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-016a397d-51b8-47dc-add2-27c052c6db50 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1845.735799] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181297MB free_disk=175GB free_vcpus=48 pci_devices=None {{(pid=62824) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1845.736057] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1846.062319] env[62824]: DEBUG oslo_vmware.api [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Task: {'id': task-2145623, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.495387} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1846.062658] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] f4157385-43ea-4e8c-887c-0985af37abd1/f4157385-43ea-4e8c-887c-0985af37abd1.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1846.062872] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] [instance: f4157385-43ea-4e8c-887c-0985af37abd1] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1846.063154] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-26bf2997-56ab-4b3c-983f-c168a6c3aa1c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1846.070268] env[62824]: DEBUG oslo_vmware.api [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Waiting for the task: (returnval){ [ 1846.070268] env[62824]: value = "task-2145624" [ 1846.070268] env[62824]: _type = "Task" [ 1846.070268] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1846.080773] env[62824]: DEBUG oslo_vmware.api [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Task: {'id': task-2145624, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1846.167442] env[62824]: INFO nova.compute.manager [None req-250af368-8463-49f4-8a2e-de088db419fa tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: 8b138cc7-8bde-48c8-b1bb-cec6ead877d3] Took 1.05 seconds to deallocate network for instance. [ 1846.206248] env[62824]: DEBUG nova.network.neutron [None req-6070f146-69b5-4aa1-b292-f106b5e9564b tempest-ServerAddressesNegativeTestJSON-1575558569 tempest-ServerAddressesNegativeTestJSON-1575558569-project-member] [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1846.581427] env[62824]: DEBUG oslo_vmware.api [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Task: {'id': task-2145624, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074951} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1846.581733] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] [instance: f4157385-43ea-4e8c-887c-0985af37abd1] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1846.582764] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc766130-d9fc-4c4d-9f14-7f4287f3086f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1846.610091] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] [instance: f4157385-43ea-4e8c-887c-0985af37abd1] Reconfiguring VM instance instance-00000009 to attach disk [datastore2] f4157385-43ea-4e8c-887c-0985af37abd1/f4157385-43ea-4e8c-887c-0985af37abd1.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1846.610091] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8cedc5a9-8b26-4b22-b376-b7c569e8786a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1846.635726] env[62824]: DEBUG nova.network.neutron [None req-6070f146-69b5-4aa1-b292-f106b5e9564b tempest-ServerAddressesNegativeTestJSON-1575558569 tempest-ServerAddressesNegativeTestJSON-1575558569-project-member] [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1846.644989] env[62824]: DEBUG oslo_vmware.api [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Waiting for the task: (returnval){ [ 1846.644989] env[62824]: value = "task-2145625" [ 1846.644989] env[62824]: _type = "Task" [ 1846.644989] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1846.663672] env[62824]: DEBUG oslo_vmware.api [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Task: {'id': task-2145625, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1847.084869] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb9dfa2f-8b99-4ed7-8745-42fb58776cde {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1847.103478] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4188513a-bf8c-4ad5-864a-bd1a0e42302e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1847.115447] env[62824]: DEBUG oslo_concurrency.lockutils [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Acquiring lock "384a96ea-40ce-43e1-a5f9-82f50b710b0b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1847.115653] env[62824]: DEBUG oslo_concurrency.lockutils [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Lock "384a96ea-40ce-43e1-a5f9-82f50b710b0b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1847.159669] env[62824]: DEBUG oslo_concurrency.lockutils [None req-6070f146-69b5-4aa1-b292-f106b5e9564b tempest-ServerAddressesNegativeTestJSON-1575558569 tempest-ServerAddressesNegativeTestJSON-1575558569-project-member] Releasing lock "refresh_cache-6cc60e74-0474-47a5-ba57-c1c646ce0e1d" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1847.159899] env[62824]: DEBUG nova.compute.manager [None req-6070f146-69b5-4aa1-b292-f106b5e9564b tempest-ServerAddressesNegativeTestJSON-1575558569 tempest-ServerAddressesNegativeTestJSON-1575558569-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3036}} [ 1847.160125] env[62824]: DEBUG nova.compute.manager [None req-6070f146-69b5-4aa1-b292-f106b5e9564b tempest-ServerAddressesNegativeTestJSON-1575558569 tempest-ServerAddressesNegativeTestJSON-1575558569-project-member] [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1847.160285] env[62824]: DEBUG nova.network.neutron [None req-6070f146-69b5-4aa1-b292-f106b5e9564b tempest-ServerAddressesNegativeTestJSON-1575558569 tempest-ServerAddressesNegativeTestJSON-1575558569-project-member] [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1847.167276] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29b81f19-97b3-4997-acbd-fb004a762656 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1847.177264] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e77fddea-8be6-4266-b538-d01f40aa5942 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1847.186549] env[62824]: DEBUG oslo_vmware.api [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Task: {'id': task-2145625, 'name': ReconfigVM_Task, 'duration_secs': 0.269075} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1847.186549] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] [instance: f4157385-43ea-4e8c-887c-0985af37abd1] Reconfigured VM instance instance-00000009 to attach disk [datastore2] f4157385-43ea-4e8c-887c-0985af37abd1/f4157385-43ea-4e8c-887c-0985af37abd1.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1847.186549] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dc6bd99c-7ad6-4b5c-b005-8c77c027bac0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1847.197319] env[62824]: DEBUG nova.compute.provider_tree [None req-34e3c185-2b6d-4b05-bb0f-3253c603dd1c tempest-InstanceActionsV221TestJSON-1308314300 tempest-InstanceActionsV221TestJSON-1308314300-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1847.201116] env[62824]: DEBUG oslo_vmware.api [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Waiting for the task: (returnval){ [ 1847.201116] env[62824]: value = "task-2145626" [ 1847.201116] env[62824]: _type = "Task" [ 1847.201116] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1847.213531] env[62824]: DEBUG oslo_vmware.api [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Task: {'id': task-2145626, 'name': Rename_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1847.214456] env[62824]: DEBUG nova.network.neutron [None req-6070f146-69b5-4aa1-b292-f106b5e9564b tempest-ServerAddressesNegativeTestJSON-1575558569 tempest-ServerAddressesNegativeTestJSON-1575558569-project-member] [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1847.235683] env[62824]: INFO nova.scheduler.client.report [None req-250af368-8463-49f4-8a2e-de088db419fa tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Deleted allocations for instance 8b138cc7-8bde-48c8-b1bb-cec6ead877d3 [ 1847.708186] env[62824]: DEBUG nova.scheduler.client.report [None req-34e3c185-2b6d-4b05-bb0f-3253c603dd1c tempest-InstanceActionsV221TestJSON-1308314300 tempest-InstanceActionsV221TestJSON-1308314300-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1847.717242] env[62824]: DEBUG nova.network.neutron [None req-6070f146-69b5-4aa1-b292-f106b5e9564b tempest-ServerAddressesNegativeTestJSON-1575558569 tempest-ServerAddressesNegativeTestJSON-1575558569-project-member] [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1847.724291] env[62824]: DEBUG oslo_vmware.api [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Task: {'id': task-2145626, 'name': Rename_Task, 'duration_secs': 0.255219} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1847.724656] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] [instance: f4157385-43ea-4e8c-887c-0985af37abd1] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1847.724904] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c01a6ca5-31c5-4121-84be-856a61826616 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1847.731928] env[62824]: DEBUG oslo_vmware.api [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Waiting for the task: (returnval){ [ 1847.731928] env[62824]: value = "task-2145628" [ 1847.731928] env[62824]: _type = "Task" [ 1847.731928] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1847.745155] env[62824]: DEBUG oslo_concurrency.lockutils [None req-250af368-8463-49f4-8a2e-de088db419fa tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Lock "8b138cc7-8bde-48c8-b1bb-cec6ead877d3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.835s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1847.745155] env[62824]: DEBUG oslo_vmware.api [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Task: {'id': task-2145628, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1848.201527] env[62824]: DEBUG nova.compute.manager [req-8bebdb83-decd-4d84-86c9-2b255a5c8a05 req-7e467813-f627-4fda-b212-82d0ce9b8161 service nova] [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] Received event network-changed-2bf4c281-010d-4311-be12-2ec99b845218 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1848.201726] env[62824]: DEBUG nova.compute.manager [req-8bebdb83-decd-4d84-86c9-2b255a5c8a05 req-7e467813-f627-4fda-b212-82d0ce9b8161 service nova] [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] Refreshing instance network info cache due to event network-changed-2bf4c281-010d-4311-be12-2ec99b845218. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1848.201981] env[62824]: DEBUG oslo_concurrency.lockutils [req-8bebdb83-decd-4d84-86c9-2b255a5c8a05 req-7e467813-f627-4fda-b212-82d0ce9b8161 service nova] Acquiring lock "refresh_cache-00fca8eb-c209-4632-abe6-5f092b04ac56" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1848.202146] env[62824]: DEBUG oslo_concurrency.lockutils [req-8bebdb83-decd-4d84-86c9-2b255a5c8a05 req-7e467813-f627-4fda-b212-82d0ce9b8161 service nova] Acquired lock "refresh_cache-00fca8eb-c209-4632-abe6-5f092b04ac56" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1848.202315] env[62824]: DEBUG nova.network.neutron [req-8bebdb83-decd-4d84-86c9-2b255a5c8a05 req-7e467813-f627-4fda-b212-82d0ce9b8161 service nova] [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] Refreshing network info cache for port 2bf4c281-010d-4311-be12-2ec99b845218 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1848.219453] env[62824]: DEBUG oslo_concurrency.lockutils [None req-34e3c185-2b6d-4b05-bb0f-3253c603dd1c tempest-InstanceActionsV221TestJSON-1308314300 tempest-InstanceActionsV221TestJSON-1308314300-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.574s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1848.220037] env[62824]: DEBUG nova.compute.manager [None req-34e3c185-2b6d-4b05-bb0f-3253c603dd1c tempest-InstanceActionsV221TestJSON-1308314300 tempest-InstanceActionsV221TestJSON-1308314300-project-member] [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1848.223094] env[62824]: INFO nova.compute.manager [None req-6070f146-69b5-4aa1-b292-f106b5e9564b tempest-ServerAddressesNegativeTestJSON-1575558569 tempest-ServerAddressesNegativeTestJSON-1575558569-project-member] [instance: 6cc60e74-0474-47a5-ba57-c1c646ce0e1d] Took 1.06 seconds to deallocate network for instance. [ 1848.226529] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5cdb74c7-a27b-45c4-a89a-19c7714f04b5 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.045s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1848.227985] env[62824]: INFO nova.compute.claims [None req-5cdb74c7-a27b-45c4-a89a-19c7714f04b5 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694-project-member] [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1848.247155] env[62824]: DEBUG oslo_vmware.api [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Task: {'id': task-2145628, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1848.248257] env[62824]: DEBUG nova.compute.manager [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] [instance: 04ac4ade-0bc8-4469-9b1b-f288b2f0a367] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1848.265566] env[62824]: ERROR nova.compute.manager [None req-9f666b55-9bba-45f0-98a3-b9d8b6261703 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 2bf4c281-010d-4311-be12-2ec99b845218, please check neutron logs for more information. [ 1848.265566] env[62824]: ERROR nova.compute.manager Traceback (most recent call last): [ 1848.265566] env[62824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1996, in _allocate_network_async [ 1848.265566] env[62824]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 1848.265566] env[62824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 1848.265566] env[62824]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 1848.265566] env[62824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 1848.265566] env[62824]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 1848.265566] env[62824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1848.265566] env[62824]: ERROR nova.compute.manager self.force_reraise() [ 1848.265566] env[62824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1848.265566] env[62824]: ERROR nova.compute.manager raise self.value [ 1848.265566] env[62824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 1848.265566] env[62824]: ERROR nova.compute.manager updated_port = self._update_port( [ 1848.265566] env[62824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 1848.265566] env[62824]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 1848.266155] env[62824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 1848.266155] env[62824]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 1848.266155] env[62824]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 2bf4c281-010d-4311-be12-2ec99b845218, please check neutron logs for more information. [ 1848.266155] env[62824]: ERROR nova.compute.manager [ 1848.266155] env[62824]: Traceback (most recent call last): [ 1848.266155] env[62824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 1848.266155] env[62824]: listener.cb(fileno) [ 1848.266155] env[62824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 1848.266155] env[62824]: result = function(*args, **kwargs) [ 1848.266155] env[62824]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 1848.266155] env[62824]: return func(*args, **kwargs) [ 1848.266155] env[62824]: File "/opt/stack/nova/nova/compute/manager.py", line 2018, in _allocate_network_async [ 1848.266155] env[62824]: raise e [ 1848.266155] env[62824]: File "/opt/stack/nova/nova/compute/manager.py", line 1996, in _allocate_network_async [ 1848.266155] env[62824]: nwinfo = self.network_api.allocate_for_instance( [ 1848.266155] env[62824]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 1848.266155] env[62824]: created_port_ids = self._update_ports_for_instance( [ 1848.266155] env[62824]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 1848.266155] env[62824]: with excutils.save_and_reraise_exception(): [ 1848.266155] env[62824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1848.266155] env[62824]: self.force_reraise() [ 1848.266155] env[62824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1848.266155] env[62824]: raise self.value [ 1848.266155] env[62824]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 1848.266155] env[62824]: updated_port = self._update_port( [ 1848.266155] env[62824]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 1848.266155] env[62824]: _ensure_no_port_binding_failure(port) [ 1848.266155] env[62824]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 1848.266155] env[62824]: raise exception.PortBindingFailed(port_id=port['id']) [ 1848.267102] env[62824]: nova.exception.PortBindingFailed: Binding failed for port 2bf4c281-010d-4311-be12-2ec99b845218, please check neutron logs for more information. [ 1848.267102] env[62824]: Removing descriptor: 15 [ 1848.267102] env[62824]: ERROR nova.compute.manager [None req-9f666b55-9bba-45f0-98a3-b9d8b6261703 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 2bf4c281-010d-4311-be12-2ec99b845218, please check neutron logs for more information. [ 1848.267102] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] Traceback (most recent call last): [ 1848.267102] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] File "/opt/stack/nova/nova/compute/manager.py", line 2901, in _build_resources [ 1848.267102] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] yield resources [ 1848.267102] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] File "/opt/stack/nova/nova/compute/manager.py", line 2648, in _build_and_run_instance [ 1848.267102] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] self.driver.spawn(context, instance, image_meta, [ 1848.267102] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 1848.267102] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] self._vmops.spawn(context, instance, image_meta, injected_files, [ 1848.267102] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 1848.267102] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] vm_ref = self.build_virtual_machine(instance, [ 1848.267498] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 1848.267498] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] vif_infos = vmwarevif.get_vif_info(self._session, [ 1848.267498] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 1848.267498] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] for vif in network_info: [ 1848.267498] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 1848.267498] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] return self._sync_wrapper(fn, *args, **kwargs) [ 1848.267498] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 1848.267498] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] self.wait() [ 1848.267498] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 1848.267498] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] self[:] = self._gt.wait() [ 1848.267498] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 1848.267498] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] return self._exit_event.wait() [ 1848.267498] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 1848.268111] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] result = hub.switch() [ 1848.268111] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 1848.268111] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] return self.greenlet.switch() [ 1848.268111] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 1848.268111] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] result = function(*args, **kwargs) [ 1848.268111] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 1848.268111] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] return func(*args, **kwargs) [ 1848.268111] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] File "/opt/stack/nova/nova/compute/manager.py", line 2018, in _allocate_network_async [ 1848.268111] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] raise e [ 1848.268111] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] File "/opt/stack/nova/nova/compute/manager.py", line 1996, in _allocate_network_async [ 1848.268111] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] nwinfo = self.network_api.allocate_for_instance( [ 1848.268111] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 1848.268111] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] created_port_ids = self._update_ports_for_instance( [ 1848.268803] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 1848.268803] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] with excutils.save_and_reraise_exception(): [ 1848.268803] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1848.268803] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] self.force_reraise() [ 1848.268803] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1848.268803] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] raise self.value [ 1848.268803] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 1848.268803] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] updated_port = self._update_port( [ 1848.268803] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 1848.268803] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] _ensure_no_port_binding_failure(port) [ 1848.268803] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 1848.268803] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] raise exception.PortBindingFailed(port_id=port['id']) [ 1848.269994] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] nova.exception.PortBindingFailed: Binding failed for port 2bf4c281-010d-4311-be12-2ec99b845218, please check neutron logs for more information. [ 1848.269994] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] [ 1848.269994] env[62824]: INFO nova.compute.manager [None req-9f666b55-9bba-45f0-98a3-b9d8b6261703 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] Terminating instance [ 1848.738506] env[62824]: DEBUG nova.compute.utils [None req-34e3c185-2b6d-4b05-bb0f-3253c603dd1c tempest-InstanceActionsV221TestJSON-1308314300 tempest-InstanceActionsV221TestJSON-1308314300-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1848.738506] env[62824]: DEBUG nova.compute.manager [None req-34e3c185-2b6d-4b05-bb0f-3253c603dd1c tempest-InstanceActionsV221TestJSON-1308314300 tempest-InstanceActionsV221TestJSON-1308314300-project-member] [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1848.738506] env[62824]: DEBUG nova.network.neutron [None req-34e3c185-2b6d-4b05-bb0f-3253c603dd1c tempest-InstanceActionsV221TestJSON-1308314300 tempest-InstanceActionsV221TestJSON-1308314300-project-member] [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1848.751177] env[62824]: DEBUG oslo_vmware.api [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Task: {'id': task-2145628, 'name': PowerOnVM_Task, 'duration_secs': 0.603685} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1848.752131] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] [instance: f4157385-43ea-4e8c-887c-0985af37abd1] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1848.752472] env[62824]: INFO nova.compute.manager [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] [instance: f4157385-43ea-4e8c-887c-0985af37abd1] Took 9.27 seconds to spawn the instance on the hypervisor. [ 1848.753370] env[62824]: DEBUG nova.compute.manager [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] [instance: f4157385-43ea-4e8c-887c-0985af37abd1] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1848.756849] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc3cdefe-d40d-4ed9-a621-cf3ec7394fc3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1848.771449] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9f666b55-9bba-45f0-98a3-b9d8b6261703 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Acquiring lock "refresh_cache-00fca8eb-c209-4632-abe6-5f092b04ac56" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1848.785037] env[62824]: DEBUG nova.network.neutron [req-8bebdb83-decd-4d84-86c9-2b255a5c8a05 req-7e467813-f627-4fda-b212-82d0ce9b8161 service nova] [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1848.791856] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1848.847271] env[62824]: DEBUG nova.policy [None req-34e3c185-2b6d-4b05-bb0f-3253c603dd1c tempest-InstanceActionsV221TestJSON-1308314300 tempest-InstanceActionsV221TestJSON-1308314300-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7ceb4919e21646cabd7c34c272e25148', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0e5a803206c044b4b889349568d27b16', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 1848.976022] env[62824]: DEBUG nova.network.neutron [req-8bebdb83-decd-4d84-86c9-2b255a5c8a05 req-7e467813-f627-4fda-b212-82d0ce9b8161 service nova] [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1849.268576] env[62824]: DEBUG nova.compute.manager [None req-34e3c185-2b6d-4b05-bb0f-3253c603dd1c tempest-InstanceActionsV221TestJSON-1308314300 tempest-InstanceActionsV221TestJSON-1308314300-project-member] [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1849.286339] env[62824]: INFO nova.scheduler.client.report [None req-6070f146-69b5-4aa1-b292-f106b5e9564b tempest-ServerAddressesNegativeTestJSON-1575558569 tempest-ServerAddressesNegativeTestJSON-1575558569-project-member] Deleted allocations for instance 6cc60e74-0474-47a5-ba57-c1c646ce0e1d [ 1849.297016] env[62824]: INFO nova.compute.manager [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] [instance: f4157385-43ea-4e8c-887c-0985af37abd1] Took 20.81 seconds to build instance. [ 1849.479429] env[62824]: DEBUG oslo_concurrency.lockutils [req-8bebdb83-decd-4d84-86c9-2b255a5c8a05 req-7e467813-f627-4fda-b212-82d0ce9b8161 service nova] Releasing lock "refresh_cache-00fca8eb-c209-4632-abe6-5f092b04ac56" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1849.479871] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9f666b55-9bba-45f0-98a3-b9d8b6261703 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Acquired lock "refresh_cache-00fca8eb-c209-4632-abe6-5f092b04ac56" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1849.480096] env[62824]: DEBUG nova.network.neutron [None req-9f666b55-9bba-45f0-98a3-b9d8b6261703 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1849.575776] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2a24dd6-aec5-4bd4-afa6-f14e214c8ce2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1849.586805] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ed0b1c0-c023-4742-a365-8d8ee139858d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1849.633335] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5c68263-89b5-4c9a-a6bc-0e1dedabb28c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1849.642557] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-128c56b1-6013-4da5-b185-e8982d58dde0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1849.658306] env[62824]: DEBUG nova.compute.provider_tree [None req-5cdb74c7-a27b-45c4-a89a-19c7714f04b5 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1849.805771] env[62824]: DEBUG oslo_concurrency.lockutils [None req-88a913cb-c25c-4200-8a98-8d69545866d2 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Lock "f4157385-43ea-4e8c-887c-0985af37abd1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.331s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1849.806212] env[62824]: DEBUG oslo_concurrency.lockutils [None req-6070f146-69b5-4aa1-b292-f106b5e9564b tempest-ServerAddressesNegativeTestJSON-1575558569 tempest-ServerAddressesNegativeTestJSON-1575558569-project-member] Lock "6cc60e74-0474-47a5-ba57-c1c646ce0e1d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.624s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1849.987882] env[62824]: DEBUG nova.network.neutron [None req-34e3c185-2b6d-4b05-bb0f-3253c603dd1c tempest-InstanceActionsV221TestJSON-1308314300 tempest-InstanceActionsV221TestJSON-1308314300-project-member] [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] Successfully created port: d07db52b-09d8-4f22-95da-96a68cc5247c {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1850.069731] env[62824]: DEBUG nova.network.neutron [None req-9f666b55-9bba-45f0-98a3-b9d8b6261703 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1850.161990] env[62824]: DEBUG nova.scheduler.client.report [None req-5cdb74c7-a27b-45c4-a89a-19c7714f04b5 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1850.290133] env[62824]: DEBUG nova.compute.manager [None req-34e3c185-2b6d-4b05-bb0f-3253c603dd1c tempest-InstanceActionsV221TestJSON-1308314300 tempest-InstanceActionsV221TestJSON-1308314300-project-member] [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1850.311409] env[62824]: DEBUG nova.compute.manager [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1850.319389] env[62824]: DEBUG nova.compute.manager [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] [instance: 82382932-7302-4441-a6f8-9aa2300ec0f6] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1850.331889] env[62824]: DEBUG nova.virt.hardware [None req-34e3c185-2b6d-4b05-bb0f-3253c603dd1c tempest-InstanceActionsV221TestJSON-1308314300 tempest-InstanceActionsV221TestJSON-1308314300-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1850.331889] env[62824]: DEBUG nova.virt.hardware [None req-34e3c185-2b6d-4b05-bb0f-3253c603dd1c tempest-InstanceActionsV221TestJSON-1308314300 tempest-InstanceActionsV221TestJSON-1308314300-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1850.331889] env[62824]: DEBUG nova.virt.hardware [None req-34e3c185-2b6d-4b05-bb0f-3253c603dd1c tempest-InstanceActionsV221TestJSON-1308314300 tempest-InstanceActionsV221TestJSON-1308314300-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1850.332182] env[62824]: DEBUG nova.virt.hardware [None req-34e3c185-2b6d-4b05-bb0f-3253c603dd1c tempest-InstanceActionsV221TestJSON-1308314300 tempest-InstanceActionsV221TestJSON-1308314300-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1850.332182] env[62824]: DEBUG nova.virt.hardware [None req-34e3c185-2b6d-4b05-bb0f-3253c603dd1c tempest-InstanceActionsV221TestJSON-1308314300 tempest-InstanceActionsV221TestJSON-1308314300-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1850.332256] env[62824]: DEBUG nova.virt.hardware [None req-34e3c185-2b6d-4b05-bb0f-3253c603dd1c tempest-InstanceActionsV221TestJSON-1308314300 tempest-InstanceActionsV221TestJSON-1308314300-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1850.332486] env[62824]: DEBUG nova.virt.hardware [None req-34e3c185-2b6d-4b05-bb0f-3253c603dd1c tempest-InstanceActionsV221TestJSON-1308314300 tempest-InstanceActionsV221TestJSON-1308314300-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1850.332624] env[62824]: DEBUG nova.virt.hardware [None req-34e3c185-2b6d-4b05-bb0f-3253c603dd1c tempest-InstanceActionsV221TestJSON-1308314300 tempest-InstanceActionsV221TestJSON-1308314300-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1850.332805] env[62824]: DEBUG nova.virt.hardware [None req-34e3c185-2b6d-4b05-bb0f-3253c603dd1c tempest-InstanceActionsV221TestJSON-1308314300 tempest-InstanceActionsV221TestJSON-1308314300-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1850.333679] env[62824]: DEBUG nova.virt.hardware [None req-34e3c185-2b6d-4b05-bb0f-3253c603dd1c tempest-InstanceActionsV221TestJSON-1308314300 tempest-InstanceActionsV221TestJSON-1308314300-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1850.333679] env[62824]: DEBUG nova.virt.hardware [None req-34e3c185-2b6d-4b05-bb0f-3253c603dd1c tempest-InstanceActionsV221TestJSON-1308314300 tempest-InstanceActionsV221TestJSON-1308314300-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1850.334093] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53f4dd9e-43b0-4122-b58d-b91fa085b435 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1850.343788] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21bfb65f-06cd-4091-8e91-faa0c1ca4f42 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1850.411076] env[62824]: DEBUG nova.network.neutron [None req-9f666b55-9bba-45f0-98a3-b9d8b6261703 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1850.669104] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5cdb74c7-a27b-45c4-a89a-19c7714f04b5 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.441s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1850.669104] env[62824]: DEBUG nova.compute.manager [None req-5cdb74c7-a27b-45c4-a89a-19c7714f04b5 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694-project-member] [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1850.671552] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2559566b-844a-4d61-b989-1f0a741aecd2 tempest-ServerDiagnosticsTest-1769488740 tempest-ServerDiagnosticsTest-1769488740-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.743s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1850.863471] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1850.868263] env[62824]: DEBUG oslo_concurrency.lockutils [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1850.916869] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9f666b55-9bba-45f0-98a3-b9d8b6261703 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Releasing lock "refresh_cache-00fca8eb-c209-4632-abe6-5f092b04ac56" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1850.916869] env[62824]: DEBUG nova.compute.manager [None req-9f666b55-9bba-45f0-98a3-b9d8b6261703 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1850.916869] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-9f666b55-9bba-45f0-98a3-b9d8b6261703 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1850.916869] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e626d11f-4e6c-492d-9d24-ff1cdaeb06be {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1850.932314] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Acquiring lock "dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1850.932314] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Lock "dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1850.935617] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ee86d7e-49cb-4dea-85a1-277d3c8b49f2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1850.967489] env[62824]: WARNING nova.virt.vmwareapi.vmops [None req-9f666b55-9bba-45f0-98a3-b9d8b6261703 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 00fca8eb-c209-4632-abe6-5f092b04ac56 could not be found. [ 1850.967694] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-9f666b55-9bba-45f0-98a3-b9d8b6261703 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1850.968223] env[62824]: INFO nova.compute.manager [None req-9f666b55-9bba-45f0-98a3-b9d8b6261703 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] Took 0.05 seconds to destroy the instance on the hypervisor. [ 1850.968223] env[62824]: DEBUG oslo.service.loopingcall [None req-9f666b55-9bba-45f0-98a3-b9d8b6261703 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1850.968830] env[62824]: DEBUG nova.compute.manager [-] [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1850.968830] env[62824]: DEBUG nova.network.neutron [-] [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1851.178673] env[62824]: DEBUG nova.compute.utils [None req-5cdb74c7-a27b-45c4-a89a-19c7714f04b5 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1851.183769] env[62824]: DEBUG nova.compute.manager [None req-5cdb74c7-a27b-45c4-a89a-19c7714f04b5 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694-project-member] [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1851.183898] env[62824]: DEBUG nova.network.neutron [None req-5cdb74c7-a27b-45c4-a89a-19c7714f04b5 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694-project-member] [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1851.208688] env[62824]: DEBUG nova.network.neutron [-] [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1851.306606] env[62824]: DEBUG nova.policy [None req-5cdb74c7-a27b-45c4-a89a-19c7714f04b5 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '87de9140ab694270a844ae40885833b3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0a48ec49a90b4cd4874ebdd318984b15', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 1851.505196] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-857b32fb-8fbc-40e8-af2b-e945f49ae5f4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1851.513090] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d136030-c781-45d5-aa38-aa8f2831f477 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1851.544475] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7853255-d00a-48f4-9326-0df13cd95329 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1851.552424] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3890bbb-c753-408a-bdff-94c658ecedca {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1851.568319] env[62824]: DEBUG nova.compute.provider_tree [None req-2559566b-844a-4d61-b989-1f0a741aecd2 tempest-ServerDiagnosticsTest-1769488740 tempest-ServerDiagnosticsTest-1769488740-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1851.684770] env[62824]: DEBUG nova.compute.manager [None req-5cdb74c7-a27b-45c4-a89a-19c7714f04b5 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694-project-member] [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1852.076221] env[62824]: DEBUG nova.scheduler.client.report [None req-2559566b-844a-4d61-b989-1f0a741aecd2 tempest-ServerDiagnosticsTest-1769488740 tempest-ServerDiagnosticsTest-1769488740-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1852.433936] env[62824]: DEBUG nova.network.neutron [None req-5cdb74c7-a27b-45c4-a89a-19c7714f04b5 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694-project-member] [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] Successfully created port: e42d44c3-e2b4-45a5-b707-7d81a18b96b0 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1852.581772] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2559566b-844a-4d61-b989-1f0a741aecd2 tempest-ServerDiagnosticsTest-1769488740 tempest-ServerDiagnosticsTest-1769488740-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.910s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1852.582536] env[62824]: ERROR nova.compute.manager [None req-2559566b-844a-4d61-b989-1f0a741aecd2 tempest-ServerDiagnosticsTest-1769488740 tempest-ServerDiagnosticsTest-1769488740-project-member] [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3fe39149-045f-4e9f-8ab4-72da2377c013, please check neutron logs for more information. [ 1852.582536] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] Traceback (most recent call last): [ 1852.582536] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] File "/opt/stack/nova/nova/compute/manager.py", line 2648, in _build_and_run_instance [ 1852.582536] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] self.driver.spawn(context, instance, image_meta, [ 1852.582536] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 1852.582536] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 1852.582536] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 1852.582536] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] vm_ref = self.build_virtual_machine(instance, [ 1852.582536] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 1852.582536] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] vif_infos = vmwarevif.get_vif_info(self._session, [ 1852.582536] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 1852.583125] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] for vif in network_info: [ 1852.583125] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 1852.583125] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] return self._sync_wrapper(fn, *args, **kwargs) [ 1852.583125] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 1852.583125] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] self.wait() [ 1852.583125] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 1852.583125] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] self[:] = self._gt.wait() [ 1852.583125] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 1852.583125] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] return self._exit_event.wait() [ 1852.583125] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 1852.583125] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] result = hub.switch() [ 1852.583125] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 1852.583125] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] return self.greenlet.switch() [ 1852.583561] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 1852.583561] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] result = function(*args, **kwargs) [ 1852.583561] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 1852.583561] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] return func(*args, **kwargs) [ 1852.583561] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] File "/opt/stack/nova/nova/compute/manager.py", line 2018, in _allocate_network_async [ 1852.583561] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] raise e [ 1852.583561] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] File "/opt/stack/nova/nova/compute/manager.py", line 1996, in _allocate_network_async [ 1852.583561] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] nwinfo = self.network_api.allocate_for_instance( [ 1852.583561] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 1852.583561] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] created_port_ids = self._update_ports_for_instance( [ 1852.583561] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 1852.583561] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] with excutils.save_and_reraise_exception(): [ 1852.583561] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1852.583979] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] self.force_reraise() [ 1852.583979] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1852.583979] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] raise self.value [ 1852.583979] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 1852.583979] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] updated_port = self._update_port( [ 1852.583979] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 1852.583979] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] _ensure_no_port_binding_failure(port) [ 1852.583979] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 1852.583979] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] raise exception.PortBindingFailed(port_id=port['id']) [ 1852.583979] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] nova.exception.PortBindingFailed: Binding failed for port 3fe39149-045f-4e9f-8ab4-72da2377c013, please check neutron logs for more information. [ 1852.583979] env[62824]: ERROR nova.compute.manager [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] [ 1852.584377] env[62824]: DEBUG nova.compute.utils [None req-2559566b-844a-4d61-b989-1f0a741aecd2 tempest-ServerDiagnosticsTest-1769488740 tempest-ServerDiagnosticsTest-1769488740-project-member] [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] Binding failed for port 3fe39149-045f-4e9f-8ab4-72da2377c013, please check neutron logs for more information. {{(pid=62824) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 1852.584377] env[62824]: DEBUG oslo_concurrency.lockutils [None req-61fb5137-5f87-475f-a5d2-0722cddc68b4 tempest-ServersTestJSON-928648138 tempest-ServersTestJSON-928648138-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 10.673s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1852.595870] env[62824]: DEBUG nova.compute.manager [None req-2559566b-844a-4d61-b989-1f0a741aecd2 tempest-ServerDiagnosticsTest-1769488740 tempest-ServerDiagnosticsTest-1769488740-project-member] [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] Build of instance 042c5f0f-bf1f-417e-be05-964cfcb20c7e was re-scheduled: Binding failed for port 3fe39149-045f-4e9f-8ab4-72da2377c013, please check neutron logs for more information. {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2487}} [ 1852.851462] env[62824]: DEBUG nova.compute.manager [None req-2559566b-844a-4d61-b989-1f0a741aecd2 tempest-ServerDiagnosticsTest-1769488740 tempest-ServerDiagnosticsTest-1769488740-project-member] [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] Unplugging VIFs for instance {{(pid=62824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3013}} [ 1852.851462] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2559566b-844a-4d61-b989-1f0a741aecd2 tempest-ServerDiagnosticsTest-1769488740 tempest-ServerDiagnosticsTest-1769488740-project-member] Acquiring lock "refresh_cache-042c5f0f-bf1f-417e-be05-964cfcb20c7e" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1852.851462] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2559566b-844a-4d61-b989-1f0a741aecd2 tempest-ServerDiagnosticsTest-1769488740 tempest-ServerDiagnosticsTest-1769488740-project-member] Acquired lock "refresh_cache-042c5f0f-bf1f-417e-be05-964cfcb20c7e" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1852.851462] env[62824]: DEBUG nova.network.neutron [None req-2559566b-844a-4d61-b989-1f0a741aecd2 tempest-ServerDiagnosticsTest-1769488740 tempest-ServerDiagnosticsTest-1769488740-project-member] [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1852.851462] env[62824]: DEBUG nova.compute.manager [None req-5cdb74c7-a27b-45c4-a89a-19c7714f04b5 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694-project-member] [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1852.854541] env[62824]: DEBUG nova.virt.hardware [None req-5cdb74c7-a27b-45c4-a89a-19c7714f04b5 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1852.854541] env[62824]: DEBUG nova.virt.hardware [None req-5cdb74c7-a27b-45c4-a89a-19c7714f04b5 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1852.854541] env[62824]: DEBUG nova.virt.hardware [None req-5cdb74c7-a27b-45c4-a89a-19c7714f04b5 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1852.858433] env[62824]: DEBUG nova.virt.hardware [None req-5cdb74c7-a27b-45c4-a89a-19c7714f04b5 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1852.858433] env[62824]: DEBUG nova.virt.hardware [None req-5cdb74c7-a27b-45c4-a89a-19c7714f04b5 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1852.858433] env[62824]: DEBUG nova.virt.hardware [None req-5cdb74c7-a27b-45c4-a89a-19c7714f04b5 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1852.858433] env[62824]: DEBUG nova.virt.hardware [None req-5cdb74c7-a27b-45c4-a89a-19c7714f04b5 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1852.858433] env[62824]: DEBUG nova.virt.hardware [None req-5cdb74c7-a27b-45c4-a89a-19c7714f04b5 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1852.858951] env[62824]: DEBUG nova.virt.hardware [None req-5cdb74c7-a27b-45c4-a89a-19c7714f04b5 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1852.858951] env[62824]: DEBUG nova.virt.hardware [None req-5cdb74c7-a27b-45c4-a89a-19c7714f04b5 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1852.858951] env[62824]: DEBUG nova.virt.hardware [None req-5cdb74c7-a27b-45c4-a89a-19c7714f04b5 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1852.858951] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c4aed13-b4db-4a80-a164-7baa05a3586d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1852.858951] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-181feacc-867b-459c-bf43-434d59966ad0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1852.892288] env[62824]: DEBUG nova.compute.manager [req-28bb9997-5608-4461-95a8-a1f7e9d26956 req-b32aa15d-ea73-4b11-8ecb-26d408ad4a9b service nova] [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] Received event network-vif-deleted-2bf4c281-010d-4311-be12-2ec99b845218 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1853.163641] env[62824]: DEBUG nova.network.neutron [None req-2559566b-844a-4d61-b989-1f0a741aecd2 tempest-ServerDiagnosticsTest-1769488740 tempest-ServerDiagnosticsTest-1769488740-project-member] [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1853.359597] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4d83ff4-2eb8-4eb8-80fe-2253b92f271b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1853.370073] env[62824]: DEBUG nova.network.neutron [None req-2559566b-844a-4d61-b989-1f0a741aecd2 tempest-ServerDiagnosticsTest-1769488740 tempest-ServerDiagnosticsTest-1769488740-project-member] [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1853.371921] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2c66f06-4220-4f23-b54d-c5ff0d92cd81 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1853.412719] env[62824]: DEBUG nova.network.neutron [-] [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1853.415560] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f380aacc-084f-4fa2-9038-0a4da1a8178b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1853.423830] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5db69be5-6200-45d4-af94-749c856ac1c1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1853.440418] env[62824]: DEBUG nova.compute.provider_tree [None req-61fb5137-5f87-475f-a5d2-0722cddc68b4 tempest-ServersTestJSON-928648138 tempest-ServersTestJSON-928648138-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1853.448815] env[62824]: DEBUG nova.compute.manager [None req-05ec47ce-505b-4b98-8e49-2ff7b777224b tempest-ServerDiagnosticsV248Test-1498148915 tempest-ServerDiagnosticsV248Test-1498148915-project-admin] [instance: f4157385-43ea-4e8c-887c-0985af37abd1] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1853.449971] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5586dd64-d231-4846-8187-25fdfbb73a40 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1853.457361] env[62824]: INFO nova.compute.manager [None req-05ec47ce-505b-4b98-8e49-2ff7b777224b tempest-ServerDiagnosticsV248Test-1498148915 tempest-ServerDiagnosticsV248Test-1498148915-project-admin] [instance: f4157385-43ea-4e8c-887c-0985af37abd1] Retrieving diagnostics [ 1853.458313] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85bfe632-66b7-4d60-904b-c133ba7cb047 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1853.499535] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Acquiring lock "4dfe9f82-7f89-486b-b159-948656d2896c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1853.499771] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Lock "4dfe9f82-7f89-486b-b159-948656d2896c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1853.714501] env[62824]: ERROR nova.compute.manager [None req-34e3c185-2b6d-4b05-bb0f-3253c603dd1c tempest-InstanceActionsV221TestJSON-1308314300 tempest-InstanceActionsV221TestJSON-1308314300-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d07db52b-09d8-4f22-95da-96a68cc5247c, please check neutron logs for more information. [ 1853.714501] env[62824]: ERROR nova.compute.manager Traceback (most recent call last): [ 1853.714501] env[62824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1996, in _allocate_network_async [ 1853.714501] env[62824]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 1853.714501] env[62824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 1853.714501] env[62824]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 1853.714501] env[62824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 1853.714501] env[62824]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 1853.714501] env[62824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1853.714501] env[62824]: ERROR nova.compute.manager self.force_reraise() [ 1853.714501] env[62824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1853.714501] env[62824]: ERROR nova.compute.manager raise self.value [ 1853.714501] env[62824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 1853.714501] env[62824]: ERROR nova.compute.manager updated_port = self._update_port( [ 1853.714501] env[62824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 1853.714501] env[62824]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 1853.715036] env[62824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 1853.715036] env[62824]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 1853.715036] env[62824]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d07db52b-09d8-4f22-95da-96a68cc5247c, please check neutron logs for more information. [ 1853.715036] env[62824]: ERROR nova.compute.manager [ 1853.715036] env[62824]: Traceback (most recent call last): [ 1853.715036] env[62824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 1853.715036] env[62824]: listener.cb(fileno) [ 1853.715036] env[62824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 1853.715036] env[62824]: result = function(*args, **kwargs) [ 1853.715036] env[62824]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 1853.715036] env[62824]: return func(*args, **kwargs) [ 1853.715036] env[62824]: File "/opt/stack/nova/nova/compute/manager.py", line 2018, in _allocate_network_async [ 1853.715036] env[62824]: raise e [ 1853.715036] env[62824]: File "/opt/stack/nova/nova/compute/manager.py", line 1996, in _allocate_network_async [ 1853.715036] env[62824]: nwinfo = self.network_api.allocate_for_instance( [ 1853.715036] env[62824]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 1853.715036] env[62824]: created_port_ids = self._update_ports_for_instance( [ 1853.715036] env[62824]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 1853.715036] env[62824]: with excutils.save_and_reraise_exception(): [ 1853.715036] env[62824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1853.715036] env[62824]: self.force_reraise() [ 1853.715036] env[62824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1853.715036] env[62824]: raise self.value [ 1853.715036] env[62824]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 1853.715036] env[62824]: updated_port = self._update_port( [ 1853.715036] env[62824]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 1853.715036] env[62824]: _ensure_no_port_binding_failure(port) [ 1853.715036] env[62824]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 1853.715036] env[62824]: raise exception.PortBindingFailed(port_id=port['id']) [ 1853.715866] env[62824]: nova.exception.PortBindingFailed: Binding failed for port d07db52b-09d8-4f22-95da-96a68cc5247c, please check neutron logs for more information. [ 1853.715866] env[62824]: Removing descriptor: 18 [ 1853.716281] env[62824]: ERROR nova.compute.manager [None req-34e3c185-2b6d-4b05-bb0f-3253c603dd1c tempest-InstanceActionsV221TestJSON-1308314300 tempest-InstanceActionsV221TestJSON-1308314300-project-member] [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d07db52b-09d8-4f22-95da-96a68cc5247c, please check neutron logs for more information. [ 1853.716281] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] Traceback (most recent call last): [ 1853.716281] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] File "/opt/stack/nova/nova/compute/manager.py", line 2901, in _build_resources [ 1853.716281] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] yield resources [ 1853.716281] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] File "/opt/stack/nova/nova/compute/manager.py", line 2648, in _build_and_run_instance [ 1853.716281] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] self.driver.spawn(context, instance, image_meta, [ 1853.716281] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 1853.716281] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 1853.716281] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 1853.716281] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] vm_ref = self.build_virtual_machine(instance, [ 1853.716281] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 1853.716680] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] vif_infos = vmwarevif.get_vif_info(self._session, [ 1853.716680] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 1853.716680] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] for vif in network_info: [ 1853.716680] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 1853.716680] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] return self._sync_wrapper(fn, *args, **kwargs) [ 1853.716680] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 1853.716680] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] self.wait() [ 1853.716680] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 1853.716680] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] self[:] = self._gt.wait() [ 1853.716680] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 1853.716680] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] return self._exit_event.wait() [ 1853.716680] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 1853.716680] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] result = hub.switch() [ 1853.717248] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 1853.717248] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] return self.greenlet.switch() [ 1853.717248] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 1853.717248] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] result = function(*args, **kwargs) [ 1853.717248] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 1853.717248] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] return func(*args, **kwargs) [ 1853.717248] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] File "/opt/stack/nova/nova/compute/manager.py", line 2018, in _allocate_network_async [ 1853.717248] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] raise e [ 1853.717248] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] File "/opt/stack/nova/nova/compute/manager.py", line 1996, in _allocate_network_async [ 1853.717248] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] nwinfo = self.network_api.allocate_for_instance( [ 1853.717248] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 1853.717248] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] created_port_ids = self._update_ports_for_instance( [ 1853.717248] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 1853.717917] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] with excutils.save_and_reraise_exception(): [ 1853.717917] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1853.717917] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] self.force_reraise() [ 1853.717917] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1853.717917] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] raise self.value [ 1853.717917] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 1853.717917] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] updated_port = self._update_port( [ 1853.717917] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 1853.717917] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] _ensure_no_port_binding_failure(port) [ 1853.717917] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 1853.717917] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] raise exception.PortBindingFailed(port_id=port['id']) [ 1853.717917] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] nova.exception.PortBindingFailed: Binding failed for port d07db52b-09d8-4f22-95da-96a68cc5247c, please check neutron logs for more information. [ 1853.717917] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] [ 1853.718351] env[62824]: INFO nova.compute.manager [None req-34e3c185-2b6d-4b05-bb0f-3253c603dd1c tempest-InstanceActionsV221TestJSON-1308314300 tempest-InstanceActionsV221TestJSON-1308314300-project-member] [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] Terminating instance [ 1853.877709] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2559566b-844a-4d61-b989-1f0a741aecd2 tempest-ServerDiagnosticsTest-1769488740 tempest-ServerDiagnosticsTest-1769488740-project-member] Releasing lock "refresh_cache-042c5f0f-bf1f-417e-be05-964cfcb20c7e" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1853.877971] env[62824]: DEBUG nova.compute.manager [None req-2559566b-844a-4d61-b989-1f0a741aecd2 tempest-ServerDiagnosticsTest-1769488740 tempest-ServerDiagnosticsTest-1769488740-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3036}} [ 1853.878171] env[62824]: DEBUG nova.compute.manager [None req-2559566b-844a-4d61-b989-1f0a741aecd2 tempest-ServerDiagnosticsTest-1769488740 tempest-ServerDiagnosticsTest-1769488740-project-member] [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1853.878343] env[62824]: DEBUG nova.network.neutron [None req-2559566b-844a-4d61-b989-1f0a741aecd2 tempest-ServerDiagnosticsTest-1769488740 tempest-ServerDiagnosticsTest-1769488740-project-member] [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1853.918972] env[62824]: INFO nova.compute.manager [-] [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] Took 2.95 seconds to deallocate network for instance. [ 1853.923039] env[62824]: DEBUG nova.network.neutron [None req-2559566b-844a-4d61-b989-1f0a741aecd2 tempest-ServerDiagnosticsTest-1769488740 tempest-ServerDiagnosticsTest-1769488740-project-member] [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1853.923260] env[62824]: DEBUG nova.compute.claims [None req-9f666b55-9bba-45f0-98a3-b9d8b6261703 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] Aborting claim: {{(pid=62824) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 1853.924339] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9f666b55-9bba-45f0-98a3-b9d8b6261703 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1853.946188] env[62824]: DEBUG nova.scheduler.client.report [None req-61fb5137-5f87-475f-a5d2-0722cddc68b4 tempest-ServersTestJSON-928648138 tempest-ServersTestJSON-928648138-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1854.224140] env[62824]: DEBUG oslo_concurrency.lockutils [None req-34e3c185-2b6d-4b05-bb0f-3253c603dd1c tempest-InstanceActionsV221TestJSON-1308314300 tempest-InstanceActionsV221TestJSON-1308314300-project-member] Acquiring lock "refresh_cache-bec6441f-72b8-4664-b125-b0aa5f21636a" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1854.224140] env[62824]: DEBUG oslo_concurrency.lockutils [None req-34e3c185-2b6d-4b05-bb0f-3253c603dd1c tempest-InstanceActionsV221TestJSON-1308314300 tempest-InstanceActionsV221TestJSON-1308314300-project-member] Acquired lock "refresh_cache-bec6441f-72b8-4664-b125-b0aa5f21636a" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1854.224140] env[62824]: DEBUG nova.network.neutron [None req-34e3c185-2b6d-4b05-bb0f-3253c603dd1c tempest-InstanceActionsV221TestJSON-1308314300 tempest-InstanceActionsV221TestJSON-1308314300-project-member] [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1854.297894] env[62824]: DEBUG nova.compute.manager [req-e85f018e-b8c0-42e5-b097-387d5f5efc5e req-6f60150f-da25-4b9c-8015-9456ea4fadfa service nova] [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] Received event network-changed-d07db52b-09d8-4f22-95da-96a68cc5247c {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1854.298589] env[62824]: DEBUG nova.compute.manager [req-e85f018e-b8c0-42e5-b097-387d5f5efc5e req-6f60150f-da25-4b9c-8015-9456ea4fadfa service nova] [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] Refreshing instance network info cache due to event network-changed-d07db52b-09d8-4f22-95da-96a68cc5247c. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1854.299306] env[62824]: DEBUG oslo_concurrency.lockutils [req-e85f018e-b8c0-42e5-b097-387d5f5efc5e req-6f60150f-da25-4b9c-8015-9456ea4fadfa service nova] Acquiring lock "refresh_cache-bec6441f-72b8-4664-b125-b0aa5f21636a" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1854.426145] env[62824]: DEBUG nova.network.neutron [None req-2559566b-844a-4d61-b989-1f0a741aecd2 tempest-ServerDiagnosticsTest-1769488740 tempest-ServerDiagnosticsTest-1769488740-project-member] [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1854.456529] env[62824]: DEBUG oslo_concurrency.lockutils [None req-61fb5137-5f87-475f-a5d2-0722cddc68b4 tempest-ServersTestJSON-928648138 tempest-ServersTestJSON-928648138-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.870s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1854.456529] env[62824]: ERROR nova.compute.manager [None req-61fb5137-5f87-475f-a5d2-0722cddc68b4 tempest-ServersTestJSON-928648138 tempest-ServersTestJSON-928648138-project-member] [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e427511e-b0f8-42bd-b24f-0037a1ac6e76, please check neutron logs for more information. [ 1854.456529] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] Traceback (most recent call last): [ 1854.456529] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] File "/opt/stack/nova/nova/compute/manager.py", line 2648, in _build_and_run_instance [ 1854.456529] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] self.driver.spawn(context, instance, image_meta, [ 1854.456529] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 1854.456529] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] self._vmops.spawn(context, instance, image_meta, injected_files, [ 1854.456529] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 1854.456529] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] vm_ref = self.build_virtual_machine(instance, [ 1854.456908] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 1854.456908] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] vif_infos = vmwarevif.get_vif_info(self._session, [ 1854.456908] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 1854.456908] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] for vif in network_info: [ 1854.456908] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 1854.456908] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] return self._sync_wrapper(fn, *args, **kwargs) [ 1854.456908] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 1854.456908] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] self.wait() [ 1854.456908] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 1854.456908] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] self[:] = self._gt.wait() [ 1854.456908] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 1854.456908] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] return self._exit_event.wait() [ 1854.456908] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 1854.457779] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] result = hub.switch() [ 1854.457779] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 1854.457779] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] return self.greenlet.switch() [ 1854.457779] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 1854.457779] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] result = function(*args, **kwargs) [ 1854.457779] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 1854.457779] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] return func(*args, **kwargs) [ 1854.457779] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] File "/opt/stack/nova/nova/compute/manager.py", line 2018, in _allocate_network_async [ 1854.457779] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] raise e [ 1854.457779] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] File "/opt/stack/nova/nova/compute/manager.py", line 1996, in _allocate_network_async [ 1854.457779] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] nwinfo = self.network_api.allocate_for_instance( [ 1854.457779] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 1854.457779] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] created_port_ids = self._update_ports_for_instance( [ 1854.458186] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 1854.458186] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] with excutils.save_and_reraise_exception(): [ 1854.458186] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1854.458186] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] self.force_reraise() [ 1854.458186] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1854.458186] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] raise self.value [ 1854.458186] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 1854.458186] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] updated_port = self._update_port( [ 1854.458186] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 1854.458186] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] _ensure_no_port_binding_failure(port) [ 1854.458186] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 1854.458186] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] raise exception.PortBindingFailed(port_id=port['id']) [ 1854.458528] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] nova.exception.PortBindingFailed: Binding failed for port e427511e-b0f8-42bd-b24f-0037a1ac6e76, please check neutron logs for more information. [ 1854.458528] env[62824]: ERROR nova.compute.manager [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] [ 1854.458528] env[62824]: DEBUG nova.compute.utils [None req-61fb5137-5f87-475f-a5d2-0722cddc68b4 tempest-ServersTestJSON-928648138 tempest-ServersTestJSON-928648138-project-member] [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] Binding failed for port e427511e-b0f8-42bd-b24f-0037a1ac6e76, please check neutron logs for more information. {{(pid=62824) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 1854.460215] env[62824]: DEBUG nova.compute.manager [None req-61fb5137-5f87-475f-a5d2-0722cddc68b4 tempest-ServersTestJSON-928648138 tempest-ServersTestJSON-928648138-project-member] [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] Build of instance 0b845d80-c8b2-4b16-bda2-08fc78fb4337 was re-scheduled: Binding failed for port e427511e-b0f8-42bd-b24f-0037a1ac6e76, please check neutron logs for more information. {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2487}} [ 1854.460215] env[62824]: DEBUG nova.compute.manager [None req-61fb5137-5f87-475f-a5d2-0722cddc68b4 tempest-ServersTestJSON-928648138 tempest-ServersTestJSON-928648138-project-member] [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] Unplugging VIFs for instance {{(pid=62824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3013}} [ 1854.460804] env[62824]: DEBUG oslo_concurrency.lockutils [None req-61fb5137-5f87-475f-a5d2-0722cddc68b4 tempest-ServersTestJSON-928648138 tempest-ServersTestJSON-928648138-project-member] Acquiring lock "refresh_cache-0b845d80-c8b2-4b16-bda2-08fc78fb4337" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1854.460804] env[62824]: DEBUG oslo_concurrency.lockutils [None req-61fb5137-5f87-475f-a5d2-0722cddc68b4 tempest-ServersTestJSON-928648138 tempest-ServersTestJSON-928648138-project-member] Acquired lock "refresh_cache-0b845d80-c8b2-4b16-bda2-08fc78fb4337" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1854.460910] env[62824]: DEBUG nova.network.neutron [None req-61fb5137-5f87-475f-a5d2-0722cddc68b4 tempest-ServersTestJSON-928648138 tempest-ServersTestJSON-928648138-project-member] [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1854.461932] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.206s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1854.467631] env[62824]: INFO nova.compute.claims [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1854.777930] env[62824]: DEBUG nova.network.neutron [None req-34e3c185-2b6d-4b05-bb0f-3253c603dd1c tempest-InstanceActionsV221TestJSON-1308314300 tempest-InstanceActionsV221TestJSON-1308314300-project-member] [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1854.929487] env[62824]: INFO nova.compute.manager [None req-2559566b-844a-4d61-b989-1f0a741aecd2 tempest-ServerDiagnosticsTest-1769488740 tempest-ServerDiagnosticsTest-1769488740-project-member] [instance: 042c5f0f-bf1f-417e-be05-964cfcb20c7e] Took 1.05 seconds to deallocate network for instance. [ 1855.041974] env[62824]: DEBUG nova.network.neutron [None req-61fb5137-5f87-475f-a5d2-0722cddc68b4 tempest-ServersTestJSON-928648138 tempest-ServersTestJSON-928648138-project-member] [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1855.124275] env[62824]: DEBUG nova.network.neutron [None req-34e3c185-2b6d-4b05-bb0f-3253c603dd1c tempest-InstanceActionsV221TestJSON-1308314300 tempest-InstanceActionsV221TestJSON-1308314300-project-member] [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1855.413960] env[62824]: DEBUG nova.network.neutron [None req-61fb5137-5f87-475f-a5d2-0722cddc68b4 tempest-ServersTestJSON-928648138 tempest-ServersTestJSON-928648138-project-member] [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1855.629581] env[62824]: DEBUG oslo_concurrency.lockutils [None req-34e3c185-2b6d-4b05-bb0f-3253c603dd1c tempest-InstanceActionsV221TestJSON-1308314300 tempest-InstanceActionsV221TestJSON-1308314300-project-member] Releasing lock "refresh_cache-bec6441f-72b8-4664-b125-b0aa5f21636a" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1855.630048] env[62824]: DEBUG nova.compute.manager [None req-34e3c185-2b6d-4b05-bb0f-3253c603dd1c tempest-InstanceActionsV221TestJSON-1308314300 tempest-InstanceActionsV221TestJSON-1308314300-project-member] [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1855.630244] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-34e3c185-2b6d-4b05-bb0f-3253c603dd1c tempest-InstanceActionsV221TestJSON-1308314300 tempest-InstanceActionsV221TestJSON-1308314300-project-member] [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1855.630787] env[62824]: DEBUG oslo_concurrency.lockutils [req-e85f018e-b8c0-42e5-b097-387d5f5efc5e req-6f60150f-da25-4b9c-8015-9456ea4fadfa service nova] Acquired lock "refresh_cache-bec6441f-72b8-4664-b125-b0aa5f21636a" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1855.630970] env[62824]: DEBUG nova.network.neutron [req-e85f018e-b8c0-42e5-b097-387d5f5efc5e req-6f60150f-da25-4b9c-8015-9456ea4fadfa service nova] [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] Refreshing network info cache for port d07db52b-09d8-4f22-95da-96a68cc5247c {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1855.632698] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6fff45a7-9558-49e1-9518-1bdfcdf17b5c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1855.644445] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdafd08d-b1e3-4cc5-8be6-7516cd070b1c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1855.681269] env[62824]: WARNING nova.virt.vmwareapi.vmops [None req-34e3c185-2b6d-4b05-bb0f-3253c603dd1c tempest-InstanceActionsV221TestJSON-1308314300 tempest-InstanceActionsV221TestJSON-1308314300-project-member] [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance bec6441f-72b8-4664-b125-b0aa5f21636a could not be found. [ 1855.683024] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-34e3c185-2b6d-4b05-bb0f-3253c603dd1c tempest-InstanceActionsV221TestJSON-1308314300 tempest-InstanceActionsV221TestJSON-1308314300-project-member] [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1855.683024] env[62824]: INFO nova.compute.manager [None req-34e3c185-2b6d-4b05-bb0f-3253c603dd1c tempest-InstanceActionsV221TestJSON-1308314300 tempest-InstanceActionsV221TestJSON-1308314300-project-member] [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] Took 0.05 seconds to destroy the instance on the hypervisor. [ 1855.683024] env[62824]: DEBUG oslo.service.loopingcall [None req-34e3c185-2b6d-4b05-bb0f-3253c603dd1c tempest-InstanceActionsV221TestJSON-1308314300 tempest-InstanceActionsV221TestJSON-1308314300-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1855.688271] env[62824]: DEBUG nova.compute.manager [-] [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1855.688777] env[62824]: DEBUG nova.network.neutron [-] [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1855.728764] env[62824]: DEBUG nova.network.neutron [-] [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1855.835863] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a34d7633-3e25-4de6-9897-9bc3574772b2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1855.844129] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-191ce54b-d328-4579-9e5c-aeb4e3522e6d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1855.879548] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65723ac9-f57a-4188-882b-f056d3964b11 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1855.888073] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-238d3055-3383-4f94-af9a-97a18baa4432 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1855.902348] env[62824]: DEBUG nova.compute.provider_tree [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1855.916423] env[62824]: DEBUG oslo_concurrency.lockutils [None req-61fb5137-5f87-475f-a5d2-0722cddc68b4 tempest-ServersTestJSON-928648138 tempest-ServersTestJSON-928648138-project-member] Releasing lock "refresh_cache-0b845d80-c8b2-4b16-bda2-08fc78fb4337" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1855.917756] env[62824]: DEBUG nova.compute.manager [None req-61fb5137-5f87-475f-a5d2-0722cddc68b4 tempest-ServersTestJSON-928648138 tempest-ServersTestJSON-928648138-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3036}} [ 1855.917756] env[62824]: DEBUG nova.compute.manager [None req-61fb5137-5f87-475f-a5d2-0722cddc68b4 tempest-ServersTestJSON-928648138 tempest-ServersTestJSON-928648138-project-member] [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1855.917756] env[62824]: DEBUG nova.network.neutron [None req-61fb5137-5f87-475f-a5d2-0722cddc68b4 tempest-ServersTestJSON-928648138 tempest-ServersTestJSON-928648138-project-member] [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1855.964275] env[62824]: DEBUG nova.network.neutron [None req-61fb5137-5f87-475f-a5d2-0722cddc68b4 tempest-ServersTestJSON-928648138 tempest-ServersTestJSON-928648138-project-member] [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1855.970977] env[62824]: INFO nova.scheduler.client.report [None req-2559566b-844a-4d61-b989-1f0a741aecd2 tempest-ServerDiagnosticsTest-1769488740 tempest-ServerDiagnosticsTest-1769488740-project-member] Deleted allocations for instance 042c5f0f-bf1f-417e-be05-964cfcb20c7e [ 1856.184809] env[62824]: DEBUG nova.network.neutron [req-e85f018e-b8c0-42e5-b097-387d5f5efc5e req-6f60150f-da25-4b9c-8015-9456ea4fadfa service nova] [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1856.232615] env[62824]: DEBUG nova.network.neutron [-] [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1856.405223] env[62824]: DEBUG nova.scheduler.client.report [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1856.470758] env[62824]: DEBUG nova.network.neutron [None req-61fb5137-5f87-475f-a5d2-0722cddc68b4 tempest-ServersTestJSON-928648138 tempest-ServersTestJSON-928648138-project-member] [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1856.482890] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2559566b-844a-4d61-b989-1f0a741aecd2 tempest-ServerDiagnosticsTest-1769488740 tempest-ServerDiagnosticsTest-1769488740-project-member] Lock "042c5f0f-bf1f-417e-be05-964cfcb20c7e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.640s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1856.566270] env[62824]: DEBUG nova.network.neutron [req-e85f018e-b8c0-42e5-b097-387d5f5efc5e req-6f60150f-da25-4b9c-8015-9456ea4fadfa service nova] [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1856.735847] env[62824]: INFO nova.compute.manager [-] [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] Took 1.05 seconds to deallocate network for instance. [ 1856.739428] env[62824]: DEBUG nova.compute.claims [None req-34e3c185-2b6d-4b05-bb0f-3253c603dd1c tempest-InstanceActionsV221TestJSON-1308314300 tempest-InstanceActionsV221TestJSON-1308314300-project-member] [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] Aborting claim: {{(pid=62824) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 1856.739428] env[62824]: DEBUG oslo_concurrency.lockutils [None req-34e3c185-2b6d-4b05-bb0f-3253c603dd1c tempest-InstanceActionsV221TestJSON-1308314300 tempest-InstanceActionsV221TestJSON-1308314300-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1856.886985] env[62824]: DEBUG nova.compute.manager [req-471b1029-f8ca-41b5-bed5-17a7eee64075 req-a94c108e-bb73-47f6-8925-005a8599aa63 service nova] [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] Received event network-vif-deleted-d07db52b-09d8-4f22-95da-96a68cc5247c {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1856.887223] env[62824]: DEBUG nova.compute.manager [req-471b1029-f8ca-41b5-bed5-17a7eee64075 req-a94c108e-bb73-47f6-8925-005a8599aa63 service nova] [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] Received event network-changed-e42d44c3-e2b4-45a5-b707-7d81a18b96b0 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1856.887376] env[62824]: DEBUG nova.compute.manager [req-471b1029-f8ca-41b5-bed5-17a7eee64075 req-a94c108e-bb73-47f6-8925-005a8599aa63 service nova] [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] Refreshing instance network info cache due to event network-changed-e42d44c3-e2b4-45a5-b707-7d81a18b96b0. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1856.887581] env[62824]: DEBUG oslo_concurrency.lockutils [req-471b1029-f8ca-41b5-bed5-17a7eee64075 req-a94c108e-bb73-47f6-8925-005a8599aa63 service nova] Acquiring lock "refresh_cache-2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1856.887714] env[62824]: DEBUG oslo_concurrency.lockutils [req-471b1029-f8ca-41b5-bed5-17a7eee64075 req-a94c108e-bb73-47f6-8925-005a8599aa63 service nova] Acquired lock "refresh_cache-2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1856.887867] env[62824]: DEBUG nova.network.neutron [req-471b1029-f8ca-41b5-bed5-17a7eee64075 req-a94c108e-bb73-47f6-8925-005a8599aa63 service nova] [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] Refreshing network info cache for port e42d44c3-e2b4-45a5-b707-7d81a18b96b0 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1856.910784] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.449s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1856.911244] env[62824]: DEBUG nova.compute.manager [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1856.913935] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7cd5dacb-b71b-429c-bfda-96302f774c53 tempest-ServersTestFqdnHostnames-528120407 tempest-ServersTestFqdnHostnames-528120407-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.153s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1856.975107] env[62824]: INFO nova.compute.manager [None req-61fb5137-5f87-475f-a5d2-0722cddc68b4 tempest-ServersTestJSON-928648138 tempest-ServersTestJSON-928648138-project-member] [instance: 0b845d80-c8b2-4b16-bda2-08fc78fb4337] Took 1.06 seconds to deallocate network for instance. [ 1856.989103] env[62824]: DEBUG nova.compute.manager [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1857.069094] env[62824]: DEBUG oslo_concurrency.lockutils [req-e85f018e-b8c0-42e5-b097-387d5f5efc5e req-6f60150f-da25-4b9c-8015-9456ea4fadfa service nova] Releasing lock "refresh_cache-bec6441f-72b8-4664-b125-b0aa5f21636a" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1857.241867] env[62824]: ERROR nova.compute.manager [None req-5cdb74c7-a27b-45c4-a89a-19c7714f04b5 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e42d44c3-e2b4-45a5-b707-7d81a18b96b0, please check neutron logs for more information. [ 1857.241867] env[62824]: ERROR nova.compute.manager Traceback (most recent call last): [ 1857.241867] env[62824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1996, in _allocate_network_async [ 1857.241867] env[62824]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 1857.241867] env[62824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 1857.241867] env[62824]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 1857.241867] env[62824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 1857.241867] env[62824]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 1857.241867] env[62824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1857.241867] env[62824]: ERROR nova.compute.manager self.force_reraise() [ 1857.241867] env[62824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1857.241867] env[62824]: ERROR nova.compute.manager raise self.value [ 1857.241867] env[62824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 1857.241867] env[62824]: ERROR nova.compute.manager updated_port = self._update_port( [ 1857.241867] env[62824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 1857.241867] env[62824]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 1857.242638] env[62824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 1857.242638] env[62824]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 1857.242638] env[62824]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e42d44c3-e2b4-45a5-b707-7d81a18b96b0, please check neutron logs for more information. [ 1857.242638] env[62824]: ERROR nova.compute.manager [ 1857.242638] env[62824]: Traceback (most recent call last): [ 1857.242638] env[62824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 1857.242638] env[62824]: listener.cb(fileno) [ 1857.242638] env[62824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 1857.242638] env[62824]: result = function(*args, **kwargs) [ 1857.242638] env[62824]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 1857.242638] env[62824]: return func(*args, **kwargs) [ 1857.242638] env[62824]: File "/opt/stack/nova/nova/compute/manager.py", line 2018, in _allocate_network_async [ 1857.242638] env[62824]: raise e [ 1857.242638] env[62824]: File "/opt/stack/nova/nova/compute/manager.py", line 1996, in _allocate_network_async [ 1857.242638] env[62824]: nwinfo = self.network_api.allocate_for_instance( [ 1857.242638] env[62824]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 1857.242638] env[62824]: created_port_ids = self._update_ports_for_instance( [ 1857.242638] env[62824]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 1857.242638] env[62824]: with excutils.save_and_reraise_exception(): [ 1857.242638] env[62824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1857.242638] env[62824]: self.force_reraise() [ 1857.242638] env[62824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1857.242638] env[62824]: raise self.value [ 1857.242638] env[62824]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 1857.242638] env[62824]: updated_port = self._update_port( [ 1857.242638] env[62824]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 1857.242638] env[62824]: _ensure_no_port_binding_failure(port) [ 1857.242638] env[62824]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 1857.242638] env[62824]: raise exception.PortBindingFailed(port_id=port['id']) [ 1857.243923] env[62824]: nova.exception.PortBindingFailed: Binding failed for port e42d44c3-e2b4-45a5-b707-7d81a18b96b0, please check neutron logs for more information. [ 1857.243923] env[62824]: Removing descriptor: 17 [ 1857.243923] env[62824]: ERROR nova.compute.manager [None req-5cdb74c7-a27b-45c4-a89a-19c7714f04b5 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694-project-member] [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e42d44c3-e2b4-45a5-b707-7d81a18b96b0, please check neutron logs for more information. [ 1857.243923] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] Traceback (most recent call last): [ 1857.243923] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] File "/opt/stack/nova/nova/compute/manager.py", line 2901, in _build_resources [ 1857.243923] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] yield resources [ 1857.243923] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] File "/opt/stack/nova/nova/compute/manager.py", line 2648, in _build_and_run_instance [ 1857.243923] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] self.driver.spawn(context, instance, image_meta, [ 1857.243923] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 1857.243923] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 1857.243923] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 1857.243923] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] vm_ref = self.build_virtual_machine(instance, [ 1857.244491] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 1857.244491] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] vif_infos = vmwarevif.get_vif_info(self._session, [ 1857.244491] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 1857.244491] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] for vif in network_info: [ 1857.244491] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 1857.244491] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] return self._sync_wrapper(fn, *args, **kwargs) [ 1857.244491] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 1857.244491] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] self.wait() [ 1857.244491] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 1857.244491] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] self[:] = self._gt.wait() [ 1857.244491] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 1857.244491] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] return self._exit_event.wait() [ 1857.244491] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 1857.244960] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] result = hub.switch() [ 1857.244960] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 1857.244960] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] return self.greenlet.switch() [ 1857.244960] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 1857.244960] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] result = function(*args, **kwargs) [ 1857.244960] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 1857.244960] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] return func(*args, **kwargs) [ 1857.244960] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] File "/opt/stack/nova/nova/compute/manager.py", line 2018, in _allocate_network_async [ 1857.244960] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] raise e [ 1857.244960] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] File "/opt/stack/nova/nova/compute/manager.py", line 1996, in _allocate_network_async [ 1857.244960] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] nwinfo = self.network_api.allocate_for_instance( [ 1857.244960] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 1857.244960] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] created_port_ids = self._update_ports_for_instance( [ 1857.246096] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 1857.246096] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] with excutils.save_and_reraise_exception(): [ 1857.246096] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1857.246096] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] self.force_reraise() [ 1857.246096] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1857.246096] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] raise self.value [ 1857.246096] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 1857.246096] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] updated_port = self._update_port( [ 1857.246096] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 1857.246096] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] _ensure_no_port_binding_failure(port) [ 1857.246096] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 1857.246096] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] raise exception.PortBindingFailed(port_id=port['id']) [ 1857.246533] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] nova.exception.PortBindingFailed: Binding failed for port e42d44c3-e2b4-45a5-b707-7d81a18b96b0, please check neutron logs for more information. [ 1857.246533] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] [ 1857.246533] env[62824]: INFO nova.compute.manager [None req-5cdb74c7-a27b-45c4-a89a-19c7714f04b5 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694-project-member] [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] Terminating instance [ 1857.418676] env[62824]: DEBUG nova.compute.utils [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1857.430250] env[62824]: DEBUG nova.compute.manager [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1857.430431] env[62824]: DEBUG nova.network.neutron [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1857.495096] env[62824]: DEBUG nova.network.neutron [req-471b1029-f8ca-41b5-bed5-17a7eee64075 req-a94c108e-bb73-47f6-8925-005a8599aa63 service nova] [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1857.512954] env[62824]: DEBUG oslo_concurrency.lockutils [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1857.628013] env[62824]: DEBUG nova.policy [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b2dc74ce3a8f4207a77ca65814be057b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e2e61ea652f94257b884d2e1b495e446', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 1857.750507] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5cdb74c7-a27b-45c4-a89a-19c7714f04b5 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694-project-member] Acquiring lock "refresh_cache-2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1857.811280] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61552166-6589-40e0-94a0-765c4efd2740 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1857.830552] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f1a510c-1cb0-40d9-817c-1e77f2b7133d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1857.870651] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04b8743e-5a7a-4b7f-8cf6-aecb5d07df4e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1857.887355] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88b4b4d7-504c-435b-9047-697a4cc1b5cb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1857.906464] env[62824]: DEBUG nova.compute.provider_tree [None req-7cd5dacb-b71b-429c-bfda-96302f774c53 tempest-ServersTestFqdnHostnames-528120407 tempest-ServersTestFqdnHostnames-528120407-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1857.922787] env[62824]: DEBUG nova.compute.manager [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1858.024644] env[62824]: INFO nova.scheduler.client.report [None req-61fb5137-5f87-475f-a5d2-0722cddc68b4 tempest-ServersTestJSON-928648138 tempest-ServersTestJSON-928648138-project-member] Deleted allocations for instance 0b845d80-c8b2-4b16-bda2-08fc78fb4337 [ 1858.386709] env[62824]: DEBUG nova.network.neutron [req-471b1029-f8ca-41b5-bed5-17a7eee64075 req-a94c108e-bb73-47f6-8925-005a8599aa63 service nova] [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1858.410628] env[62824]: DEBUG nova.scheduler.client.report [None req-7cd5dacb-b71b-429c-bfda-96302f774c53 tempest-ServersTestFqdnHostnames-528120407 tempest-ServersTestFqdnHostnames-528120407-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1858.538243] env[62824]: DEBUG oslo_concurrency.lockutils [None req-61fb5137-5f87-475f-a5d2-0722cddc68b4 tempest-ServersTestJSON-928648138 tempest-ServersTestJSON-928648138-project-member] Lock "0b845d80-c8b2-4b16-bda2-08fc78fb4337" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.090s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1858.895500] env[62824]: DEBUG oslo_concurrency.lockutils [req-471b1029-f8ca-41b5-bed5-17a7eee64075 req-a94c108e-bb73-47f6-8925-005a8599aa63 service nova] Releasing lock "refresh_cache-2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1858.895978] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5cdb74c7-a27b-45c4-a89a-19c7714f04b5 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694-project-member] Acquired lock "refresh_cache-2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1858.898470] env[62824]: DEBUG nova.network.neutron [None req-5cdb74c7-a27b-45c4-a89a-19c7714f04b5 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694-project-member] [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1858.921941] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7cd5dacb-b71b-429c-bfda-96302f774c53 tempest-ServersTestFqdnHostnames-528120407 tempest-ServersTestFqdnHostnames-528120407-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.008s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1858.923168] env[62824]: ERROR nova.compute.manager [None req-7cd5dacb-b71b-429c-bfda-96302f774c53 tempest-ServersTestFqdnHostnames-528120407 tempest-ServersTestFqdnHostnames-528120407-project-member] [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3a529f42-989d-46e2-ae3f-1e1a4fa18860, please check neutron logs for more information. [ 1858.923168] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] Traceback (most recent call last): [ 1858.923168] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] File "/opt/stack/nova/nova/compute/manager.py", line 2648, in _build_and_run_instance [ 1858.923168] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] self.driver.spawn(context, instance, image_meta, [ 1858.923168] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 1858.923168] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] self._vmops.spawn(context, instance, image_meta, injected_files, [ 1858.923168] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 1858.923168] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] vm_ref = self.build_virtual_machine(instance, [ 1858.923168] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 1858.923168] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] vif_infos = vmwarevif.get_vif_info(self._session, [ 1858.923168] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 1858.923558] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] for vif in network_info: [ 1858.923558] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 1858.923558] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] return self._sync_wrapper(fn, *args, **kwargs) [ 1858.923558] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 1858.923558] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] self.wait() [ 1858.923558] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 1858.923558] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] self[:] = self._gt.wait() [ 1858.923558] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 1858.923558] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] return self._exit_event.wait() [ 1858.923558] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 1858.923558] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] result = hub.switch() [ 1858.923558] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 1858.923558] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] return self.greenlet.switch() [ 1858.923917] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 1858.923917] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] result = function(*args, **kwargs) [ 1858.923917] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 1858.923917] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] return func(*args, **kwargs) [ 1858.923917] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] File "/opt/stack/nova/nova/compute/manager.py", line 2018, in _allocate_network_async [ 1858.923917] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] raise e [ 1858.923917] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] File "/opt/stack/nova/nova/compute/manager.py", line 1996, in _allocate_network_async [ 1858.923917] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] nwinfo = self.network_api.allocate_for_instance( [ 1858.923917] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 1858.923917] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] created_port_ids = self._update_ports_for_instance( [ 1858.923917] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 1858.923917] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] with excutils.save_and_reraise_exception(): [ 1858.923917] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1858.924379] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] self.force_reraise() [ 1858.924379] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1858.924379] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] raise self.value [ 1858.924379] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 1858.924379] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] updated_port = self._update_port( [ 1858.924379] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 1858.924379] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] _ensure_no_port_binding_failure(port) [ 1858.924379] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 1858.924379] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] raise exception.PortBindingFailed(port_id=port['id']) [ 1858.924379] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] nova.exception.PortBindingFailed: Binding failed for port 3a529f42-989d-46e2-ae3f-1e1a4fa18860, please check neutron logs for more information. [ 1858.924379] env[62824]: ERROR nova.compute.manager [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] [ 1858.924784] env[62824]: DEBUG nova.compute.utils [None req-7cd5dacb-b71b-429c-bfda-96302f774c53 tempest-ServersTestFqdnHostnames-528120407 tempest-ServersTestFqdnHostnames-528120407-project-member] [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] Binding failed for port 3a529f42-989d-46e2-ae3f-1e1a4fa18860, please check neutron logs for more information. {{(pid=62824) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 1858.926869] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 13.191s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1858.934932] env[62824]: DEBUG nova.compute.manager [None req-7cd5dacb-b71b-429c-bfda-96302f774c53 tempest-ServersTestFqdnHostnames-528120407 tempest-ServersTestFqdnHostnames-528120407-project-member] [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] Build of instance 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf was re-scheduled: Binding failed for port 3a529f42-989d-46e2-ae3f-1e1a4fa18860, please check neutron logs for more information. {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2487}} [ 1858.934932] env[62824]: DEBUG nova.compute.manager [None req-7cd5dacb-b71b-429c-bfda-96302f774c53 tempest-ServersTestFqdnHostnames-528120407 tempest-ServersTestFqdnHostnames-528120407-project-member] [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] Unplugging VIFs for instance {{(pid=62824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3013}} [ 1858.934932] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7cd5dacb-b71b-429c-bfda-96302f774c53 tempest-ServersTestFqdnHostnames-528120407 tempest-ServersTestFqdnHostnames-528120407-project-member] Acquiring lock "refresh_cache-27e6ff4d-c4f2-41e7-92c2-a0e2445830bf" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1858.934932] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7cd5dacb-b71b-429c-bfda-96302f774c53 tempest-ServersTestFqdnHostnames-528120407 tempest-ServersTestFqdnHostnames-528120407-project-member] Acquired lock "refresh_cache-27e6ff4d-c4f2-41e7-92c2-a0e2445830bf" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1858.936094] env[62824]: DEBUG nova.network.neutron [None req-7cd5dacb-b71b-429c-bfda-96302f774c53 tempest-ServersTestFqdnHostnames-528120407 tempest-ServersTestFqdnHostnames-528120407-project-member] [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1858.937700] env[62824]: DEBUG nova.compute.manager [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1858.972117] env[62824]: DEBUG nova.virt.hardware [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1858.972345] env[62824]: DEBUG nova.virt.hardware [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1858.972592] env[62824]: DEBUG nova.virt.hardware [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1858.973720] env[62824]: DEBUG nova.virt.hardware [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1858.973720] env[62824]: DEBUG nova.virt.hardware [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1858.973720] env[62824]: DEBUG nova.virt.hardware [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1858.973720] env[62824]: DEBUG nova.virt.hardware [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1858.973720] env[62824]: DEBUG nova.virt.hardware [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1858.974037] env[62824]: DEBUG nova.virt.hardware [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1858.974037] env[62824]: DEBUG nova.virt.hardware [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1858.974037] env[62824]: DEBUG nova.virt.hardware [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1858.975063] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a248ab7b-2585-4e85-afd2-e2d14a50ea94 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1858.986899] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef6f7e08-b381-4542-9155-55eb641651fb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1859.041611] env[62824]: DEBUG nova.compute.manager [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] [instance: 517bf2cf-e142-4f36-bf53-79a8d2d8a530] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1859.131515] env[62824]: DEBUG nova.network.neutron [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Successfully created port: 2d861cfc-3ad4-4d40-ad77-e4530d363421 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1859.476621] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Acquiring lock "e158b5d4-c120-4e6c-89c4-7668e097926d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1859.477262] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Lock "e158b5d4-c120-4e6c-89c4-7668e097926d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1859.514152] env[62824]: DEBUG nova.network.neutron [None req-5cdb74c7-a27b-45c4-a89a-19c7714f04b5 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694-project-member] [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1859.528118] env[62824]: DEBUG nova.network.neutron [None req-7cd5dacb-b71b-429c-bfda-96302f774c53 tempest-ServersTestFqdnHostnames-528120407 tempest-ServersTestFqdnHostnames-528120407-project-member] [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1859.575545] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1860.072902] env[62824]: DEBUG nova.network.neutron [None req-7cd5dacb-b71b-429c-bfda-96302f774c53 tempest-ServersTestFqdnHostnames-528120407 tempest-ServersTestFqdnHostnames-528120407-project-member] [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1860.101360] env[62824]: DEBUG nova.network.neutron [None req-5cdb74c7-a27b-45c4-a89a-19c7714f04b5 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694-project-member] [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1860.485422] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1860.485422] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance f4157385-43ea-4e8c-887c-0985af37abd1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1860.485422] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 00fca8eb-c209-4632-abe6-5f092b04ac56 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1860.485422] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance bec6441f-72b8-4664-b125-b0aa5f21636a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1860.485921] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1860.485921] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 29b8dd5f-1855-490c-a01b-54840073a753 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1860.576134] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7cd5dacb-b71b-429c-bfda-96302f774c53 tempest-ServersTestFqdnHostnames-528120407 tempest-ServersTestFqdnHostnames-528120407-project-member] Releasing lock "refresh_cache-27e6ff4d-c4f2-41e7-92c2-a0e2445830bf" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1860.576134] env[62824]: DEBUG nova.compute.manager [None req-7cd5dacb-b71b-429c-bfda-96302f774c53 tempest-ServersTestFqdnHostnames-528120407 tempest-ServersTestFqdnHostnames-528120407-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3036}} [ 1860.576134] env[62824]: DEBUG nova.compute.manager [None req-7cd5dacb-b71b-429c-bfda-96302f774c53 tempest-ServersTestFqdnHostnames-528120407 tempest-ServersTestFqdnHostnames-528120407-project-member] [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1860.577785] env[62824]: DEBUG nova.network.neutron [None req-7cd5dacb-b71b-429c-bfda-96302f774c53 tempest-ServersTestFqdnHostnames-528120407 tempest-ServersTestFqdnHostnames-528120407-project-member] [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1860.604825] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5cdb74c7-a27b-45c4-a89a-19c7714f04b5 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694-project-member] Releasing lock "refresh_cache-2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1860.605379] env[62824]: DEBUG nova.compute.manager [None req-5cdb74c7-a27b-45c4-a89a-19c7714f04b5 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694-project-member] [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1860.605569] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5cdb74c7-a27b-45c4-a89a-19c7714f04b5 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694-project-member] [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1860.606286] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-157aa781-fac6-41d5-88f0-c63828063fa3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1860.618024] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28b86fda-ac69-4405-a45f-460b58fcdea7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1860.635926] env[62824]: DEBUG nova.network.neutron [None req-7cd5dacb-b71b-429c-bfda-96302f774c53 tempest-ServersTestFqdnHostnames-528120407 tempest-ServersTestFqdnHostnames-528120407-project-member] [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1860.647578] env[62824]: WARNING nova.virt.vmwareapi.vmops [None req-5cdb74c7-a27b-45c4-a89a-19c7714f04b5 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694-project-member] [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c could not be found. [ 1860.647832] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5cdb74c7-a27b-45c4-a89a-19c7714f04b5 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694-project-member] [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1860.648032] env[62824]: INFO nova.compute.manager [None req-5cdb74c7-a27b-45c4-a89a-19c7714f04b5 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694-project-member] [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] Took 0.04 seconds to destroy the instance on the hypervisor. [ 1860.649153] env[62824]: DEBUG oslo.service.loopingcall [None req-5cdb74c7-a27b-45c4-a89a-19c7714f04b5 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1860.649153] env[62824]: DEBUG nova.compute.manager [-] [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1860.649153] env[62824]: DEBUG nova.network.neutron [-] [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1860.660192] env[62824]: DEBUG nova.compute.manager [req-6a5dee52-097d-4663-b041-537556186ed9 req-1b1e1d0e-eaf9-4b5b-985b-8ed30f47d94f service nova] [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] Received event network-vif-deleted-e42d44c3-e2b4-45a5-b707-7d81a18b96b0 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1860.729939] env[62824]: DEBUG nova.network.neutron [-] [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1860.992538] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 04ac4ade-0bc8-4469-9b1b-f288b2f0a367 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1861.150076] env[62824]: DEBUG nova.network.neutron [None req-7cd5dacb-b71b-429c-bfda-96302f774c53 tempest-ServersTestFqdnHostnames-528120407 tempest-ServersTestFqdnHostnames-528120407-project-member] [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1861.237432] env[62824]: DEBUG nova.network.neutron [-] [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1861.497383] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 82382932-7302-4441-a6f8-9aa2300ec0f6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1861.656087] env[62824]: INFO nova.compute.manager [None req-7cd5dacb-b71b-429c-bfda-96302f774c53 tempest-ServersTestFqdnHostnames-528120407 tempest-ServersTestFqdnHostnames-528120407-project-member] [instance: 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf] Took 1.08 seconds to deallocate network for instance. [ 1861.739176] env[62824]: INFO nova.compute.manager [-] [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] Took 1.09 seconds to deallocate network for instance. [ 1861.745042] env[62824]: DEBUG nova.compute.claims [None req-5cdb74c7-a27b-45c4-a89a-19c7714f04b5 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694-project-member] [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] Aborting claim: {{(pid=62824) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 1861.745042] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5cdb74c7-a27b-45c4-a89a-19c7714f04b5 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1861.893266] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Acquiring lock "07ce931f-45ef-409b-b714-9f1cd47a3a88" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1861.893489] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Lock "07ce931f-45ef-409b-b714-9f1cd47a3a88" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1862.000410] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance f4d63a93-23af-470c-b36a-662af81dc386 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1862.329223] env[62824]: DEBUG nova.network.neutron [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Successfully updated port: 2d861cfc-3ad4-4d40-ad77-e4530d363421 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1862.507220] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1862.704440] env[62824]: INFO nova.scheduler.client.report [None req-7cd5dacb-b71b-429c-bfda-96302f774c53 tempest-ServersTestFqdnHostnames-528120407 tempest-ServersTestFqdnHostnames-528120407-project-member] Deleted allocations for instance 27e6ff4d-c4f2-41e7-92c2-a0e2445830bf [ 1862.838634] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Acquiring lock "refresh_cache-29b8dd5f-1855-490c-a01b-54840073a753" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1862.838780] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Acquired lock "refresh_cache-29b8dd5f-1855-490c-a01b-54840073a753" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1862.838933] env[62824]: DEBUG nova.network.neutron [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1863.011047] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 517bf2cf-e142-4f36-bf53-79a8d2d8a530 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1863.213429] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7cd5dacb-b71b-429c-bfda-96302f774c53 tempest-ServersTestFqdnHostnames-528120407 tempest-ServersTestFqdnHostnames-528120407-project-member] Lock "27e6ff4d-c4f2-41e7-92c2-a0e2445830bf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.450s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1863.293230] env[62824]: DEBUG nova.compute.manager [req-d2639cd6-e1b7-4867-945c-dfd2da0a778b req-fd5003d0-5dcd-4b22-a0f8-558cb6cc1da9 service nova] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Received event network-vif-plugged-2d861cfc-3ad4-4d40-ad77-e4530d363421 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1863.293890] env[62824]: DEBUG oslo_concurrency.lockutils [req-d2639cd6-e1b7-4867-945c-dfd2da0a778b req-fd5003d0-5dcd-4b22-a0f8-558cb6cc1da9 service nova] Acquiring lock "29b8dd5f-1855-490c-a01b-54840073a753-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1863.293890] env[62824]: DEBUG oslo_concurrency.lockutils [req-d2639cd6-e1b7-4867-945c-dfd2da0a778b req-fd5003d0-5dcd-4b22-a0f8-558cb6cc1da9 service nova] Lock "29b8dd5f-1855-490c-a01b-54840073a753-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1863.295553] env[62824]: DEBUG oslo_concurrency.lockutils [req-d2639cd6-e1b7-4867-945c-dfd2da0a778b req-fd5003d0-5dcd-4b22-a0f8-558cb6cc1da9 service nova] Lock "29b8dd5f-1855-490c-a01b-54840073a753-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1863.295553] env[62824]: DEBUG nova.compute.manager [req-d2639cd6-e1b7-4867-945c-dfd2da0a778b req-fd5003d0-5dcd-4b22-a0f8-558cb6cc1da9 service nova] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] No waiting events found dispatching network-vif-plugged-2d861cfc-3ad4-4d40-ad77-e4530d363421 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1863.295553] env[62824]: WARNING nova.compute.manager [req-d2639cd6-e1b7-4867-945c-dfd2da0a778b req-fd5003d0-5dcd-4b22-a0f8-558cb6cc1da9 service nova] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Received unexpected event network-vif-plugged-2d861cfc-3ad4-4d40-ad77-e4530d363421 for instance with vm_state building and task_state spawning. [ 1863.423225] env[62824]: DEBUG nova.network.neutron [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1863.514189] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 9062e606-511f-4fe7-9621-90b3c0e51eec has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1863.639839] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Acquiring lock "f3740d9a-b21a-4a79-9e28-2a89ecd08bb6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1863.639839] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Lock "f3740d9a-b21a-4a79-9e28-2a89ecd08bb6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1863.709629] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Acquiring lock "ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1863.709629] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Lock "ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1863.719611] env[62824]: DEBUG nova.compute.manager [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] [instance: 9062e606-511f-4fe7-9621-90b3c0e51eec] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1863.864023] env[62824]: DEBUG nova.network.neutron [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Updating instance_info_cache with network_info: [{"id": "2d861cfc-3ad4-4d40-ad77-e4530d363421", "address": "fa:16:3e:59:84:dd", "network": {"id": "f18b4230-e950-4957-a02b-107d27729346", "bridge": "br-int", "label": "tempest-VolumesAssistedSnapshotsTest-633702241-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e2e61ea652f94257b884d2e1b495e446", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4b5c60ce-845e-4506-bc10-348461fece6d", "external-id": "nsx-vlan-transportzone-831", "segmentation_id": 831, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2d861cfc-3a", "ovs_interfaceid": "2d861cfc-3ad4-4d40-ad77-e4530d363421", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1864.021784] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 384a96ea-40ce-43e1-a5f9-82f50b710b0b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1864.240321] env[62824]: DEBUG oslo_concurrency.lockutils [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1864.367240] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Releasing lock "refresh_cache-29b8dd5f-1855-490c-a01b-54840073a753" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1864.368115] env[62824]: DEBUG nova.compute.manager [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Instance network_info: |[{"id": "2d861cfc-3ad4-4d40-ad77-e4530d363421", "address": "fa:16:3e:59:84:dd", "network": {"id": "f18b4230-e950-4957-a02b-107d27729346", "bridge": "br-int", "label": "tempest-VolumesAssistedSnapshotsTest-633702241-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e2e61ea652f94257b884d2e1b495e446", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4b5c60ce-845e-4506-bc10-348461fece6d", "external-id": "nsx-vlan-transportzone-831", "segmentation_id": 831, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2d861cfc-3a", "ovs_interfaceid": "2d861cfc-3ad4-4d40-ad77-e4530d363421", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1864.368530] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:59:84:dd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4b5c60ce-845e-4506-bc10-348461fece6d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2d861cfc-3ad4-4d40-ad77-e4530d363421', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1864.378872] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Creating folder: Project (e2e61ea652f94257b884d2e1b495e446). Parent ref: group-v438503. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1864.379857] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bf9bda91-387f-42fc-a308-d9edd1de6498 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1864.395051] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Created folder: Project (e2e61ea652f94257b884d2e1b495e446) in parent group-v438503. [ 1864.395279] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Creating folder: Instances. Parent ref: group-v438511. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1864.395540] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c925ef27-9b5a-4bd1-9d0f-a2747c525138 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1864.406461] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Created folder: Instances in parent group-v438511. [ 1864.406461] env[62824]: DEBUG oslo.service.loopingcall [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1864.406461] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1864.406461] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0c56d545-1dca-4752-b793-7b58c7ef46aa {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1864.428278] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1864.428278] env[62824]: value = "task-2145637" [ 1864.428278] env[62824]: _type = "Task" [ 1864.428278] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1864.438626] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145637, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1864.530390] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1864.596388] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Acquiring lock "16b466c3-c749-4f96-a82c-32dad31138ec" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1864.596388] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Lock "16b466c3-c749-4f96-a82c-32dad31138ec" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1864.942919] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145637, 'name': CreateVM_Task, 'duration_secs': 0.487695} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1864.943995] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1864.962122] env[62824]: DEBUG oslo_vmware.service [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ea310cd-09d9-4060-9eaf-d0465d9c53d9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1864.971463] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1864.971463] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1864.971792] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1864.972695] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-76dbde72-d6d7-4ec5-9a38-dcc3b943d398 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1864.976876] env[62824]: DEBUG oslo_vmware.api [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Waiting for the task: (returnval){ [ 1864.976876] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52db60a9-6873-a2c2-8bd5-c5e72c238439" [ 1864.976876] env[62824]: _type = "Task" [ 1864.976876] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1864.986937] env[62824]: DEBUG oslo_vmware.api [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52db60a9-6873-a2c2-8bd5-c5e72c238439, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1865.034949] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 4dfe9f82-7f89-486b-b159-948656d2896c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1865.044365] env[62824]: DEBUG oslo_concurrency.lockutils [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Acquiring lock "b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1865.044638] env[62824]: DEBUG oslo_concurrency.lockutils [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Lock "b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1865.120512] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Acquiring lock "b8cc8cd3-ea03-40bf-b867-7ad193365552" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1865.120750] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Lock "b8cc8cd3-ea03-40bf-b867-7ad193365552" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1865.497233] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1865.497233] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1865.497402] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1865.497564] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1865.497786] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1865.498069] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-29b3362f-3c7c-4b00-856b-10373093eb34 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1865.518077] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1865.518286] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1865.521988] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a43e6c4-d4cb-4948-ae92-07f44dc34888 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1865.527431] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Acquiring lock "c36ece43-3d70-4e67-a740-9057f413c722" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1865.527649] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Lock "c36ece43-3d70-4e67-a740-9057f413c722" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1865.532273] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ecea032a-d51e-4f77-a60c-ce89272a2097 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1865.539621] env[62824]: DEBUG oslo_vmware.api [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Waiting for the task: (returnval){ [ 1865.539621] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52c97680-4630-5ccd-0042-9028e90b1fa0" [ 1865.539621] env[62824]: _type = "Task" [ 1865.539621] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1865.546024] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance e158b5d4-c120-4e6c-89c4-7668e097926d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1865.546472] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Total usable vcpus: 48, total allocated vcpus: 5 {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1865.546620] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1472MB phys_disk=200GB used_disk=5GB total_vcpus=48 used_vcpus=5 pci_stats=[] {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1865.549037] env[62824]: DEBUG oslo_vmware.api [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52c97680-4630-5ccd-0042-9028e90b1fa0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1865.775063] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Acquiring lock "b96d1351-f3a7-4bac-998c-a34ab2606041" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1865.775407] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Lock "b96d1351-f3a7-4bac-998c-a34ab2606041" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1865.875222] env[62824]: DEBUG nova.compute.manager [None req-1d97f079-2839-4d4c-a9db-bbee07006718 tempest-ServerDiagnosticsV248Test-1498148915 tempest-ServerDiagnosticsV248Test-1498148915-project-admin] [instance: f4157385-43ea-4e8c-887c-0985af37abd1] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1865.876660] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b1f0889-82a7-4ed3-a539-45f3d5ee37fb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1865.884407] env[62824]: INFO nova.compute.manager [None req-1d97f079-2839-4d4c-a9db-bbee07006718 tempest-ServerDiagnosticsV248Test-1498148915 tempest-ServerDiagnosticsV248Test-1498148915-project-admin] [instance: f4157385-43ea-4e8c-887c-0985af37abd1] Retrieving diagnostics [ 1865.885480] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d86b1814-2b8a-41f0-abdb-c1de0e405df7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1865.969474] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4aaaeed8-a6f3-4b9b-b8fe-18becfccafdf {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1865.977861] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ed17e8c-fef8-4163-87b5-89cf444efbb7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1866.021181] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c67fc54-654b-420c-9c29-ceae88f5974c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1866.033065] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7610d9f9-8e06-41ad-bda6-00f4c1c595a3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1866.049442] env[62824]: DEBUG nova.compute.provider_tree [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1866.055563] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Preparing fetch location {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1866.055802] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Creating directory with path [datastore1] vmware_temp/b40ede64-0abe-49ff-bfc5-8316c59660cf/9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1866.056087] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-470e61ca-6364-452f-a71a-ba11b470339c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1866.085533] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Created directory with path [datastore1] vmware_temp/b40ede64-0abe-49ff-bfc5-8316c59660cf/9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1866.085901] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Fetch image to [datastore1] vmware_temp/b40ede64-0abe-49ff-bfc5-8316c59660cf/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/tmp-sparse.vmdk {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1866.085980] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Downloading image file data 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 to [datastore1] vmware_temp/b40ede64-0abe-49ff-bfc5-8316c59660cf/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/tmp-sparse.vmdk on the data store datastore1 {{(pid=62824) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 1866.086775] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbbd4dcd-5134-4f57-af64-ffa679f88104 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1866.099020] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-867d645b-d594-4d59-a3a3-be083eaab6f5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1866.111474] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a240feb6-e60e-4f19-a147-24f6253f26f4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1866.144888] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8883ae9-4899-49a1-866c-886c611bb75d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1866.151389] env[62824]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-bc550322-8199-4ebe-91ee-62e64e9944a4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1866.244042] env[62824]: DEBUG nova.virt.vmwareapi.images [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Downloading image file data 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 to the data store datastore1 {{(pid=62824) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 1866.294915] env[62824]: DEBUG oslo_vmware.rw_handles [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/b40ede64-0abe-49ff-bfc5-8316c59660cf/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=62824) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1866.555087] env[62824]: DEBUG nova.scheduler.client.report [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1866.896816] env[62824]: DEBUG nova.compute.manager [req-a5a88b7f-ba2e-4ff7-88b0-7dc213e7ce80 req-13aca724-0ec3-4273-9b9a-d23e892b22dd service nova] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Received event network-changed-2d861cfc-3ad4-4d40-ad77-e4530d363421 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1866.896816] env[62824]: DEBUG nova.compute.manager [req-a5a88b7f-ba2e-4ff7-88b0-7dc213e7ce80 req-13aca724-0ec3-4273-9b9a-d23e892b22dd service nova] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Refreshing instance network info cache due to event network-changed-2d861cfc-3ad4-4d40-ad77-e4530d363421. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1866.896816] env[62824]: DEBUG oslo_concurrency.lockutils [req-a5a88b7f-ba2e-4ff7-88b0-7dc213e7ce80 req-13aca724-0ec3-4273-9b9a-d23e892b22dd service nova] Acquiring lock "refresh_cache-29b8dd5f-1855-490c-a01b-54840073a753" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1866.897064] env[62824]: DEBUG oslo_concurrency.lockutils [req-a5a88b7f-ba2e-4ff7-88b0-7dc213e7ce80 req-13aca724-0ec3-4273-9b9a-d23e892b22dd service nova] Acquired lock "refresh_cache-29b8dd5f-1855-490c-a01b-54840073a753" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1866.897064] env[62824]: DEBUG nova.network.neutron [req-a5a88b7f-ba2e-4ff7-88b0-7dc213e7ce80 req-13aca724-0ec3-4273-9b9a-d23e892b22dd service nova] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Refreshing network info cache for port 2d861cfc-3ad4-4d40-ad77-e4530d363421 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1866.959650] env[62824]: DEBUG oslo_vmware.rw_handles [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Completed reading data from the image iterator. {{(pid=62824) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1866.959813] env[62824]: DEBUG oslo_vmware.rw_handles [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Closing write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/b40ede64-0abe-49ff-bfc5-8316c59660cf/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=62824) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 1867.063995] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62824) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1867.063995] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 8.134s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1867.063995] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.269s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1867.063995] env[62824]: INFO nova.compute.claims [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] [instance: 04ac4ade-0bc8-4469-9b1b-f288b2f0a367] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1867.088127] env[62824]: DEBUG nova.virt.vmwareapi.images [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Downloaded image file data 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 to vmware_temp/b40ede64-0abe-49ff-bfc5-8316c59660cf/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/tmp-sparse.vmdk on the data store datastore1 {{(pid=62824) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 1867.089073] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Caching image {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1867.089909] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Copying Virtual Disk [datastore1] vmware_temp/b40ede64-0abe-49ff-bfc5-8316c59660cf/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/tmp-sparse.vmdk to [datastore1] vmware_temp/b40ede64-0abe-49ff-bfc5-8316c59660cf/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1867.090286] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1bb4dd01-b782-4c8a-88a7-37c011ab29a2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1867.101063] env[62824]: DEBUG oslo_vmware.api [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Waiting for the task: (returnval){ [ 1867.101063] env[62824]: value = "task-2145638" [ 1867.101063] env[62824]: _type = "Task" [ 1867.101063] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1867.109899] env[62824]: DEBUG oslo_vmware.api [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Task: {'id': task-2145638, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1867.387547] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Acquiring lock "b3003c4b-ae5a-48df-8c12-a915a76253f4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1867.387547] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Lock "b3003c4b-ae5a-48df-8c12-a915a76253f4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1867.617552] env[62824]: DEBUG oslo_vmware.api [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Task: {'id': task-2145638, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1867.789735] env[62824]: DEBUG nova.network.neutron [req-a5a88b7f-ba2e-4ff7-88b0-7dc213e7ce80 req-13aca724-0ec3-4273-9b9a-d23e892b22dd service nova] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Updated VIF entry in instance network info cache for port 2d861cfc-3ad4-4d40-ad77-e4530d363421. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1867.790393] env[62824]: DEBUG nova.network.neutron [req-a5a88b7f-ba2e-4ff7-88b0-7dc213e7ce80 req-13aca724-0ec3-4273-9b9a-d23e892b22dd service nova] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Updating instance_info_cache with network_info: [{"id": "2d861cfc-3ad4-4d40-ad77-e4530d363421", "address": "fa:16:3e:59:84:dd", "network": {"id": "f18b4230-e950-4957-a02b-107d27729346", "bridge": "br-int", "label": "tempest-VolumesAssistedSnapshotsTest-633702241-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e2e61ea652f94257b884d2e1b495e446", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4b5c60ce-845e-4506-bc10-348461fece6d", "external-id": "nsx-vlan-transportzone-831", "segmentation_id": 831, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2d861cfc-3a", "ovs_interfaceid": "2d861cfc-3ad4-4d40-ad77-e4530d363421", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1867.818929] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ba162a0b-a017-488a-838b-b0c0ba13e0d5 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Acquiring lock "f4157385-43ea-4e8c-887c-0985af37abd1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1867.819274] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ba162a0b-a017-488a-838b-b0c0ba13e0d5 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Lock "f4157385-43ea-4e8c-887c-0985af37abd1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1867.819503] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ba162a0b-a017-488a-838b-b0c0ba13e0d5 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Acquiring lock "f4157385-43ea-4e8c-887c-0985af37abd1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1867.819708] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ba162a0b-a017-488a-838b-b0c0ba13e0d5 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Lock "f4157385-43ea-4e8c-887c-0985af37abd1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1867.819920] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ba162a0b-a017-488a-838b-b0c0ba13e0d5 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Lock "f4157385-43ea-4e8c-887c-0985af37abd1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1867.822113] env[62824]: INFO nova.compute.manager [None req-ba162a0b-a017-488a-838b-b0c0ba13e0d5 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] [instance: f4157385-43ea-4e8c-887c-0985af37abd1] Terminating instance [ 1868.117293] env[62824]: DEBUG oslo_vmware.api [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Task: {'id': task-2145638, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.663082} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1868.117293] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Copied Virtual Disk [datastore1] vmware_temp/b40ede64-0abe-49ff-bfc5-8316c59660cf/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/tmp-sparse.vmdk to [datastore1] vmware_temp/b40ede64-0abe-49ff-bfc5-8316c59660cf/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1868.117293] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Deleting the datastore file [datastore1] vmware_temp/b40ede64-0abe-49ff-bfc5-8316c59660cf/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/tmp-sparse.vmdk {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1868.117710] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-033ebadf-399e-4004-82ae-6d4ac698aa5a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1868.126087] env[62824]: DEBUG oslo_vmware.api [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Waiting for the task: (returnval){ [ 1868.126087] env[62824]: value = "task-2145639" [ 1868.126087] env[62824]: _type = "Task" [ 1868.126087] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1868.137163] env[62824]: DEBUG oslo_vmware.api [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Task: {'id': task-2145639, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1868.293402] env[62824]: DEBUG oslo_concurrency.lockutils [req-a5a88b7f-ba2e-4ff7-88b0-7dc213e7ce80 req-13aca724-0ec3-4273-9b9a-d23e892b22dd service nova] Releasing lock "refresh_cache-29b8dd5f-1855-490c-a01b-54840073a753" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1868.327200] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ba162a0b-a017-488a-838b-b0c0ba13e0d5 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Acquiring lock "refresh_cache-f4157385-43ea-4e8c-887c-0985af37abd1" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1868.327402] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ba162a0b-a017-488a-838b-b0c0ba13e0d5 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Acquired lock "refresh_cache-f4157385-43ea-4e8c-887c-0985af37abd1" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1868.327580] env[62824]: DEBUG nova.network.neutron [None req-ba162a0b-a017-488a-838b-b0c0ba13e0d5 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] [instance: f4157385-43ea-4e8c-887c-0985af37abd1] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1868.496501] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5eaa800-26b8-4870-851c-a8bc847275d2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1868.508417] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6194fc86-1e1a-4564-a9c3-dc5ed8b2fdf4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1868.547831] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38b3a4ea-4618-42ec-80cd-83282f950830 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1868.555922] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b722dac0-06cf-40eb-b987-0cecdac4f467 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1868.570973] env[62824]: DEBUG nova.compute.provider_tree [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1868.633529] env[62824]: DEBUG oslo_vmware.api [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Task: {'id': task-2145639, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.026691} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1868.633529] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1868.633724] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Moving file from [datastore1] vmware_temp/b40ede64-0abe-49ff-bfc5-8316c59660cf/9e2a7d30-212d-4ab8-9606-c5c6d52629e8 to [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8. {{(pid=62824) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 1868.633943] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-cd7ad520-34cf-4121-9269-822d42554b22 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1868.641807] env[62824]: DEBUG oslo_vmware.api [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Waiting for the task: (returnval){ [ 1868.641807] env[62824]: value = "task-2145640" [ 1868.641807] env[62824]: _type = "Task" [ 1868.641807] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1868.651038] env[62824]: DEBUG oslo_vmware.api [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Task: {'id': task-2145640, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1868.851871] env[62824]: DEBUG nova.network.neutron [None req-ba162a0b-a017-488a-838b-b0c0ba13e0d5 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] [instance: f4157385-43ea-4e8c-887c-0985af37abd1] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1868.907880] env[62824]: DEBUG nova.network.neutron [None req-ba162a0b-a017-488a-838b-b0c0ba13e0d5 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] [instance: f4157385-43ea-4e8c-887c-0985af37abd1] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1869.075472] env[62824]: DEBUG nova.scheduler.client.report [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1869.153430] env[62824]: DEBUG oslo_vmware.api [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Task: {'id': task-2145640, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.024036} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1869.153694] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] File moved {{(pid=62824) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 1869.153883] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Cleaning up location [datastore1] vmware_temp/b40ede64-0abe-49ff-bfc5-8316c59660cf {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1869.154059] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Deleting the datastore file [datastore1] vmware_temp/b40ede64-0abe-49ff-bfc5-8316c59660cf {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1869.154328] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e182f836-5640-47d8-a899-20c4808d0e46 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1869.161024] env[62824]: DEBUG oslo_vmware.api [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Waiting for the task: (returnval){ [ 1869.161024] env[62824]: value = "task-2145641" [ 1869.161024] env[62824]: _type = "Task" [ 1869.161024] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1869.168641] env[62824]: DEBUG oslo_vmware.api [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Task: {'id': task-2145641, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1869.413892] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ba162a0b-a017-488a-838b-b0c0ba13e0d5 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Releasing lock "refresh_cache-f4157385-43ea-4e8c-887c-0985af37abd1" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1869.413892] env[62824]: DEBUG nova.compute.manager [None req-ba162a0b-a017-488a-838b-b0c0ba13e0d5 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] [instance: f4157385-43ea-4e8c-887c-0985af37abd1] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1869.413892] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-ba162a0b-a017-488a-838b-b0c0ba13e0d5 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] [instance: f4157385-43ea-4e8c-887c-0985af37abd1] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1869.413892] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b65442a-bf3e-4add-b2b0-a79d9685ac8b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1869.422651] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba162a0b-a017-488a-838b-b0c0ba13e0d5 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] [instance: f4157385-43ea-4e8c-887c-0985af37abd1] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1869.422924] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fe7319d1-8189-48ad-bd38-e9dfa3fad1bc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1869.430415] env[62824]: DEBUG oslo_vmware.api [None req-ba162a0b-a017-488a-838b-b0c0ba13e0d5 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Waiting for the task: (returnval){ [ 1869.430415] env[62824]: value = "task-2145642" [ 1869.430415] env[62824]: _type = "Task" [ 1869.430415] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1869.440660] env[62824]: DEBUG oslo_vmware.api [None req-ba162a0b-a017-488a-838b-b0c0ba13e0d5 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Task: {'id': task-2145642, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1869.582087] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.521s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1869.582357] env[62824]: DEBUG nova.compute.manager [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] [instance: 04ac4ade-0bc8-4469-9b1b-f288b2f0a367] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1869.585167] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.723s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1869.586757] env[62824]: INFO nova.compute.claims [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1869.681364] env[62824]: DEBUG oslo_vmware.api [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Task: {'id': task-2145641, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.02444} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1869.681657] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1869.682424] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ce354a78-a1c5-489a-a9f8-348f3e0b2f29 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1869.689989] env[62824]: DEBUG oslo_vmware.api [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Waiting for the task: (returnval){ [ 1869.689989] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52445dfb-80e2-ae9f-121f-4137b500fd47" [ 1869.689989] env[62824]: _type = "Task" [ 1869.689989] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1869.702811] env[62824]: DEBUG oslo_vmware.api [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52445dfb-80e2-ae9f-121f-4137b500fd47, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1869.784193] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Acquiring lock "6d80ec06-8559-4964-8577-a2512aa366ed" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1869.784693] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Lock "6d80ec06-8559-4964-8577-a2512aa366ed" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1869.941188] env[62824]: DEBUG oslo_vmware.api [None req-ba162a0b-a017-488a-838b-b0c0ba13e0d5 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Task: {'id': task-2145642, 'name': PowerOffVM_Task, 'duration_secs': 0.215344} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1869.941449] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba162a0b-a017-488a-838b-b0c0ba13e0d5 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] [instance: f4157385-43ea-4e8c-887c-0985af37abd1] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1869.941617] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-ba162a0b-a017-488a-838b-b0c0ba13e0d5 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] [instance: f4157385-43ea-4e8c-887c-0985af37abd1] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1869.941882] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c1051073-e33b-42f4-8dd2-5ad10096fa1b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1869.969823] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-ba162a0b-a017-488a-838b-b0c0ba13e0d5 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] [instance: f4157385-43ea-4e8c-887c-0985af37abd1] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1869.970074] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-ba162a0b-a017-488a-838b-b0c0ba13e0d5 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] [instance: f4157385-43ea-4e8c-887c-0985af37abd1] Deleting contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1869.970265] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-ba162a0b-a017-488a-838b-b0c0ba13e0d5 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Deleting the datastore file [datastore2] f4157385-43ea-4e8c-887c-0985af37abd1 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1869.970529] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-32d6be5a-378c-4f20-ae1c-55e551f28aeb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1869.977922] env[62824]: DEBUG oslo_vmware.api [None req-ba162a0b-a017-488a-838b-b0c0ba13e0d5 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Waiting for the task: (returnval){ [ 1869.977922] env[62824]: value = "task-2145644" [ 1869.977922] env[62824]: _type = "Task" [ 1869.977922] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1869.985389] env[62824]: DEBUG oslo_vmware.api [None req-ba162a0b-a017-488a-838b-b0c0ba13e0d5 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Task: {'id': task-2145644, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1870.091512] env[62824]: DEBUG nova.compute.utils [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1870.096593] env[62824]: DEBUG nova.compute.manager [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] [instance: 04ac4ade-0bc8-4469-9b1b-f288b2f0a367] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1870.096861] env[62824]: DEBUG nova.network.neutron [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] [instance: 04ac4ade-0bc8-4469-9b1b-f288b2f0a367] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1870.207905] env[62824]: DEBUG oslo_vmware.api [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52445dfb-80e2-ae9f-121f-4137b500fd47, 'name': SearchDatastore_Task, 'duration_secs': 0.009978} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1870.208211] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1870.208454] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 29b8dd5f-1855-490c-a01b-54840073a753/29b8dd5f-1855-490c-a01b-54840073a753.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1870.208711] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1a2cd9e1-cda4-49f9-b36f-bfe1ea33616e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1870.216153] env[62824]: DEBUG oslo_vmware.api [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Waiting for the task: (returnval){ [ 1870.216153] env[62824]: value = "task-2145645" [ 1870.216153] env[62824]: _type = "Task" [ 1870.216153] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1870.224641] env[62824]: DEBUG oslo_vmware.api [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Task: {'id': task-2145645, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1870.441733] env[62824]: DEBUG nova.policy [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0b22066e51cd4e5481c4f725deecab84', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1af614b0b5ed4997b0b844648bc44eb4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 1870.491093] env[62824]: DEBUG oslo_vmware.api [None req-ba162a0b-a017-488a-838b-b0c0ba13e0d5 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Task: {'id': task-2145644, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.084072} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1870.491329] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-ba162a0b-a017-488a-838b-b0c0ba13e0d5 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1870.491559] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-ba162a0b-a017-488a-838b-b0c0ba13e0d5 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] [instance: f4157385-43ea-4e8c-887c-0985af37abd1] Deleted contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1870.491775] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-ba162a0b-a017-488a-838b-b0c0ba13e0d5 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] [instance: f4157385-43ea-4e8c-887c-0985af37abd1] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1870.491998] env[62824]: INFO nova.compute.manager [None req-ba162a0b-a017-488a-838b-b0c0ba13e0d5 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] [instance: f4157385-43ea-4e8c-887c-0985af37abd1] Took 1.08 seconds to destroy the instance on the hypervisor. [ 1870.492664] env[62824]: DEBUG oslo.service.loopingcall [None req-ba162a0b-a017-488a-838b-b0c0ba13e0d5 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1870.492945] env[62824]: DEBUG nova.compute.manager [-] [instance: f4157385-43ea-4e8c-887c-0985af37abd1] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1870.493081] env[62824]: DEBUG nova.network.neutron [-] [instance: f4157385-43ea-4e8c-887c-0985af37abd1] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1870.527740] env[62824]: DEBUG nova.network.neutron [-] [instance: f4157385-43ea-4e8c-887c-0985af37abd1] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1870.610395] env[62824]: DEBUG nova.compute.manager [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] [instance: 04ac4ade-0bc8-4469-9b1b-f288b2f0a367] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1870.731397] env[62824]: DEBUG oslo_vmware.api [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Task: {'id': task-2145645, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.463853} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1870.731677] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 29b8dd5f-1855-490c-a01b-54840073a753/29b8dd5f-1855-490c-a01b-54840073a753.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1870.731929] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1870.732223] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8914d980-4811-419e-bafd-d569c44ff4a0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1870.740074] env[62824]: DEBUG oslo_vmware.api [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Waiting for the task: (returnval){ [ 1870.740074] env[62824]: value = "task-2145646" [ 1870.740074] env[62824]: _type = "Task" [ 1870.740074] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1870.749580] env[62824]: DEBUG oslo_vmware.api [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Task: {'id': task-2145646, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1870.937059] env[62824]: DEBUG oslo_concurrency.lockutils [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Acquiring lock "59b5b883-4188-471c-8862-444f3ce08cb0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1870.937331] env[62824]: DEBUG oslo_concurrency.lockutils [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Lock "59b5b883-4188-471c-8862-444f3ce08cb0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1871.012699] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-163e3d23-9914-42df-b5ef-60b9265f3adf {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1871.021224] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-139b9f7f-7af3-4314-adcc-26785ad637fd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1871.063879] env[62824]: DEBUG nova.network.neutron [-] [instance: f4157385-43ea-4e8c-887c-0985af37abd1] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1871.066728] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90ddd5b2-74df-415a-9cb9-80c712dd4a95 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1871.070148] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1871.081325] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-970b65b9-bf2b-4857-bb51-cb71b2192367 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1871.095388] env[62824]: DEBUG nova.compute.provider_tree [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1871.256743] env[62824]: DEBUG oslo_vmware.api [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Task: {'id': task-2145646, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065638} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1871.256743] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1871.257122] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18070b72-2c3d-47eb-8671-a584b51f0d0e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1871.290751] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Reconfiguring VM instance instance-0000000d to attach disk [datastore1] 29b8dd5f-1855-490c-a01b-54840073a753/29b8dd5f-1855-490c-a01b-54840073a753.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1871.291705] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-58e7c22b-ff07-4241-a7d8-29dd8a2a8c63 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1871.314336] env[62824]: DEBUG oslo_vmware.api [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Waiting for the task: (returnval){ [ 1871.314336] env[62824]: value = "task-2145647" [ 1871.314336] env[62824]: _type = "Task" [ 1871.314336] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1871.328625] env[62824]: DEBUG oslo_vmware.api [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Task: {'id': task-2145647, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1871.354609] env[62824]: DEBUG nova.network.neutron [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] [instance: 04ac4ade-0bc8-4469-9b1b-f288b2f0a367] Successfully created port: 086b038c-de36-4e83-b5dd-5857c8b6225e {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1871.572245] env[62824]: INFO nova.compute.manager [-] [instance: f4157385-43ea-4e8c-887c-0985af37abd1] Took 1.08 seconds to deallocate network for instance. [ 1871.580894] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1871.585119] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Starting heal instance info cache {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10445}} [ 1871.585119] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Rebuilding the list of instances to heal {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10449}} [ 1871.598986] env[62824]: DEBUG nova.scheduler.client.report [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1871.624034] env[62824]: DEBUG nova.compute.manager [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] [instance: 04ac4ade-0bc8-4469-9b1b-f288b2f0a367] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1871.668744] env[62824]: DEBUG nova.virt.hardware [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1871.668988] env[62824]: DEBUG nova.virt.hardware [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1871.672738] env[62824]: DEBUG nova.virt.hardware [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1871.672980] env[62824]: DEBUG nova.virt.hardware [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1871.673401] env[62824]: DEBUG nova.virt.hardware [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1871.673401] env[62824]: DEBUG nova.virt.hardware [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1871.673512] env[62824]: DEBUG nova.virt.hardware [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1871.673732] env[62824]: DEBUG nova.virt.hardware [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1871.673832] env[62824]: DEBUG nova.virt.hardware [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1871.673995] env[62824]: DEBUG nova.virt.hardware [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1871.674214] env[62824]: DEBUG nova.virt.hardware [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1871.675685] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-975f0bf9-1c71-4c87-8b7f-f79c9d8cbf16 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1871.689411] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa343043-d36f-4b63-92b7-58d07b67c0a2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1871.834724] env[62824]: DEBUG oslo_vmware.api [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Task: {'id': task-2145647, 'name': ReconfigVM_Task, 'duration_secs': 0.281993} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1871.834724] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Reconfigured VM instance instance-0000000d to attach disk [datastore1] 29b8dd5f-1855-490c-a01b-54840073a753/29b8dd5f-1855-490c-a01b-54840073a753.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1871.835597] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e5313b07-053d-4a65-8beb-afa9bb9068dd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1871.845664] env[62824]: DEBUG oslo_vmware.api [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Waiting for the task: (returnval){ [ 1871.845664] env[62824]: value = "task-2145648" [ 1871.845664] env[62824]: _type = "Task" [ 1871.845664] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1871.856665] env[62824]: DEBUG oslo_vmware.api [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Task: {'id': task-2145648, 'name': Rename_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1872.080368] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ba162a0b-a017-488a-838b-b0c0ba13e0d5 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1872.088608] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: f4157385-43ea-4e8c-887c-0985af37abd1] Skipping network cache update for instance because it is being deleted. {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10462}} [ 1872.088771] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] Skipping network cache update for instance because it is Building. {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10458}} [ 1872.089237] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] Skipping network cache update for instance because it is Building. {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10458}} [ 1872.089237] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] Skipping network cache update for instance because it is Building. {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10458}} [ 1872.089237] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Skipping network cache update for instance because it is Building. {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10458}} [ 1872.089449] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 04ac4ade-0bc8-4469-9b1b-f288b2f0a367] Skipping network cache update for instance because it is Building. {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10458}} [ 1872.089449] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Skipping network cache update for instance because it is Building. {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10458}} [ 1872.089519] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Didn't find any instances for network info cache update. {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10531}} [ 1872.089715] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1872.090000] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1872.090168] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1872.090359] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1872.090543] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1872.090703] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62824) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11064}} [ 1872.107021] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.521s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1872.107021] env[62824]: DEBUG nova.compute.manager [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1872.109728] env[62824]: DEBUG oslo_concurrency.lockutils [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.243s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1872.111242] env[62824]: INFO nova.compute.claims [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] [instance: 82382932-7302-4441-a6f8-9aa2300ec0f6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1872.359465] env[62824]: DEBUG oslo_vmware.api [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Task: {'id': task-2145648, 'name': Rename_Task, 'duration_secs': 0.144257} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1872.359745] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1872.360079] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-76373299-9fa9-4a46-a4ed-5e7d83ff45aa {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1872.367205] env[62824]: DEBUG oslo_vmware.api [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Waiting for the task: (returnval){ [ 1872.367205] env[62824]: value = "task-2145649" [ 1872.367205] env[62824]: _type = "Task" [ 1872.367205] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1872.377655] env[62824]: DEBUG oslo_vmware.api [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Task: {'id': task-2145649, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1872.617665] env[62824]: DEBUG nova.compute.utils [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1872.620053] env[62824]: DEBUG nova.compute.manager [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1872.620053] env[62824]: DEBUG nova.network.neutron [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: f4d63a93-23af-470c-b36a-662af81dc386] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1872.720361] env[62824]: DEBUG nova.policy [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b989eb89498b45dab6604116870d2fa1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c987c5f844be42c99cc9d57a9ca84c31', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 1872.880766] env[62824]: DEBUG oslo_vmware.api [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Task: {'id': task-2145649, 'name': PowerOnVM_Task} progress is 71%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1873.129247] env[62824]: DEBUG nova.compute.manager [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1873.235401] env[62824]: DEBUG nova.network.neutron [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Successfully created port: 5a010977-7d05-47e0-8688-bbf9a0e70900 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1873.258501] env[62824]: DEBUG nova.network.neutron [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] [instance: 04ac4ade-0bc8-4469-9b1b-f288b2f0a367] Successfully updated port: 086b038c-de36-4e83-b5dd-5857c8b6225e {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1873.381501] env[62824]: DEBUG oslo_vmware.api [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Task: {'id': task-2145649, 'name': PowerOnVM_Task, 'duration_secs': 0.681939} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1873.381745] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1873.381779] env[62824]: INFO nova.compute.manager [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Took 14.44 seconds to spawn the instance on the hypervisor. [ 1873.381935] env[62824]: DEBUG nova.compute.manager [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1873.384685] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ce7bea7-8129-4918-a922-622f76f446a4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1873.527204] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d3db3bc-0fff-45a2-ad70-05f6b7819fc9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1873.536034] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca4d1cb5-ac25-4b37-baac-15bdb250a63e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1873.572327] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-248b9b14-9cb2-4b0a-a49b-cc901690ee8b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1873.580532] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8929926f-c495-46cf-b9dc-4353b0e019fc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1873.594770] env[62824]: DEBUG nova.compute.provider_tree [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1873.762407] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Acquiring lock "refresh_cache-04ac4ade-0bc8-4469-9b1b-f288b2f0a367" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1873.762584] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Acquired lock "refresh_cache-04ac4ade-0bc8-4469-9b1b-f288b2f0a367" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1873.762742] env[62824]: DEBUG nova.network.neutron [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] [instance: 04ac4ade-0bc8-4469-9b1b-f288b2f0a367] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1873.923148] env[62824]: INFO nova.compute.manager [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Took 29.70 seconds to build instance. [ 1874.046801] env[62824]: DEBUG nova.compute.manager [req-4d1e4fe3-abd9-4c29-b23e-c744ce73fb60 req-fbe19470-0dee-4c6f-a651-49cb33885c45 service nova] [instance: 04ac4ade-0bc8-4469-9b1b-f288b2f0a367] Received event network-vif-plugged-086b038c-de36-4e83-b5dd-5857c8b6225e {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1874.047540] env[62824]: DEBUG oslo_concurrency.lockutils [req-4d1e4fe3-abd9-4c29-b23e-c744ce73fb60 req-fbe19470-0dee-4c6f-a651-49cb33885c45 service nova] Acquiring lock "04ac4ade-0bc8-4469-9b1b-f288b2f0a367-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1874.047540] env[62824]: DEBUG oslo_concurrency.lockutils [req-4d1e4fe3-abd9-4c29-b23e-c744ce73fb60 req-fbe19470-0dee-4c6f-a651-49cb33885c45 service nova] Lock "04ac4ade-0bc8-4469-9b1b-f288b2f0a367-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1874.047718] env[62824]: DEBUG oslo_concurrency.lockutils [req-4d1e4fe3-abd9-4c29-b23e-c744ce73fb60 req-fbe19470-0dee-4c6f-a651-49cb33885c45 service nova] Lock "04ac4ade-0bc8-4469-9b1b-f288b2f0a367-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1874.047778] env[62824]: DEBUG nova.compute.manager [req-4d1e4fe3-abd9-4c29-b23e-c744ce73fb60 req-fbe19470-0dee-4c6f-a651-49cb33885c45 service nova] [instance: 04ac4ade-0bc8-4469-9b1b-f288b2f0a367] No waiting events found dispatching network-vif-plugged-086b038c-de36-4e83-b5dd-5857c8b6225e {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1874.047938] env[62824]: WARNING nova.compute.manager [req-4d1e4fe3-abd9-4c29-b23e-c744ce73fb60 req-fbe19470-0dee-4c6f-a651-49cb33885c45 service nova] [instance: 04ac4ade-0bc8-4469-9b1b-f288b2f0a367] Received unexpected event network-vif-plugged-086b038c-de36-4e83-b5dd-5857c8b6225e for instance with vm_state building and task_state spawning. [ 1874.097717] env[62824]: DEBUG nova.scheduler.client.report [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1874.147023] env[62824]: DEBUG nova.compute.manager [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1874.168231] env[62824]: DEBUG nova.virt.hardware [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1874.168493] env[62824]: DEBUG nova.virt.hardware [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1874.168647] env[62824]: DEBUG nova.virt.hardware [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1874.168829] env[62824]: DEBUG nova.virt.hardware [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1874.168975] env[62824]: DEBUG nova.virt.hardware [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1874.169146] env[62824]: DEBUG nova.virt.hardware [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1874.169360] env[62824]: DEBUG nova.virt.hardware [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1874.169521] env[62824]: DEBUG nova.virt.hardware [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1874.169691] env[62824]: DEBUG nova.virt.hardware [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1874.169853] env[62824]: DEBUG nova.virt.hardware [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1874.170045] env[62824]: DEBUG nova.virt.hardware [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1874.170905] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1aa95d25-25a1-4caa-be02-4df49ceb2d92 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1874.182406] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5e4f158-a7a1-4ff1-8c4a-18b6735d6f4f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1874.312158] env[62824]: DEBUG nova.network.neutron [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] [instance: 04ac4ade-0bc8-4469-9b1b-f288b2f0a367] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1874.425832] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ae8e5d94-cc21-4022-95ef-aef5c866d0e4 tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Lock "29b8dd5f-1855-490c-a01b-54840073a753" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.051s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1874.515691] env[62824]: DEBUG oslo_concurrency.lockutils [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquiring lock "487c2c9d-2cd2-4912-8613-e1bfac732c40" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1874.515691] env[62824]: DEBUG oslo_concurrency.lockutils [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "487c2c9d-2cd2-4912-8613-e1bfac732c40" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1874.534889] env[62824]: DEBUG nova.network.neutron [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] [instance: 04ac4ade-0bc8-4469-9b1b-f288b2f0a367] Updating instance_info_cache with network_info: [{"id": "086b038c-de36-4e83-b5dd-5857c8b6225e", "address": "fa:16:3e:2c:9b:54", "network": {"id": "b6403dc3-30ef-4c76-b521-380c179f345d", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.51", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "22c5c40c2e004babaa15896c89307303", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1002b79b-224e-41e3-a484-4245a767147a", "external-id": "nsx-vlan-transportzone-353", "segmentation_id": 353, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap086b038c-de", "ovs_interfaceid": "086b038c-de36-4e83-b5dd-5857c8b6225e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1874.602555] env[62824]: DEBUG oslo_concurrency.lockutils [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.493s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1874.603154] env[62824]: DEBUG nova.compute.manager [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] [instance: 82382932-7302-4441-a6f8-9aa2300ec0f6] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1874.606255] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9f666b55-9bba-45f0-98a3-b9d8b6261703 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.683s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1874.928859] env[62824]: DEBUG nova.compute.manager [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1875.042655] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Releasing lock "refresh_cache-04ac4ade-0bc8-4469-9b1b-f288b2f0a367" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1875.043188] env[62824]: DEBUG nova.compute.manager [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] [instance: 04ac4ade-0bc8-4469-9b1b-f288b2f0a367] Instance network_info: |[{"id": "086b038c-de36-4e83-b5dd-5857c8b6225e", "address": "fa:16:3e:2c:9b:54", "network": {"id": "b6403dc3-30ef-4c76-b521-380c179f345d", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.51", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "22c5c40c2e004babaa15896c89307303", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1002b79b-224e-41e3-a484-4245a767147a", "external-id": "nsx-vlan-transportzone-353", "segmentation_id": 353, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap086b038c-de", "ovs_interfaceid": "086b038c-de36-4e83-b5dd-5857c8b6225e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1875.047055] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] [instance: 04ac4ade-0bc8-4469-9b1b-f288b2f0a367] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2c:9b:54', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1002b79b-224e-41e3-a484-4245a767147a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '086b038c-de36-4e83-b5dd-5857c8b6225e', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1875.054620] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Creating folder: Project (1af614b0b5ed4997b0b844648bc44eb4). Parent ref: group-v438503. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1875.055438] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4d8fe7d9-4915-45cb-b7b0-475788b0365a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1875.071019] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Created folder: Project (1af614b0b5ed4997b0b844648bc44eb4) in parent group-v438503. [ 1875.071019] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Creating folder: Instances. Parent ref: group-v438514. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1875.071019] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9eb2724c-53bd-4185-a696-001ec2f4fc3e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1875.082197] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Created folder: Instances in parent group-v438514. [ 1875.082724] env[62824]: DEBUG oslo.service.loopingcall [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1875.086337] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 04ac4ade-0bc8-4469-9b1b-f288b2f0a367] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1875.086577] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a0a8d5b5-1c6e-4228-a11d-d8332179dc25 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1875.112438] env[62824]: DEBUG nova.compute.utils [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1875.116200] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1875.116200] env[62824]: value = "task-2145652" [ 1875.116200] env[62824]: _type = "Task" [ 1875.116200] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1875.116909] env[62824]: DEBUG nova.compute.manager [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] [instance: 82382932-7302-4441-a6f8-9aa2300ec0f6] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1875.117088] env[62824]: DEBUG nova.network.neutron [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] [instance: 82382932-7302-4441-a6f8-9aa2300ec0f6] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1875.130263] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145652, 'name': CreateVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1875.194988] env[62824]: DEBUG nova.policy [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8418cb2af3484bd99daff2045756b468', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0cf3a345a7d54625885c2159edcf0877', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 1875.438398] env[62824]: DEBUG nova.network.neutron [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Successfully updated port: 5a010977-7d05-47e0-8688-bbf9a0e70900 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1875.458624] env[62824]: DEBUG oslo_concurrency.lockutils [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1875.587562] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfbf4f69-ede4-4a31-b3dc-6a43cec9be9b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1875.600962] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-309e7795-1cf6-4d21-b586-baad97bf7bd6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1875.655174] env[62824]: DEBUG nova.compute.manager [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] [instance: 82382932-7302-4441-a6f8-9aa2300ec0f6] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1875.666048] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40a958f4-e2b8-4fb0-aead-cf5ed75dfb07 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1875.670022] env[62824]: DEBUG nova.compute.manager [req-8e23775a-5da9-4f31-808c-e99c0327a01e req-89eea978-4d44-4774-9276-013591714268 service nova] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Received event network-vif-plugged-5a010977-7d05-47e0-8688-bbf9a0e70900 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1875.670022] env[62824]: DEBUG oslo_concurrency.lockutils [req-8e23775a-5da9-4f31-808c-e99c0327a01e req-89eea978-4d44-4774-9276-013591714268 service nova] Acquiring lock "f4d63a93-23af-470c-b36a-662af81dc386-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1875.670022] env[62824]: DEBUG oslo_concurrency.lockutils [req-8e23775a-5da9-4f31-808c-e99c0327a01e req-89eea978-4d44-4774-9276-013591714268 service nova] Lock "f4d63a93-23af-470c-b36a-662af81dc386-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1875.670022] env[62824]: DEBUG oslo_concurrency.lockutils [req-8e23775a-5da9-4f31-808c-e99c0327a01e req-89eea978-4d44-4774-9276-013591714268 service nova] Lock "f4d63a93-23af-470c-b36a-662af81dc386-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1875.670022] env[62824]: DEBUG nova.compute.manager [req-8e23775a-5da9-4f31-808c-e99c0327a01e req-89eea978-4d44-4774-9276-013591714268 service nova] [instance: f4d63a93-23af-470c-b36a-662af81dc386] No waiting events found dispatching network-vif-plugged-5a010977-7d05-47e0-8688-bbf9a0e70900 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1875.670407] env[62824]: WARNING nova.compute.manager [req-8e23775a-5da9-4f31-808c-e99c0327a01e req-89eea978-4d44-4774-9276-013591714268 service nova] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Received unexpected event network-vif-plugged-5a010977-7d05-47e0-8688-bbf9a0e70900 for instance with vm_state building and task_state spawning. [ 1875.678316] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145652, 'name': CreateVM_Task, 'duration_secs': 0.424921} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1875.679546] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6e6c6ef-e2a8-42e1-87dd-8746811142bd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1875.683559] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 04ac4ade-0bc8-4469-9b1b-f288b2f0a367] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1875.684305] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1875.684477] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1875.684797] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1875.685449] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7901bc42-a437-49b0-ab3d-770615498f87 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1875.699620] env[62824]: DEBUG nova.compute.provider_tree [None req-9f666b55-9bba-45f0-98a3-b9d8b6261703 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1875.700346] env[62824]: DEBUG oslo_vmware.api [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Waiting for the task: (returnval){ [ 1875.700346] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]521d0f87-3022-09b7-8515-b7fa7c209771" [ 1875.700346] env[62824]: _type = "Task" [ 1875.700346] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1875.713161] env[62824]: DEBUG oslo_vmware.api [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]521d0f87-3022-09b7-8515-b7fa7c209771, 'name': SearchDatastore_Task, 'duration_secs': 0.011993} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1875.713492] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1875.713729] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] [instance: 04ac4ade-0bc8-4469-9b1b-f288b2f0a367] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1875.713952] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1875.714298] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1875.714366] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1875.715221] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ef98c1e2-2579-4a57-aa48-7059a1dd5fd8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1875.725857] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1875.726072] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1875.726863] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6d215508-9965-4f58-a4f7-423946c8ad49 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1875.734437] env[62824]: DEBUG oslo_vmware.api [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Waiting for the task: (returnval){ [ 1875.734437] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]526b1c3e-dc49-a4a0-ed31-c4f18cfd1c3e" [ 1875.734437] env[62824]: _type = "Task" [ 1875.734437] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1875.745229] env[62824]: DEBUG oslo_vmware.api [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]526b1c3e-dc49-a4a0-ed31-c4f18cfd1c3e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1875.846806] env[62824]: DEBUG nova.network.neutron [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] [instance: 82382932-7302-4441-a6f8-9aa2300ec0f6] Successfully created port: 760062c3-a8b6-4863-a557-0d44a27e3bb3 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1875.940569] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquiring lock "refresh_cache-f4d63a93-23af-470c-b36a-662af81dc386" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1875.940569] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquired lock "refresh_cache-f4d63a93-23af-470c-b36a-662af81dc386" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1875.940729] env[62824]: DEBUG nova.network.neutron [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1876.203017] env[62824]: DEBUG nova.scheduler.client.report [None req-9f666b55-9bba-45f0-98a3-b9d8b6261703 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1876.248443] env[62824]: DEBUG oslo_vmware.api [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]526b1c3e-dc49-a4a0-ed31-c4f18cfd1c3e, 'name': SearchDatastore_Task, 'duration_secs': 0.010215} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1876.249600] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d5e15844-4278-4212-b643-bbd97e2aa693 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1876.256305] env[62824]: DEBUG oslo_vmware.api [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Waiting for the task: (returnval){ [ 1876.256305] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5236b4fc-8737-afc2-eccd-fe50647be38b" [ 1876.256305] env[62824]: _type = "Task" [ 1876.256305] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1876.265316] env[62824]: DEBUG oslo_vmware.api [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5236b4fc-8737-afc2-eccd-fe50647be38b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1876.507529] env[62824]: DEBUG nova.network.neutron [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1876.666642] env[62824]: DEBUG nova.compute.manager [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] [instance: 82382932-7302-4441-a6f8-9aa2300ec0f6] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1876.690821] env[62824]: DEBUG nova.virt.hardware [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1876.690896] env[62824]: DEBUG nova.virt.hardware [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1876.691076] env[62824]: DEBUG nova.virt.hardware [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1876.691262] env[62824]: DEBUG nova.virt.hardware [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1876.691365] env[62824]: DEBUG nova.virt.hardware [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1876.691501] env[62824]: DEBUG nova.virt.hardware [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1876.691697] env[62824]: DEBUG nova.virt.hardware [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1876.691842] env[62824]: DEBUG nova.virt.hardware [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1876.692333] env[62824]: DEBUG nova.virt.hardware [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1876.692333] env[62824]: DEBUG nova.virt.hardware [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1876.692333] env[62824]: DEBUG nova.virt.hardware [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1876.693217] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d9f61fb-10b9-4da9-918d-fc43d6bf3ca3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1876.702846] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f90746b-df0a-4880-863a-9d520d310513 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1876.707705] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9f666b55-9bba-45f0-98a3-b9d8b6261703 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.101s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1876.708078] env[62824]: ERROR nova.compute.manager [None req-9f666b55-9bba-45f0-98a3-b9d8b6261703 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 2bf4c281-010d-4311-be12-2ec99b845218, please check neutron logs for more information. [ 1876.708078] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] Traceback (most recent call last): [ 1876.708078] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] File "/opt/stack/nova/nova/compute/manager.py", line 2648, in _build_and_run_instance [ 1876.708078] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] self.driver.spawn(context, instance, image_meta, [ 1876.708078] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 1876.708078] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] self._vmops.spawn(context, instance, image_meta, injected_files, [ 1876.708078] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 1876.708078] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] vm_ref = self.build_virtual_machine(instance, [ 1876.708078] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 1876.708078] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] vif_infos = vmwarevif.get_vif_info(self._session, [ 1876.708078] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 1876.708479] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] for vif in network_info: [ 1876.708479] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 1876.708479] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] return self._sync_wrapper(fn, *args, **kwargs) [ 1876.708479] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 1876.708479] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] self.wait() [ 1876.708479] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 1876.708479] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] self[:] = self._gt.wait() [ 1876.708479] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 1876.708479] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] return self._exit_event.wait() [ 1876.708479] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 1876.708479] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] result = hub.switch() [ 1876.708479] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 1876.708479] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] return self.greenlet.switch() [ 1876.708803] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 1876.708803] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] result = function(*args, **kwargs) [ 1876.708803] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 1876.708803] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] return func(*args, **kwargs) [ 1876.708803] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] File "/opt/stack/nova/nova/compute/manager.py", line 2018, in _allocate_network_async [ 1876.708803] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] raise e [ 1876.708803] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] File "/opt/stack/nova/nova/compute/manager.py", line 1996, in _allocate_network_async [ 1876.708803] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] nwinfo = self.network_api.allocate_for_instance( [ 1876.708803] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 1876.708803] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] created_port_ids = self._update_ports_for_instance( [ 1876.708803] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 1876.708803] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] with excutils.save_and_reraise_exception(): [ 1876.708803] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1876.709080] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] self.force_reraise() [ 1876.709080] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1876.709080] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] raise self.value [ 1876.709080] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 1876.709080] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] updated_port = self._update_port( [ 1876.709080] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 1876.709080] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] _ensure_no_port_binding_failure(port) [ 1876.709080] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 1876.709080] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] raise exception.PortBindingFailed(port_id=port['id']) [ 1876.709080] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] nova.exception.PortBindingFailed: Binding failed for port 2bf4c281-010d-4311-be12-2ec99b845218, please check neutron logs for more information. [ 1876.709080] env[62824]: ERROR nova.compute.manager [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] [ 1876.709308] env[62824]: DEBUG nova.compute.utils [None req-9f666b55-9bba-45f0-98a3-b9d8b6261703 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] Binding failed for port 2bf4c281-010d-4311-be12-2ec99b845218, please check neutron logs for more information. {{(pid=62824) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 1876.709862] env[62824]: DEBUG oslo_concurrency.lockutils [None req-34e3c185-2b6d-4b05-bb0f-3253c603dd1c tempest-InstanceActionsV221TestJSON-1308314300 tempest-InstanceActionsV221TestJSON-1308314300-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.970s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1876.714570] env[62824]: DEBUG nova.compute.manager [None req-9f666b55-9bba-45f0-98a3-b9d8b6261703 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] Build of instance 00fca8eb-c209-4632-abe6-5f092b04ac56 was re-scheduled: Binding failed for port 2bf4c281-010d-4311-be12-2ec99b845218, please check neutron logs for more information. {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2487}} [ 1876.714570] env[62824]: DEBUG nova.compute.manager [None req-9f666b55-9bba-45f0-98a3-b9d8b6261703 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] Unplugging VIFs for instance {{(pid=62824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3013}} [ 1876.714570] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9f666b55-9bba-45f0-98a3-b9d8b6261703 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Acquiring lock "refresh_cache-00fca8eb-c209-4632-abe6-5f092b04ac56" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1876.714695] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9f666b55-9bba-45f0-98a3-b9d8b6261703 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Acquired lock "refresh_cache-00fca8eb-c209-4632-abe6-5f092b04ac56" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1876.714842] env[62824]: DEBUG nova.network.neutron [None req-9f666b55-9bba-45f0-98a3-b9d8b6261703 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1876.768068] env[62824]: DEBUG oslo_vmware.api [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5236b4fc-8737-afc2-eccd-fe50647be38b, 'name': SearchDatastore_Task, 'duration_secs': 0.017368} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1876.768361] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1876.768698] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 04ac4ade-0bc8-4469-9b1b-f288b2f0a367/04ac4ade-0bc8-4469-9b1b-f288b2f0a367.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1876.768926] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-eb3e2e00-bc29-4b51-81a5-51debbb55401 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1876.778137] env[62824]: DEBUG oslo_vmware.api [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Waiting for the task: (returnval){ [ 1876.778137] env[62824]: value = "task-2145653" [ 1876.778137] env[62824]: _type = "Task" [ 1876.778137] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1876.788471] env[62824]: DEBUG oslo_vmware.api [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Task: {'id': task-2145653, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1877.085560] env[62824]: DEBUG nova.network.neutron [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Updating instance_info_cache with network_info: [{"id": "5a010977-7d05-47e0-8688-bbf9a0e70900", "address": "fa:16:3e:1e:14:06", "network": {"id": "aa1c09e5-8d72-43ad-a903-a0f5e711da80", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1643616572-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c987c5f844be42c99cc9d57a9ca84c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9bc2632-36f9-4912-8782-8bbb789f909d", "external-id": "nsx-vlan-transportzone-897", "segmentation_id": 897, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5a010977-7d", "ovs_interfaceid": "5a010977-7d05-47e0-8688-bbf9a0e70900", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1877.257762] env[62824]: DEBUG nova.network.neutron [None req-9f666b55-9bba-45f0-98a3-b9d8b6261703 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1877.298068] env[62824]: DEBUG oslo_vmware.api [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Task: {'id': task-2145653, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1877.417797] env[62824]: DEBUG nova.compute.manager [req-8777dabb-2352-4aa5-a2e8-767bf1cacab3 req-1204f426-8fe2-4474-889d-114b1dc06a71 service nova] [instance: 04ac4ade-0bc8-4469-9b1b-f288b2f0a367] Received event network-changed-086b038c-de36-4e83-b5dd-5857c8b6225e {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1877.418489] env[62824]: DEBUG nova.compute.manager [req-8777dabb-2352-4aa5-a2e8-767bf1cacab3 req-1204f426-8fe2-4474-889d-114b1dc06a71 service nova] [instance: 04ac4ade-0bc8-4469-9b1b-f288b2f0a367] Refreshing instance network info cache due to event network-changed-086b038c-de36-4e83-b5dd-5857c8b6225e. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1877.420856] env[62824]: DEBUG oslo_concurrency.lockutils [req-8777dabb-2352-4aa5-a2e8-767bf1cacab3 req-1204f426-8fe2-4474-889d-114b1dc06a71 service nova] Acquiring lock "refresh_cache-04ac4ade-0bc8-4469-9b1b-f288b2f0a367" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1877.420856] env[62824]: DEBUG oslo_concurrency.lockutils [req-8777dabb-2352-4aa5-a2e8-767bf1cacab3 req-1204f426-8fe2-4474-889d-114b1dc06a71 service nova] Acquired lock "refresh_cache-04ac4ade-0bc8-4469-9b1b-f288b2f0a367" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1877.420856] env[62824]: DEBUG nova.network.neutron [req-8777dabb-2352-4aa5-a2e8-767bf1cacab3 req-1204f426-8fe2-4474-889d-114b1dc06a71 service nova] [instance: 04ac4ade-0bc8-4469-9b1b-f288b2f0a367] Refreshing network info cache for port 086b038c-de36-4e83-b5dd-5857c8b6225e {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1877.453136] env[62824]: DEBUG nova.network.neutron [None req-9f666b55-9bba-45f0-98a3-b9d8b6261703 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1877.589851] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Releasing lock "refresh_cache-f4d63a93-23af-470c-b36a-662af81dc386" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1877.591579] env[62824]: DEBUG nova.compute.manager [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Instance network_info: |[{"id": "5a010977-7d05-47e0-8688-bbf9a0e70900", "address": "fa:16:3e:1e:14:06", "network": {"id": "aa1c09e5-8d72-43ad-a903-a0f5e711da80", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1643616572-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c987c5f844be42c99cc9d57a9ca84c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9bc2632-36f9-4912-8782-8bbb789f909d", "external-id": "nsx-vlan-transportzone-897", "segmentation_id": 897, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5a010977-7d", "ovs_interfaceid": "5a010977-7d05-47e0-8688-bbf9a0e70900", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1877.592321] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1e:14:06', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c9bc2632-36f9-4912-8782-8bbb789f909d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5a010977-7d05-47e0-8688-bbf9a0e70900', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1877.601337] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Creating folder: Project (c987c5f844be42c99cc9d57a9ca84c31). Parent ref: group-v438503. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1877.601539] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-21afa843-5035-4942-9add-cd58950ea640 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1877.622040] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Created folder: Project (c987c5f844be42c99cc9d57a9ca84c31) in parent group-v438503. [ 1877.624115] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Creating folder: Instances. Parent ref: group-v438517. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1877.626906] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6c82dc39-ccb9-42d9-ac55-0fe8b749bcb5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1877.638007] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Created folder: Instances in parent group-v438517. [ 1877.638113] env[62824]: DEBUG oslo.service.loopingcall [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1877.638257] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1877.638452] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d9d282df-15c4-40a4-b11e-218949a2691f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1877.662625] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1877.662625] env[62824]: value = "task-2145656" [ 1877.662625] env[62824]: _type = "Task" [ 1877.662625] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1877.670845] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145656, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1877.779243] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91b63d69-20a4-459c-8ad6-34bac1ab9c91 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1877.796118] env[62824]: DEBUG oslo_vmware.api [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Task: {'id': task-2145653, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.868274} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1877.798101] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f897f67-52e8-439a-8b13-28bfdda38aba {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1877.800359] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 04ac4ade-0bc8-4469-9b1b-f288b2f0a367/04ac4ade-0bc8-4469-9b1b-f288b2f0a367.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1877.800577] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] [instance: 04ac4ade-0bc8-4469-9b1b-f288b2f0a367] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1877.800817] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-51fce8ae-36c8-4d9b-8ff4-42b09592bf5a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1877.846300] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5ad4ce0-8488-4585-bc23-02fe764a29be {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1877.850346] env[62824]: DEBUG oslo_vmware.api [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Waiting for the task: (returnval){ [ 1877.850346] env[62824]: value = "task-2145657" [ 1877.850346] env[62824]: _type = "Task" [ 1877.850346] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1877.858235] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14813e77-8a16-4ac9-b06f-fc3fbe407439 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1877.865515] env[62824]: DEBUG oslo_vmware.api [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Task: {'id': task-2145657, 'name': ExtendVirtualDisk_Task} progress is 50%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1877.877500] env[62824]: DEBUG nova.compute.provider_tree [None req-34e3c185-2b6d-4b05-bb0f-3253c603dd1c tempest-InstanceActionsV221TestJSON-1308314300 tempest-InstanceActionsV221TestJSON-1308314300-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1877.961036] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9f666b55-9bba-45f0-98a3-b9d8b6261703 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Releasing lock "refresh_cache-00fca8eb-c209-4632-abe6-5f092b04ac56" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1877.961036] env[62824]: DEBUG nova.compute.manager [None req-9f666b55-9bba-45f0-98a3-b9d8b6261703 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3036}} [ 1877.961036] env[62824]: DEBUG nova.compute.manager [None req-9f666b55-9bba-45f0-98a3-b9d8b6261703 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1877.961036] env[62824]: DEBUG nova.network.neutron [None req-9f666b55-9bba-45f0-98a3-b9d8b6261703 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1877.962312] env[62824]: DEBUG oslo_concurrency.lockutils [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Acquiring lock "69bcac33-d6ac-4ee7-b674-58ab25389149" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1877.962570] env[62824]: DEBUG oslo_concurrency.lockutils [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Lock "69bcac33-d6ac-4ee7-b674-58ab25389149" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1878.008885] env[62824]: DEBUG nova.network.neutron [None req-9f666b55-9bba-45f0-98a3-b9d8b6261703 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1878.010513] env[62824]: DEBUG oslo_concurrency.lockutils [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Acquiring lock "5f1237ff-5841-41f1-8d77-df9fabaa62b6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1878.010637] env[62824]: DEBUG oslo_concurrency.lockutils [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Lock "5f1237ff-5841-41f1-8d77-df9fabaa62b6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1878.057149] env[62824]: DEBUG nova.network.neutron [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] [instance: 82382932-7302-4441-a6f8-9aa2300ec0f6] Successfully updated port: 760062c3-a8b6-4863-a557-0d44a27e3bb3 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1878.173034] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145656, 'name': CreateVM_Task, 'duration_secs': 0.391474} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1878.176037] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1878.176703] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1878.176867] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1878.177201] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1878.177756] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-498d005b-c5b8-471f-b9e4-d1bf0e1c4b37 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1878.183283] env[62824]: DEBUG oslo_vmware.api [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Waiting for the task: (returnval){ [ 1878.183283] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52038ca0-dba3-bbe5-edd1-3ee87fa075de" [ 1878.183283] env[62824]: _type = "Task" [ 1878.183283] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1878.192021] env[62824]: DEBUG oslo_vmware.api [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52038ca0-dba3-bbe5-edd1-3ee87fa075de, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1878.335463] env[62824]: DEBUG nova.network.neutron [req-8777dabb-2352-4aa5-a2e8-767bf1cacab3 req-1204f426-8fe2-4474-889d-114b1dc06a71 service nova] [instance: 04ac4ade-0bc8-4469-9b1b-f288b2f0a367] Updated VIF entry in instance network info cache for port 086b038c-de36-4e83-b5dd-5857c8b6225e. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1878.335883] env[62824]: DEBUG nova.network.neutron [req-8777dabb-2352-4aa5-a2e8-767bf1cacab3 req-1204f426-8fe2-4474-889d-114b1dc06a71 service nova] [instance: 04ac4ade-0bc8-4469-9b1b-f288b2f0a367] Updating instance_info_cache with network_info: [{"id": "086b038c-de36-4e83-b5dd-5857c8b6225e", "address": "fa:16:3e:2c:9b:54", "network": {"id": "b6403dc3-30ef-4c76-b521-380c179f345d", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.51", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "22c5c40c2e004babaa15896c89307303", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1002b79b-224e-41e3-a484-4245a767147a", "external-id": "nsx-vlan-transportzone-353", "segmentation_id": 353, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap086b038c-de", "ovs_interfaceid": "086b038c-de36-4e83-b5dd-5857c8b6225e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1878.361091] env[62824]: DEBUG oslo_vmware.api [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Task: {'id': task-2145657, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066102} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1878.361374] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] [instance: 04ac4ade-0bc8-4469-9b1b-f288b2f0a367] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1878.362464] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02d73a88-533e-45b0-bc9b-644c9bb61a2c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1878.386438] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] [instance: 04ac4ade-0bc8-4469-9b1b-f288b2f0a367] Reconfiguring VM instance instance-0000000e to attach disk [datastore2] 04ac4ade-0bc8-4469-9b1b-f288b2f0a367/04ac4ade-0bc8-4469-9b1b-f288b2f0a367.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1878.387339] env[62824]: DEBUG nova.scheduler.client.report [None req-34e3c185-2b6d-4b05-bb0f-3253c603dd1c tempest-InstanceActionsV221TestJSON-1308314300 tempest-InstanceActionsV221TestJSON-1308314300-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1878.394982] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1098893b-25ce-4243-941c-fd08a3aea63e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1878.422829] env[62824]: DEBUG oslo_vmware.api [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Waiting for the task: (returnval){ [ 1878.422829] env[62824]: value = "task-2145658" [ 1878.422829] env[62824]: _type = "Task" [ 1878.422829] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1878.432127] env[62824]: DEBUG oslo_vmware.api [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Task: {'id': task-2145658, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1878.501572] env[62824]: DEBUG nova.compute.manager [req-da15937c-1bd1-4413-a9a5-48febeddc0b5 req-65ebc1d5-3187-436b-aa8c-14fdbc479a93 service nova] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Received event network-changed-5a010977-7d05-47e0-8688-bbf9a0e70900 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1878.501893] env[62824]: DEBUG nova.compute.manager [req-da15937c-1bd1-4413-a9a5-48febeddc0b5 req-65ebc1d5-3187-436b-aa8c-14fdbc479a93 service nova] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Refreshing instance network info cache due to event network-changed-5a010977-7d05-47e0-8688-bbf9a0e70900. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1878.502013] env[62824]: DEBUG oslo_concurrency.lockutils [req-da15937c-1bd1-4413-a9a5-48febeddc0b5 req-65ebc1d5-3187-436b-aa8c-14fdbc479a93 service nova] Acquiring lock "refresh_cache-f4d63a93-23af-470c-b36a-662af81dc386" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1878.502185] env[62824]: DEBUG oslo_concurrency.lockutils [req-da15937c-1bd1-4413-a9a5-48febeddc0b5 req-65ebc1d5-3187-436b-aa8c-14fdbc479a93 service nova] Acquired lock "refresh_cache-f4d63a93-23af-470c-b36a-662af81dc386" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1878.502405] env[62824]: DEBUG nova.network.neutron [req-da15937c-1bd1-4413-a9a5-48febeddc0b5 req-65ebc1d5-3187-436b-aa8c-14fdbc479a93 service nova] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Refreshing network info cache for port 5a010977-7d05-47e0-8688-bbf9a0e70900 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1878.511992] env[62824]: DEBUG nova.network.neutron [None req-9f666b55-9bba-45f0-98a3-b9d8b6261703 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1878.560856] env[62824]: DEBUG oslo_concurrency.lockutils [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Acquiring lock "refresh_cache-82382932-7302-4441-a6f8-9aa2300ec0f6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1878.561434] env[62824]: DEBUG oslo_concurrency.lockutils [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Acquired lock "refresh_cache-82382932-7302-4441-a6f8-9aa2300ec0f6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1878.561434] env[62824]: DEBUG nova.network.neutron [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] [instance: 82382932-7302-4441-a6f8-9aa2300ec0f6] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1878.698632] env[62824]: DEBUG oslo_vmware.api [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52038ca0-dba3-bbe5-edd1-3ee87fa075de, 'name': SearchDatastore_Task, 'duration_secs': 0.020702} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1878.698936] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1878.698936] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1878.699142] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1878.699960] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1878.699960] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1878.699960] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c4b61045-265b-4e17-a732-ba7a1a551070 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1878.717560] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1878.718146] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1878.718772] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eb38c1cf-77f5-451e-8e42-e20e860059c9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1878.725057] env[62824]: DEBUG oslo_vmware.api [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Waiting for the task: (returnval){ [ 1878.725057] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52880171-f341-1a48-62b8-7c54cc02b4cb" [ 1878.725057] env[62824]: _type = "Task" [ 1878.725057] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1878.733847] env[62824]: DEBUG oslo_vmware.api [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52880171-f341-1a48-62b8-7c54cc02b4cb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1878.839695] env[62824]: DEBUG oslo_concurrency.lockutils [req-8777dabb-2352-4aa5-a2e8-767bf1cacab3 req-1204f426-8fe2-4474-889d-114b1dc06a71 service nova] Releasing lock "refresh_cache-04ac4ade-0bc8-4469-9b1b-f288b2f0a367" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1878.896726] env[62824]: DEBUG oslo_concurrency.lockutils [None req-34e3c185-2b6d-4b05-bb0f-3253c603dd1c tempest-InstanceActionsV221TestJSON-1308314300 tempest-InstanceActionsV221TestJSON-1308314300-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.186s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1878.897152] env[62824]: ERROR nova.compute.manager [None req-34e3c185-2b6d-4b05-bb0f-3253c603dd1c tempest-InstanceActionsV221TestJSON-1308314300 tempest-InstanceActionsV221TestJSON-1308314300-project-member] [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d07db52b-09d8-4f22-95da-96a68cc5247c, please check neutron logs for more information. [ 1878.897152] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] Traceback (most recent call last): [ 1878.897152] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] File "/opt/stack/nova/nova/compute/manager.py", line 2648, in _build_and_run_instance [ 1878.897152] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] self.driver.spawn(context, instance, image_meta, [ 1878.897152] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 1878.897152] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 1878.897152] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 1878.897152] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] vm_ref = self.build_virtual_machine(instance, [ 1878.897152] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 1878.897152] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] vif_infos = vmwarevif.get_vif_info(self._session, [ 1878.897152] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 1878.897461] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] for vif in network_info: [ 1878.897461] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 1878.897461] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] return self._sync_wrapper(fn, *args, **kwargs) [ 1878.897461] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 1878.897461] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] self.wait() [ 1878.897461] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 1878.897461] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] self[:] = self._gt.wait() [ 1878.897461] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 1878.897461] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] return self._exit_event.wait() [ 1878.897461] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 1878.897461] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] result = hub.switch() [ 1878.897461] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 1878.897461] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] return self.greenlet.switch() [ 1878.897758] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 1878.897758] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] result = function(*args, **kwargs) [ 1878.897758] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 1878.897758] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] return func(*args, **kwargs) [ 1878.897758] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] File "/opt/stack/nova/nova/compute/manager.py", line 2018, in _allocate_network_async [ 1878.897758] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] raise e [ 1878.897758] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] File "/opt/stack/nova/nova/compute/manager.py", line 1996, in _allocate_network_async [ 1878.897758] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] nwinfo = self.network_api.allocate_for_instance( [ 1878.897758] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 1878.897758] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] created_port_ids = self._update_ports_for_instance( [ 1878.897758] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 1878.897758] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] with excutils.save_and_reraise_exception(): [ 1878.897758] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1878.898067] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] self.force_reraise() [ 1878.898067] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1878.898067] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] raise self.value [ 1878.898067] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 1878.898067] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] updated_port = self._update_port( [ 1878.898067] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 1878.898067] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] _ensure_no_port_binding_failure(port) [ 1878.898067] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 1878.898067] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] raise exception.PortBindingFailed(port_id=port['id']) [ 1878.898067] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] nova.exception.PortBindingFailed: Binding failed for port d07db52b-09d8-4f22-95da-96a68cc5247c, please check neutron logs for more information. [ 1878.898067] env[62824]: ERROR nova.compute.manager [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] [ 1878.898349] env[62824]: DEBUG nova.compute.utils [None req-34e3c185-2b6d-4b05-bb0f-3253c603dd1c tempest-InstanceActionsV221TestJSON-1308314300 tempest-InstanceActionsV221TestJSON-1308314300-project-member] [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] Binding failed for port d07db52b-09d8-4f22-95da-96a68cc5247c, please check neutron logs for more information. {{(pid=62824) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 1878.899694] env[62824]: DEBUG nova.compute.manager [None req-34e3c185-2b6d-4b05-bb0f-3253c603dd1c tempest-InstanceActionsV221TestJSON-1308314300 tempest-InstanceActionsV221TestJSON-1308314300-project-member] [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] Build of instance bec6441f-72b8-4664-b125-b0aa5f21636a was re-scheduled: Binding failed for port d07db52b-09d8-4f22-95da-96a68cc5247c, please check neutron logs for more information. {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2487}} [ 1878.900175] env[62824]: DEBUG nova.compute.manager [None req-34e3c185-2b6d-4b05-bb0f-3253c603dd1c tempest-InstanceActionsV221TestJSON-1308314300 tempest-InstanceActionsV221TestJSON-1308314300-project-member] [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] Unplugging VIFs for instance {{(pid=62824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3013}} [ 1878.900450] env[62824]: DEBUG oslo_concurrency.lockutils [None req-34e3c185-2b6d-4b05-bb0f-3253c603dd1c tempest-InstanceActionsV221TestJSON-1308314300 tempest-InstanceActionsV221TestJSON-1308314300-project-member] Acquiring lock "refresh_cache-bec6441f-72b8-4664-b125-b0aa5f21636a" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1878.900586] env[62824]: DEBUG oslo_concurrency.lockutils [None req-34e3c185-2b6d-4b05-bb0f-3253c603dd1c tempest-InstanceActionsV221TestJSON-1308314300 tempest-InstanceActionsV221TestJSON-1308314300-project-member] Acquired lock "refresh_cache-bec6441f-72b8-4664-b125-b0aa5f21636a" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1878.900752] env[62824]: DEBUG nova.network.neutron [None req-34e3c185-2b6d-4b05-bb0f-3253c603dd1c tempest-InstanceActionsV221TestJSON-1308314300 tempest-InstanceActionsV221TestJSON-1308314300-project-member] [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1878.902144] env[62824]: DEBUG oslo_concurrency.lockutils [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.389s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1878.904033] env[62824]: INFO nova.compute.claims [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1878.935656] env[62824]: DEBUG oslo_vmware.api [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Task: {'id': task-2145658, 'name': ReconfigVM_Task, 'duration_secs': 0.284068} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1878.936021] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] [instance: 04ac4ade-0bc8-4469-9b1b-f288b2f0a367] Reconfigured VM instance instance-0000000e to attach disk [datastore2] 04ac4ade-0bc8-4469-9b1b-f288b2f0a367/04ac4ade-0bc8-4469-9b1b-f288b2f0a367.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1878.936715] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9d38d01e-04be-42a2-9057-c7d47830ee9d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1878.944855] env[62824]: DEBUG oslo_vmware.api [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Waiting for the task: (returnval){ [ 1878.944855] env[62824]: value = "task-2145659" [ 1878.944855] env[62824]: _type = "Task" [ 1878.944855] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1878.954263] env[62824]: DEBUG oslo_vmware.api [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Task: {'id': task-2145659, 'name': Rename_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1879.014436] env[62824]: INFO nova.compute.manager [None req-9f666b55-9bba-45f0-98a3-b9d8b6261703 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: 00fca8eb-c209-4632-abe6-5f092b04ac56] Took 1.05 seconds to deallocate network for instance. [ 1879.101778] env[62824]: DEBUG nova.network.neutron [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] [instance: 82382932-7302-4441-a6f8-9aa2300ec0f6] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1879.238621] env[62824]: DEBUG oslo_vmware.api [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52880171-f341-1a48-62b8-7c54cc02b4cb, 'name': SearchDatastore_Task, 'duration_secs': 0.015482} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1879.242757] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-17b6dc8f-4117-42cd-ade7-bea230b25de6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1879.249288] env[62824]: DEBUG oslo_vmware.api [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Waiting for the task: (returnval){ [ 1879.249288] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]529a92c4-fc46-3552-42ee-8ac5c3698074" [ 1879.249288] env[62824]: _type = "Task" [ 1879.249288] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1879.258140] env[62824]: DEBUG oslo_vmware.api [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]529a92c4-fc46-3552-42ee-8ac5c3698074, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1879.294082] env[62824]: DEBUG nova.network.neutron [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] [instance: 82382932-7302-4441-a6f8-9aa2300ec0f6] Updating instance_info_cache with network_info: [{"id": "760062c3-a8b6-4863-a557-0d44a27e3bb3", "address": "fa:16:3e:d4:15:ee", "network": {"id": "668ff2d5-a4e1-488b-b71f-7f45c130d0e0", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-616618532-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0cf3a345a7d54625885c2159edcf0877", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "57691231-2b8d-4d71-8f79-d4a6a1d95ec8", "external-id": "nsx-vlan-transportzone-373", "segmentation_id": 373, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap760062c3-a8", "ovs_interfaceid": "760062c3-a8b6-4863-a557-0d44a27e3bb3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1879.432234] env[62824]: DEBUG nova.network.neutron [None req-34e3c185-2b6d-4b05-bb0f-3253c603dd1c tempest-InstanceActionsV221TestJSON-1308314300 tempest-InstanceActionsV221TestJSON-1308314300-project-member] [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1879.460021] env[62824]: DEBUG oslo_vmware.api [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Task: {'id': task-2145659, 'name': Rename_Task, 'duration_secs': 0.139477} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1879.460021] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] [instance: 04ac4ade-0bc8-4469-9b1b-f288b2f0a367] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1879.460021] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1d6f70d5-fdbf-4cdc-9718-de4b673635af {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1879.468753] env[62824]: DEBUG oslo_vmware.api [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Waiting for the task: (returnval){ [ 1879.468753] env[62824]: value = "task-2145660" [ 1879.468753] env[62824]: _type = "Task" [ 1879.468753] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1879.479803] env[62824]: DEBUG oslo_vmware.api [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Task: {'id': task-2145660, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1879.530073] env[62824]: DEBUG nova.network.neutron [req-da15937c-1bd1-4413-a9a5-48febeddc0b5 req-65ebc1d5-3187-436b-aa8c-14fdbc479a93 service nova] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Updated VIF entry in instance network info cache for port 5a010977-7d05-47e0-8688-bbf9a0e70900. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1879.530480] env[62824]: DEBUG nova.network.neutron [req-da15937c-1bd1-4413-a9a5-48febeddc0b5 req-65ebc1d5-3187-436b-aa8c-14fdbc479a93 service nova] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Updating instance_info_cache with network_info: [{"id": "5a010977-7d05-47e0-8688-bbf9a0e70900", "address": "fa:16:3e:1e:14:06", "network": {"id": "aa1c09e5-8d72-43ad-a903-a0f5e711da80", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1643616572-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c987c5f844be42c99cc9d57a9ca84c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9bc2632-36f9-4912-8782-8bbb789f909d", "external-id": "nsx-vlan-transportzone-897", "segmentation_id": 897, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5a010977-7d", "ovs_interfaceid": "5a010977-7d05-47e0-8688-bbf9a0e70900", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1879.536447] env[62824]: DEBUG nova.network.neutron [None req-34e3c185-2b6d-4b05-bb0f-3253c603dd1c tempest-InstanceActionsV221TestJSON-1308314300 tempest-InstanceActionsV221TestJSON-1308314300-project-member] [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1879.652029] env[62824]: DEBUG nova.compute.manager [req-405f179b-ac52-4fc9-8c14-be91c1810781 req-10535818-3596-4440-b80c-a50c9f0f1246 service nova] [instance: 82382932-7302-4441-a6f8-9aa2300ec0f6] Received event network-vif-plugged-760062c3-a8b6-4863-a557-0d44a27e3bb3 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1879.652295] env[62824]: DEBUG oslo_concurrency.lockutils [req-405f179b-ac52-4fc9-8c14-be91c1810781 req-10535818-3596-4440-b80c-a50c9f0f1246 service nova] Acquiring lock "82382932-7302-4441-a6f8-9aa2300ec0f6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1879.652518] env[62824]: DEBUG oslo_concurrency.lockutils [req-405f179b-ac52-4fc9-8c14-be91c1810781 req-10535818-3596-4440-b80c-a50c9f0f1246 service nova] Lock "82382932-7302-4441-a6f8-9aa2300ec0f6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1879.652691] env[62824]: DEBUG oslo_concurrency.lockutils [req-405f179b-ac52-4fc9-8c14-be91c1810781 req-10535818-3596-4440-b80c-a50c9f0f1246 service nova] Lock "82382932-7302-4441-a6f8-9aa2300ec0f6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1879.652862] env[62824]: DEBUG nova.compute.manager [req-405f179b-ac52-4fc9-8c14-be91c1810781 req-10535818-3596-4440-b80c-a50c9f0f1246 service nova] [instance: 82382932-7302-4441-a6f8-9aa2300ec0f6] No waiting events found dispatching network-vif-plugged-760062c3-a8b6-4863-a557-0d44a27e3bb3 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1879.653276] env[62824]: WARNING nova.compute.manager [req-405f179b-ac52-4fc9-8c14-be91c1810781 req-10535818-3596-4440-b80c-a50c9f0f1246 service nova] [instance: 82382932-7302-4441-a6f8-9aa2300ec0f6] Received unexpected event network-vif-plugged-760062c3-a8b6-4863-a557-0d44a27e3bb3 for instance with vm_state building and task_state spawning. [ 1879.653533] env[62824]: DEBUG nova.compute.manager [req-405f179b-ac52-4fc9-8c14-be91c1810781 req-10535818-3596-4440-b80c-a50c9f0f1246 service nova] [instance: 82382932-7302-4441-a6f8-9aa2300ec0f6] Received event network-changed-760062c3-a8b6-4863-a557-0d44a27e3bb3 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1879.653705] env[62824]: DEBUG nova.compute.manager [req-405f179b-ac52-4fc9-8c14-be91c1810781 req-10535818-3596-4440-b80c-a50c9f0f1246 service nova] [instance: 82382932-7302-4441-a6f8-9aa2300ec0f6] Refreshing instance network info cache due to event network-changed-760062c3-a8b6-4863-a557-0d44a27e3bb3. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1879.653866] env[62824]: DEBUG oslo_concurrency.lockutils [req-405f179b-ac52-4fc9-8c14-be91c1810781 req-10535818-3596-4440-b80c-a50c9f0f1246 service nova] Acquiring lock "refresh_cache-82382932-7302-4441-a6f8-9aa2300ec0f6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1879.761856] env[62824]: DEBUG oslo_vmware.api [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]529a92c4-fc46-3552-42ee-8ac5c3698074, 'name': SearchDatastore_Task, 'duration_secs': 0.036952} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1879.762117] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1879.763851] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] f4d63a93-23af-470c-b36a-662af81dc386/f4d63a93-23af-470c-b36a-662af81dc386.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1879.764030] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-be9e72ed-7394-4154-adb2-0c9477ea9dcb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1879.774121] env[62824]: DEBUG oslo_vmware.api [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Waiting for the task: (returnval){ [ 1879.774121] env[62824]: value = "task-2145661" [ 1879.774121] env[62824]: _type = "Task" [ 1879.774121] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1879.786733] env[62824]: DEBUG oslo_vmware.api [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2145661, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1879.799434] env[62824]: DEBUG oslo_concurrency.lockutils [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Releasing lock "refresh_cache-82382932-7302-4441-a6f8-9aa2300ec0f6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1879.799776] env[62824]: DEBUG nova.compute.manager [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] [instance: 82382932-7302-4441-a6f8-9aa2300ec0f6] Instance network_info: |[{"id": "760062c3-a8b6-4863-a557-0d44a27e3bb3", "address": "fa:16:3e:d4:15:ee", "network": {"id": "668ff2d5-a4e1-488b-b71f-7f45c130d0e0", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-616618532-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0cf3a345a7d54625885c2159edcf0877", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "57691231-2b8d-4d71-8f79-d4a6a1d95ec8", "external-id": "nsx-vlan-transportzone-373", "segmentation_id": 373, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap760062c3-a8", "ovs_interfaceid": "760062c3-a8b6-4863-a557-0d44a27e3bb3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1879.800085] env[62824]: DEBUG oslo_concurrency.lockutils [req-405f179b-ac52-4fc9-8c14-be91c1810781 req-10535818-3596-4440-b80c-a50c9f0f1246 service nova] Acquired lock "refresh_cache-82382932-7302-4441-a6f8-9aa2300ec0f6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1879.800268] env[62824]: DEBUG nova.network.neutron [req-405f179b-ac52-4fc9-8c14-be91c1810781 req-10535818-3596-4440-b80c-a50c9f0f1246 service nova] [instance: 82382932-7302-4441-a6f8-9aa2300ec0f6] Refreshing network info cache for port 760062c3-a8b6-4863-a557-0d44a27e3bb3 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1879.801570] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] [instance: 82382932-7302-4441-a6f8-9aa2300ec0f6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d4:15:ee', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '57691231-2b8d-4d71-8f79-d4a6a1d95ec8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '760062c3-a8b6-4863-a557-0d44a27e3bb3', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1879.809788] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Creating folder: Project (0cf3a345a7d54625885c2159edcf0877). Parent ref: group-v438503. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1879.810454] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bb2aebe1-aacd-463b-b53d-ab0120000bed {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1879.824218] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Created folder: Project (0cf3a345a7d54625885c2159edcf0877) in parent group-v438503. [ 1879.824450] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Creating folder: Instances. Parent ref: group-v438520. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1879.824704] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-166c34c2-ce9e-4a11-885c-d04706528231 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1879.838076] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Created folder: Instances in parent group-v438520. [ 1879.838076] env[62824]: DEBUG oslo.service.loopingcall [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1879.838076] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 82382932-7302-4441-a6f8-9aa2300ec0f6] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1879.838076] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7f7c4518-bf10-4203-967d-d01ac8db5259 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1879.858649] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1879.858649] env[62824]: value = "task-2145664" [ 1879.858649] env[62824]: _type = "Task" [ 1879.858649] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1879.868521] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145664, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1879.986603] env[62824]: DEBUG oslo_vmware.api [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Task: {'id': task-2145660, 'name': PowerOnVM_Task, 'duration_secs': 0.458245} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1879.986972] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] [instance: 04ac4ade-0bc8-4469-9b1b-f288b2f0a367] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1879.987202] env[62824]: INFO nova.compute.manager [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] [instance: 04ac4ade-0bc8-4469-9b1b-f288b2f0a367] Took 8.36 seconds to spawn the instance on the hypervisor. [ 1879.987451] env[62824]: DEBUG nova.compute.manager [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] [instance: 04ac4ade-0bc8-4469-9b1b-f288b2f0a367] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1879.988508] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d1efb3b-ba4b-4c09-852f-87c2b8be4577 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1880.033078] env[62824]: DEBUG oslo_concurrency.lockutils [req-da15937c-1bd1-4413-a9a5-48febeddc0b5 req-65ebc1d5-3187-436b-aa8c-14fdbc479a93 service nova] Releasing lock "refresh_cache-f4d63a93-23af-470c-b36a-662af81dc386" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1880.033177] env[62824]: DEBUG nova.compute.manager [req-da15937c-1bd1-4413-a9a5-48febeddc0b5 req-65ebc1d5-3187-436b-aa8c-14fdbc479a93 service nova] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Received event network-changed-2d861cfc-3ad4-4d40-ad77-e4530d363421 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1880.033296] env[62824]: DEBUG nova.compute.manager [req-da15937c-1bd1-4413-a9a5-48febeddc0b5 req-65ebc1d5-3187-436b-aa8c-14fdbc479a93 service nova] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Refreshing instance network info cache due to event network-changed-2d861cfc-3ad4-4d40-ad77-e4530d363421. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1880.033503] env[62824]: DEBUG oslo_concurrency.lockutils [req-da15937c-1bd1-4413-a9a5-48febeddc0b5 req-65ebc1d5-3187-436b-aa8c-14fdbc479a93 service nova] Acquiring lock "refresh_cache-29b8dd5f-1855-490c-a01b-54840073a753" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1880.033929] env[62824]: DEBUG oslo_concurrency.lockutils [req-da15937c-1bd1-4413-a9a5-48febeddc0b5 req-65ebc1d5-3187-436b-aa8c-14fdbc479a93 service nova] Acquired lock "refresh_cache-29b8dd5f-1855-490c-a01b-54840073a753" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1880.033929] env[62824]: DEBUG nova.network.neutron [req-da15937c-1bd1-4413-a9a5-48febeddc0b5 req-65ebc1d5-3187-436b-aa8c-14fdbc479a93 service nova] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Refreshing network info cache for port 2d861cfc-3ad4-4d40-ad77-e4530d363421 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1880.039658] env[62824]: DEBUG oslo_concurrency.lockutils [None req-34e3c185-2b6d-4b05-bb0f-3253c603dd1c tempest-InstanceActionsV221TestJSON-1308314300 tempest-InstanceActionsV221TestJSON-1308314300-project-member] Releasing lock "refresh_cache-bec6441f-72b8-4664-b125-b0aa5f21636a" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1880.041214] env[62824]: DEBUG nova.compute.manager [None req-34e3c185-2b6d-4b05-bb0f-3253c603dd1c tempest-InstanceActionsV221TestJSON-1308314300 tempest-InstanceActionsV221TestJSON-1308314300-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3036}} [ 1880.041214] env[62824]: DEBUG nova.compute.manager [None req-34e3c185-2b6d-4b05-bb0f-3253c603dd1c tempest-InstanceActionsV221TestJSON-1308314300 tempest-InstanceActionsV221TestJSON-1308314300-project-member] [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1880.041214] env[62824]: DEBUG nova.network.neutron [None req-34e3c185-2b6d-4b05-bb0f-3253c603dd1c tempest-InstanceActionsV221TestJSON-1308314300 tempest-InstanceActionsV221TestJSON-1308314300-project-member] [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1880.047301] env[62824]: INFO nova.scheduler.client.report [None req-9f666b55-9bba-45f0-98a3-b9d8b6261703 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Deleted allocations for instance 00fca8eb-c209-4632-abe6-5f092b04ac56 [ 1880.064189] env[62824]: DEBUG nova.network.neutron [None req-34e3c185-2b6d-4b05-bb0f-3253c603dd1c tempest-InstanceActionsV221TestJSON-1308314300 tempest-InstanceActionsV221TestJSON-1308314300-project-member] [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1880.289525] env[62824]: DEBUG oslo_vmware.api [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2145661, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1880.375285] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145664, 'name': CreateVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1880.443330] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b21b8c84-684c-4e7e-8a7a-b0b1b7529b7d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1880.451458] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aff05854-5c45-4b0d-9cf9-d586e578671e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1880.487442] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87ef6773-3fd6-442d-b4bc-1cbc48a16c50 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1880.499634] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4899bea5-aa9d-4aca-b345-274855f5ea89 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1880.520045] env[62824]: DEBUG nova.compute.provider_tree [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1880.523283] env[62824]: INFO nova.compute.manager [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] [instance: 04ac4ade-0bc8-4469-9b1b-f288b2f0a367] Took 31.77 seconds to build instance. [ 1880.558022] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9f666b55-9bba-45f0-98a3-b9d8b6261703 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Lock "00fca8eb-c209-4632-abe6-5f092b04ac56" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 52.515s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1880.568132] env[62824]: DEBUG nova.network.neutron [None req-34e3c185-2b6d-4b05-bb0f-3253c603dd1c tempest-InstanceActionsV221TestJSON-1308314300 tempest-InstanceActionsV221TestJSON-1308314300-project-member] [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1880.597919] env[62824]: DEBUG nova.network.neutron [req-405f179b-ac52-4fc9-8c14-be91c1810781 req-10535818-3596-4440-b80c-a50c9f0f1246 service nova] [instance: 82382932-7302-4441-a6f8-9aa2300ec0f6] Updated VIF entry in instance network info cache for port 760062c3-a8b6-4863-a557-0d44a27e3bb3. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1880.598404] env[62824]: DEBUG nova.network.neutron [req-405f179b-ac52-4fc9-8c14-be91c1810781 req-10535818-3596-4440-b80c-a50c9f0f1246 service nova] [instance: 82382932-7302-4441-a6f8-9aa2300ec0f6] Updating instance_info_cache with network_info: [{"id": "760062c3-a8b6-4863-a557-0d44a27e3bb3", "address": "fa:16:3e:d4:15:ee", "network": {"id": "668ff2d5-a4e1-488b-b71f-7f45c130d0e0", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-616618532-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0cf3a345a7d54625885c2159edcf0877", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "57691231-2b8d-4d71-8f79-d4a6a1d95ec8", "external-id": "nsx-vlan-transportzone-373", "segmentation_id": 373, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap760062c3-a8", "ovs_interfaceid": "760062c3-a8b6-4863-a557-0d44a27e3bb3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1880.787878] env[62824]: DEBUG oslo_vmware.api [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2145661, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.631436} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1880.788272] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] f4d63a93-23af-470c-b36a-662af81dc386/f4d63a93-23af-470c-b36a-662af81dc386.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1880.788344] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1880.788586] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-80e26e97-61b7-44e0-980c-6df1b9b4b6e1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1880.797603] env[62824]: DEBUG oslo_vmware.api [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Waiting for the task: (returnval){ [ 1880.797603] env[62824]: value = "task-2145665" [ 1880.797603] env[62824]: _type = "Task" [ 1880.797603] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1880.807838] env[62824]: DEBUG oslo_vmware.api [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2145665, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1880.878496] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145664, 'name': CreateVM_Task, 'duration_secs': 0.573431} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1880.878681] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 82382932-7302-4441-a6f8-9aa2300ec0f6] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1880.879379] env[62824]: DEBUG oslo_concurrency.lockutils [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1880.879544] env[62824]: DEBUG oslo_concurrency.lockutils [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1880.879901] env[62824]: DEBUG oslo_concurrency.lockutils [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1880.880184] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-77afc043-c99b-47e0-99e1-93abb03752f7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1880.887502] env[62824]: DEBUG oslo_vmware.api [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Waiting for the task: (returnval){ [ 1880.887502] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52c6e27d-7411-8892-7f5d-f31b24380327" [ 1880.887502] env[62824]: _type = "Task" [ 1880.887502] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1880.897777] env[62824]: DEBUG oslo_vmware.api [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52c6e27d-7411-8892-7f5d-f31b24380327, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1880.944454] env[62824]: DEBUG nova.network.neutron [req-da15937c-1bd1-4413-a9a5-48febeddc0b5 req-65ebc1d5-3187-436b-aa8c-14fdbc479a93 service nova] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Updated VIF entry in instance network info cache for port 2d861cfc-3ad4-4d40-ad77-e4530d363421. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1880.944823] env[62824]: DEBUG nova.network.neutron [req-da15937c-1bd1-4413-a9a5-48febeddc0b5 req-65ebc1d5-3187-436b-aa8c-14fdbc479a93 service nova] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Updating instance_info_cache with network_info: [{"id": "2d861cfc-3ad4-4d40-ad77-e4530d363421", "address": "fa:16:3e:59:84:dd", "network": {"id": "f18b4230-e950-4957-a02b-107d27729346", "bridge": "br-int", "label": "tempest-VolumesAssistedSnapshotsTest-633702241-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.144", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e2e61ea652f94257b884d2e1b495e446", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4b5c60ce-845e-4506-bc10-348461fece6d", "external-id": "nsx-vlan-transportzone-831", "segmentation_id": 831, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2d861cfc-3a", "ovs_interfaceid": "2d861cfc-3ad4-4d40-ad77-e4530d363421", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1881.024481] env[62824]: DEBUG nova.scheduler.client.report [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1881.030025] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0437e645-42c5-4488-a72b-2afcd8e301b9 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Lock "04ac4ade-0bc8-4469-9b1b-f288b2f0a367" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 47.151s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1881.060565] env[62824]: DEBUG nova.compute.manager [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] [instance: dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1881.072253] env[62824]: INFO nova.compute.manager [None req-34e3c185-2b6d-4b05-bb0f-3253c603dd1c tempest-InstanceActionsV221TestJSON-1308314300 tempest-InstanceActionsV221TestJSON-1308314300-project-member] [instance: bec6441f-72b8-4664-b125-b0aa5f21636a] Took 1.03 seconds to deallocate network for instance. [ 1881.103743] env[62824]: DEBUG oslo_concurrency.lockutils [req-405f179b-ac52-4fc9-8c14-be91c1810781 req-10535818-3596-4440-b80c-a50c9f0f1246 service nova] Releasing lock "refresh_cache-82382932-7302-4441-a6f8-9aa2300ec0f6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1881.309572] env[62824]: DEBUG oslo_vmware.api [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2145665, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.0728} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1881.309649] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1881.311484] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cec3338-be01-4101-ba0e-ac40f956ab05 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1881.339488] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Reconfiguring VM instance instance-00000010 to attach disk [datastore2] f4d63a93-23af-470c-b36a-662af81dc386/f4d63a93-23af-470c-b36a-662af81dc386.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1881.340296] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dbdd51fb-9103-4783-90e0-884009b75c9b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1881.365354] env[62824]: DEBUG oslo_vmware.api [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Waiting for the task: (returnval){ [ 1881.365354] env[62824]: value = "task-2145666" [ 1881.365354] env[62824]: _type = "Task" [ 1881.365354] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1881.377077] env[62824]: DEBUG oslo_vmware.api [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2145666, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1881.401243] env[62824]: DEBUG oslo_vmware.api [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52c6e27d-7411-8892-7f5d-f31b24380327, 'name': SearchDatastore_Task, 'duration_secs': 0.021291} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1881.401243] env[62824]: DEBUG oslo_concurrency.lockutils [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1881.401243] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] [instance: 82382932-7302-4441-a6f8-9aa2300ec0f6] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1881.401243] env[62824]: DEBUG oslo_concurrency.lockutils [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1881.401454] env[62824]: DEBUG oslo_concurrency.lockutils [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1881.401482] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1881.402369] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e5d899fd-aa8d-4272-b56b-704dfad726c9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1881.412122] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1881.412316] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1881.413119] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f2bf6f96-2046-4026-90f5-92822447ddb6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1881.419520] env[62824]: DEBUG oslo_vmware.api [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Waiting for the task: (returnval){ [ 1881.419520] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52a9139f-4f70-8eca-4260-913edc44d8a2" [ 1881.419520] env[62824]: _type = "Task" [ 1881.419520] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1881.431363] env[62824]: DEBUG oslo_vmware.api [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52a9139f-4f70-8eca-4260-913edc44d8a2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1881.447905] env[62824]: DEBUG oslo_concurrency.lockutils [req-da15937c-1bd1-4413-a9a5-48febeddc0b5 req-65ebc1d5-3187-436b-aa8c-14fdbc479a93 service nova] Releasing lock "refresh_cache-29b8dd5f-1855-490c-a01b-54840073a753" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1881.531045] env[62824]: DEBUG oslo_concurrency.lockutils [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.628s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1881.533201] env[62824]: DEBUG nova.compute.manager [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1881.535744] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.960s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1881.537184] env[62824]: INFO nova.compute.claims [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] [instance: 517bf2cf-e142-4f36-bf53-79a8d2d8a530] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1881.540605] env[62824]: DEBUG nova.compute.manager [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] [instance: 4dfe9f82-7f89-486b-b159-948656d2896c] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1881.600460] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1881.879430] env[62824]: DEBUG oslo_vmware.api [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2145666, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1881.932884] env[62824]: DEBUG oslo_vmware.api [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52a9139f-4f70-8eca-4260-913edc44d8a2, 'name': SearchDatastore_Task, 'duration_secs': 0.011089} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1881.933776] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0f505803-2ce4-46e4-b6b4-2ad98f2ba949 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1881.940971] env[62824]: DEBUG oslo_vmware.api [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Waiting for the task: (returnval){ [ 1881.940971] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]524985b8-4584-58f6-0662-807989aef40a" [ 1881.940971] env[62824]: _type = "Task" [ 1881.940971] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1881.952911] env[62824]: DEBUG oslo_vmware.api [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]524985b8-4584-58f6-0662-807989aef40a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1881.955475] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fa6dedf3-2458-4e1e-a821-8c6b0570fd79 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Acquiring lock "04ac4ade-0bc8-4469-9b1b-f288b2f0a367" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1881.956215] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fa6dedf3-2458-4e1e-a821-8c6b0570fd79 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Lock "04ac4ade-0bc8-4469-9b1b-f288b2f0a367" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1881.956215] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fa6dedf3-2458-4e1e-a821-8c6b0570fd79 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Acquiring lock "04ac4ade-0bc8-4469-9b1b-f288b2f0a367-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1881.956215] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fa6dedf3-2458-4e1e-a821-8c6b0570fd79 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Lock "04ac4ade-0bc8-4469-9b1b-f288b2f0a367-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1881.956379] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fa6dedf3-2458-4e1e-a821-8c6b0570fd79 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Lock "04ac4ade-0bc8-4469-9b1b-f288b2f0a367-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1881.958846] env[62824]: INFO nova.compute.manager [None req-fa6dedf3-2458-4e1e-a821-8c6b0570fd79 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] [instance: 04ac4ade-0bc8-4469-9b1b-f288b2f0a367] Terminating instance [ 1882.048542] env[62824]: DEBUG nova.compute.utils [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1882.051729] env[62824]: DEBUG nova.compute.manager [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1882.051729] env[62824]: DEBUG nova.network.neutron [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1882.080656] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1882.112186] env[62824]: INFO nova.scheduler.client.report [None req-34e3c185-2b6d-4b05-bb0f-3253c603dd1c tempest-InstanceActionsV221TestJSON-1308314300 tempest-InstanceActionsV221TestJSON-1308314300-project-member] Deleted allocations for instance bec6441f-72b8-4664-b125-b0aa5f21636a [ 1882.165663] env[62824]: DEBUG nova.policy [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0c461e862de44bb8b66545d6ec3b1b0c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd041a8209e3848eba8e47251e013ca17', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 1882.376722] env[62824]: DEBUG oslo_vmware.api [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2145666, 'name': ReconfigVM_Task, 'duration_secs': 0.836337} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1882.377099] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Reconfigured VM instance instance-00000010 to attach disk [datastore2] f4d63a93-23af-470c-b36a-662af81dc386/f4d63a93-23af-470c-b36a-662af81dc386.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1882.377861] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4cf70de6-a331-4a59-8c4e-383349e79c7c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1882.385623] env[62824]: DEBUG oslo_vmware.api [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Waiting for the task: (returnval){ [ 1882.385623] env[62824]: value = "task-2145667" [ 1882.385623] env[62824]: _type = "Task" [ 1882.385623] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1882.396109] env[62824]: DEBUG oslo_vmware.api [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2145667, 'name': Rename_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1882.455762] env[62824]: DEBUG oslo_vmware.api [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]524985b8-4584-58f6-0662-807989aef40a, 'name': SearchDatastore_Task, 'duration_secs': 0.00974} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1882.457096] env[62824]: DEBUG oslo_concurrency.lockutils [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1882.457096] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 82382932-7302-4441-a6f8-9aa2300ec0f6/82382932-7302-4441-a6f8-9aa2300ec0f6.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1882.457096] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7f11cf69-65df-44d1-b38e-74095da6bb8a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1882.464868] env[62824]: DEBUG nova.compute.manager [None req-fa6dedf3-2458-4e1e-a821-8c6b0570fd79 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] [instance: 04ac4ade-0bc8-4469-9b1b-f288b2f0a367] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1882.465192] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-fa6dedf3-2458-4e1e-a821-8c6b0570fd79 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] [instance: 04ac4ade-0bc8-4469-9b1b-f288b2f0a367] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1882.467233] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d06a0f2b-49c4-4777-9be3-a83d55e8bec5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1882.472186] env[62824]: DEBUG oslo_vmware.api [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Waiting for the task: (returnval){ [ 1882.472186] env[62824]: value = "task-2145668" [ 1882.472186] env[62824]: _type = "Task" [ 1882.472186] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1882.478251] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa6dedf3-2458-4e1e-a821-8c6b0570fd79 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] [instance: 04ac4ade-0bc8-4469-9b1b-f288b2f0a367] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1882.478965] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-726075e9-b5db-4c7e-8f62-9b7e7b76efb2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1882.485272] env[62824]: DEBUG oslo_vmware.api [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Task: {'id': task-2145668, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1882.488696] env[62824]: DEBUG oslo_vmware.api [None req-fa6dedf3-2458-4e1e-a821-8c6b0570fd79 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Waiting for the task: (returnval){ [ 1882.488696] env[62824]: value = "task-2145669" [ 1882.488696] env[62824]: _type = "Task" [ 1882.488696] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1882.503644] env[62824]: DEBUG oslo_vmware.api [None req-fa6dedf3-2458-4e1e-a821-8c6b0570fd79 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Task: {'id': task-2145669, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1882.557587] env[62824]: DEBUG nova.compute.manager [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1882.626438] env[62824]: DEBUG oslo_concurrency.lockutils [None req-34e3c185-2b6d-4b05-bb0f-3253c603dd1c tempest-InstanceActionsV221TestJSON-1308314300 tempest-InstanceActionsV221TestJSON-1308314300-project-member] Lock "bec6441f-72b8-4664-b125-b0aa5f21636a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 53.955s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1882.899920] env[62824]: DEBUG oslo_vmware.api [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2145667, 'name': Rename_Task, 'duration_secs': 0.232532} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1882.900237] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1882.900466] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-83a082c5-6cbb-4708-bc5a-06077b90dce1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1882.913192] env[62824]: DEBUG oslo_vmware.api [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Waiting for the task: (returnval){ [ 1882.913192] env[62824]: value = "task-2145670" [ 1882.913192] env[62824]: _type = "Task" [ 1882.913192] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1882.933594] env[62824]: DEBUG oslo_vmware.api [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2145670, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1882.935037] env[62824]: DEBUG nova.network.neutron [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c] Successfully created port: 084f765d-6af5-4867-bb76-af443b0bf834 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1882.986482] env[62824]: DEBUG oslo_vmware.api [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Task: {'id': task-2145668, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1883.011877] env[62824]: DEBUG oslo_vmware.api [None req-fa6dedf3-2458-4e1e-a821-8c6b0570fd79 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Task: {'id': task-2145669, 'name': PowerOffVM_Task, 'duration_secs': 0.209172} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1883.012647] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa6dedf3-2458-4e1e-a821-8c6b0570fd79 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] [instance: 04ac4ade-0bc8-4469-9b1b-f288b2f0a367] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1883.012845] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-fa6dedf3-2458-4e1e-a821-8c6b0570fd79 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] [instance: 04ac4ade-0bc8-4469-9b1b-f288b2f0a367] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1883.013279] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0c01ece7-8c48-4da3-a99a-3c590c85d7ba {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1883.068130] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22160f44-3e69-46f8-9bc4-313f941c990e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1883.077970] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a31903d8-0e7d-472b-80a9-06557392fd6d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1883.123022] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fcea255-5553-4390-b379-2dcaae453bcf {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1883.123022] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-fa6dedf3-2458-4e1e-a821-8c6b0570fd79 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] [instance: 04ac4ade-0bc8-4469-9b1b-f288b2f0a367] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1883.123212] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-fa6dedf3-2458-4e1e-a821-8c6b0570fd79 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] [instance: 04ac4ade-0bc8-4469-9b1b-f288b2f0a367] Deleting contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1883.123212] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-fa6dedf3-2458-4e1e-a821-8c6b0570fd79 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Deleting the datastore file [datastore2] 04ac4ade-0bc8-4469-9b1b-f288b2f0a367 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1883.123999] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9be0ae38-a726-4a5d-b137-481dcd7f6418 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1883.130456] env[62824]: DEBUG nova.compute.manager [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1883.135794] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70c3fd8a-8c7a-4612-852d-0b1bb0cb57e8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1883.140413] env[62824]: DEBUG oslo_vmware.api [None req-fa6dedf3-2458-4e1e-a821-8c6b0570fd79 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Waiting for the task: (returnval){ [ 1883.140413] env[62824]: value = "task-2145672" [ 1883.140413] env[62824]: _type = "Task" [ 1883.140413] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1883.153337] env[62824]: DEBUG nova.compute.provider_tree [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1883.158484] env[62824]: DEBUG oslo_vmware.api [None req-fa6dedf3-2458-4e1e-a821-8c6b0570fd79 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Task: {'id': task-2145672, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1883.424808] env[62824]: DEBUG oslo_vmware.api [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2145670, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1883.484427] env[62824]: DEBUG oslo_vmware.api [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Task: {'id': task-2145668, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.555589} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1883.484427] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 82382932-7302-4441-a6f8-9aa2300ec0f6/82382932-7302-4441-a6f8-9aa2300ec0f6.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1883.484696] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] [instance: 82382932-7302-4441-a6f8-9aa2300ec0f6] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1883.484973] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2818831d-b9ad-4f3c-bb02-3e46154607a3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1883.494957] env[62824]: DEBUG oslo_vmware.api [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Waiting for the task: (returnval){ [ 1883.494957] env[62824]: value = "task-2145673" [ 1883.494957] env[62824]: _type = "Task" [ 1883.494957] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1883.505847] env[62824]: DEBUG oslo_vmware.api [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Task: {'id': task-2145673, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1883.568163] env[62824]: DEBUG nova.compute.manager [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1883.591736] env[62824]: DEBUG nova.virt.hardware [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:50:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='64227196',id=22,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_0-1469961458',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1883.592050] env[62824]: DEBUG nova.virt.hardware [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1883.592183] env[62824]: DEBUG nova.virt.hardware [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1883.592372] env[62824]: DEBUG nova.virt.hardware [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1883.592521] env[62824]: DEBUG nova.virt.hardware [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1883.592683] env[62824]: DEBUG nova.virt.hardware [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1883.592895] env[62824]: DEBUG nova.virt.hardware [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1883.593150] env[62824]: DEBUG nova.virt.hardware [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1883.593240] env[62824]: DEBUG nova.virt.hardware [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1883.593406] env[62824]: DEBUG nova.virt.hardware [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1883.593579] env[62824]: DEBUG nova.virt.hardware [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1883.594509] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bfc1b00-be47-4990-9811-a153e5de534a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1883.603374] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab68c5f8-fc56-4703-a986-28cbb25770a7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1883.656265] env[62824]: DEBUG nova.scheduler.client.report [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1883.659906] env[62824]: DEBUG oslo_vmware.api [None req-fa6dedf3-2458-4e1e-a821-8c6b0570fd79 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Task: {'id': task-2145672, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.157528} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1883.660405] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-fa6dedf3-2458-4e1e-a821-8c6b0570fd79 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1883.660664] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-fa6dedf3-2458-4e1e-a821-8c6b0570fd79 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] [instance: 04ac4ade-0bc8-4469-9b1b-f288b2f0a367] Deleted contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1883.660941] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-fa6dedf3-2458-4e1e-a821-8c6b0570fd79 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] [instance: 04ac4ade-0bc8-4469-9b1b-f288b2f0a367] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1883.661225] env[62824]: INFO nova.compute.manager [None req-fa6dedf3-2458-4e1e-a821-8c6b0570fd79 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] [instance: 04ac4ade-0bc8-4469-9b1b-f288b2f0a367] Took 1.20 seconds to destroy the instance on the hypervisor. [ 1883.661611] env[62824]: DEBUG oslo.service.loopingcall [None req-fa6dedf3-2458-4e1e-a821-8c6b0570fd79 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1883.662717] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1883.663256] env[62824]: DEBUG nova.compute.manager [-] [instance: 04ac4ade-0bc8-4469-9b1b-f288b2f0a367] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1883.663452] env[62824]: DEBUG nova.network.neutron [-] [instance: 04ac4ade-0bc8-4469-9b1b-f288b2f0a367] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1883.932874] env[62824]: DEBUG oslo_vmware.api [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2145670, 'name': PowerOnVM_Task, 'duration_secs': 0.58767} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1883.932874] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1883.932874] env[62824]: INFO nova.compute.manager [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Took 9.79 seconds to spawn the instance on the hypervisor. [ 1883.932874] env[62824]: DEBUG nova.compute.manager [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1883.932874] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f306078-acad-46ca-b63a-c14a15e46e88 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1884.022449] env[62824]: DEBUG oslo_vmware.api [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Task: {'id': task-2145673, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.097938} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1884.023287] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] [instance: 82382932-7302-4441-a6f8-9aa2300ec0f6] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1884.024264] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e78158b3-314e-412a-9dc2-6400593bd1fc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1884.055879] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] [instance: 82382932-7302-4441-a6f8-9aa2300ec0f6] Reconfiguring VM instance instance-0000000f to attach disk [datastore2] 82382932-7302-4441-a6f8-9aa2300ec0f6/82382932-7302-4441-a6f8-9aa2300ec0f6.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1884.056353] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8951dc7b-de9b-4b38-953a-6e9722f51462 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1884.089024] env[62824]: DEBUG oslo_vmware.api [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Waiting for the task: (returnval){ [ 1884.089024] env[62824]: value = "task-2145674" [ 1884.089024] env[62824]: _type = "Task" [ 1884.089024] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1884.100557] env[62824]: DEBUG oslo_vmware.api [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Task: {'id': task-2145674, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1884.164558] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.629s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1884.165112] env[62824]: DEBUG nova.compute.manager [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] [instance: 517bf2cf-e142-4f36-bf53-79a8d2d8a530] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1884.168393] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5cdb74c7-a27b-45c4-a89a-19c7714f04b5 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 22.424s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1884.381897] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Acquiring lock "dc3317e4-6958-4df6-8e39-c89b4a5d861c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1884.381897] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Lock "dc3317e4-6958-4df6-8e39-c89b4a5d861c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1884.388105] env[62824]: DEBUG nova.compute.manager [req-2c45252d-07b6-45a1-b2d1-598efd8e812c req-ccce690e-382f-45cd-b03d-0eccfff01eb6 service nova] [instance: 04ac4ade-0bc8-4469-9b1b-f288b2f0a367] Received event network-vif-deleted-086b038c-de36-4e83-b5dd-5857c8b6225e {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1884.388303] env[62824]: INFO nova.compute.manager [req-2c45252d-07b6-45a1-b2d1-598efd8e812c req-ccce690e-382f-45cd-b03d-0eccfff01eb6 service nova] [instance: 04ac4ade-0bc8-4469-9b1b-f288b2f0a367] Neutron deleted interface 086b038c-de36-4e83-b5dd-5857c8b6225e; detaching it from the instance and deleting it from the info cache [ 1884.388534] env[62824]: DEBUG nova.network.neutron [req-2c45252d-07b6-45a1-b2d1-598efd8e812c req-ccce690e-382f-45cd-b03d-0eccfff01eb6 service nova] [instance: 04ac4ade-0bc8-4469-9b1b-f288b2f0a367] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1884.459448] env[62824]: INFO nova.compute.manager [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Took 33.64 seconds to build instance. [ 1884.600627] env[62824]: DEBUG oslo_vmware.api [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Task: {'id': task-2145674, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1884.603758] env[62824]: DEBUG nova.network.neutron [-] [instance: 04ac4ade-0bc8-4469-9b1b-f288b2f0a367] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1884.673554] env[62824]: DEBUG nova.compute.utils [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1884.681588] env[62824]: DEBUG nova.compute.manager [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] [instance: 517bf2cf-e142-4f36-bf53-79a8d2d8a530] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1884.682148] env[62824]: DEBUG nova.network.neutron [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] [instance: 517bf2cf-e142-4f36-bf53-79a8d2d8a530] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1884.752683] env[62824]: DEBUG nova.compute.manager [req-b34f4cef-3ea4-4750-b843-63f19643eaf5 req-6accb973-1214-492c-8c06-165839a6a006 service nova] [instance: 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c] Received event network-vif-plugged-084f765d-6af5-4867-bb76-af443b0bf834 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1884.752907] env[62824]: DEBUG oslo_concurrency.lockutils [req-b34f4cef-3ea4-4750-b843-63f19643eaf5 req-6accb973-1214-492c-8c06-165839a6a006 service nova] Acquiring lock "9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1884.753133] env[62824]: DEBUG oslo_concurrency.lockutils [req-b34f4cef-3ea4-4750-b843-63f19643eaf5 req-6accb973-1214-492c-8c06-165839a6a006 service nova] Lock "9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1884.753302] env[62824]: DEBUG oslo_concurrency.lockutils [req-b34f4cef-3ea4-4750-b843-63f19643eaf5 req-6accb973-1214-492c-8c06-165839a6a006 service nova] Lock "9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1884.753473] env[62824]: DEBUG nova.compute.manager [req-b34f4cef-3ea4-4750-b843-63f19643eaf5 req-6accb973-1214-492c-8c06-165839a6a006 service nova] [instance: 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c] No waiting events found dispatching network-vif-plugged-084f765d-6af5-4867-bb76-af443b0bf834 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1884.753637] env[62824]: WARNING nova.compute.manager [req-b34f4cef-3ea4-4750-b843-63f19643eaf5 req-6accb973-1214-492c-8c06-165839a6a006 service nova] [instance: 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c] Received unexpected event network-vif-plugged-084f765d-6af5-4867-bb76-af443b0bf834 for instance with vm_state building and task_state spawning. [ 1884.897019] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-00c938ee-ddef-49d7-a284-abd3d56926c7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1884.902035] env[62824]: DEBUG nova.network.neutron [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c] Successfully updated port: 084f765d-6af5-4867-bb76-af443b0bf834 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1884.908071] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-329c7af0-7f80-4ed9-b4f6-eb2041734c0d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1884.941147] env[62824]: DEBUG nova.compute.manager [req-2c45252d-07b6-45a1-b2d1-598efd8e812c req-ccce690e-382f-45cd-b03d-0eccfff01eb6 service nova] [instance: 04ac4ade-0bc8-4469-9b1b-f288b2f0a367] Detach interface failed, port_id=086b038c-de36-4e83-b5dd-5857c8b6225e, reason: Instance 04ac4ade-0bc8-4469-9b1b-f288b2f0a367 could not be found. {{(pid=62824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11482}} [ 1884.962856] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2d0c0380-db44-4191-9c31-02bd535c6449 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lock "f4d63a93-23af-470c-b36a-662af81dc386" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 48.892s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1884.979525] env[62824]: DEBUG nova.policy [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '16d8bfd2c5264817a312b1911adc1a37', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dbe0dd011b0c43328d69799d09fc050c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 1885.105268] env[62824]: DEBUG oslo_vmware.api [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Task: {'id': task-2145674, 'name': ReconfigVM_Task, 'duration_secs': 0.774396} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1885.107029] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] [instance: 82382932-7302-4441-a6f8-9aa2300ec0f6] Reconfigured VM instance instance-0000000f to attach disk [datastore2] 82382932-7302-4441-a6f8-9aa2300ec0f6/82382932-7302-4441-a6f8-9aa2300ec0f6.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1885.107029] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-94727966-e56a-43a6-9bc8-7130c9c821b4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1885.109467] env[62824]: INFO nova.compute.manager [-] [instance: 04ac4ade-0bc8-4469-9b1b-f288b2f0a367] Took 1.44 seconds to deallocate network for instance. [ 1885.118466] env[62824]: DEBUG oslo_vmware.api [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Waiting for the task: (returnval){ [ 1885.118466] env[62824]: value = "task-2145675" [ 1885.118466] env[62824]: _type = "Task" [ 1885.118466] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1885.128570] env[62824]: DEBUG oslo_vmware.api [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Task: {'id': task-2145675, 'name': Rename_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1885.182599] env[62824]: DEBUG nova.compute.manager [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] [instance: 517bf2cf-e142-4f36-bf53-79a8d2d8a530] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1885.206871] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1aadc913-823e-4f73-bc15-c048c8517422 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1885.215393] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2fbeff2-a338-4ba6-b5c9-299f31ce3335 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1885.257034] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f2ee072-c676-4a17-9da0-341cde286384 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1885.266087] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90eecac1-66b0-4e41-aaa0-674bf17dcab4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1885.282882] env[62824]: DEBUG nova.compute.provider_tree [None req-5cdb74c7-a27b-45c4-a89a-19c7714f04b5 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1885.405711] env[62824]: DEBUG oslo_concurrency.lockutils [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Acquiring lock "refresh_cache-9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1885.409104] env[62824]: DEBUG oslo_concurrency.lockutils [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Acquired lock "refresh_cache-9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1885.409104] env[62824]: DEBUG nova.network.neutron [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1885.471792] env[62824]: DEBUG nova.compute.manager [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1885.619661] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fa6dedf3-2458-4e1e-a821-8c6b0570fd79 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1885.639168] env[62824]: DEBUG oslo_vmware.api [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Task: {'id': task-2145675, 'name': Rename_Task, 'duration_secs': 0.464665} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1885.640035] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] [instance: 82382932-7302-4441-a6f8-9aa2300ec0f6] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1885.640035] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-07101e2d-efa2-4d61-b99a-7d3c6aa75f27 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1885.648248] env[62824]: DEBUG oslo_vmware.api [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Waiting for the task: (returnval){ [ 1885.648248] env[62824]: value = "task-2145676" [ 1885.648248] env[62824]: _type = "Task" [ 1885.648248] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1885.666145] env[62824]: DEBUG oslo_vmware.api [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Task: {'id': task-2145676, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1885.691312] env[62824]: DEBUG nova.network.neutron [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] [instance: 517bf2cf-e142-4f36-bf53-79a8d2d8a530] Successfully created port: a245bb15-a02d-4ed0-a558-e3be831df931 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1885.787288] env[62824]: DEBUG nova.scheduler.client.report [None req-5cdb74c7-a27b-45c4-a89a-19c7714f04b5 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1885.946923] env[62824]: DEBUG nova.network.neutron [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1886.015586] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1886.160659] env[62824]: DEBUG oslo_vmware.api [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Task: {'id': task-2145676, 'name': PowerOnVM_Task} progress is 1%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1886.196190] env[62824]: DEBUG nova.compute.manager [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] [instance: 517bf2cf-e142-4f36-bf53-79a8d2d8a530] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1886.234763] env[62824]: DEBUG nova.virt.hardware [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1886.235121] env[62824]: DEBUG nova.virt.hardware [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1886.235290] env[62824]: DEBUG nova.virt.hardware [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1886.235480] env[62824]: DEBUG nova.virt.hardware [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1886.235627] env[62824]: DEBUG nova.virt.hardware [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1886.235773] env[62824]: DEBUG nova.virt.hardware [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1886.235988] env[62824]: DEBUG nova.virt.hardware [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1886.236180] env[62824]: DEBUG nova.virt.hardware [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1886.236353] env[62824]: DEBUG nova.virt.hardware [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1886.236522] env[62824]: DEBUG nova.virt.hardware [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1886.236723] env[62824]: DEBUG nova.virt.hardware [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1886.238136] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81e38a9e-26ab-4467-b209-54397a738621 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1886.248459] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ac56699-c645-45b8-bcd2-964e5ad1be7f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1886.294791] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5cdb74c7-a27b-45c4-a89a-19c7714f04b5 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.126s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1886.297290] env[62824]: ERROR nova.compute.manager [None req-5cdb74c7-a27b-45c4-a89a-19c7714f04b5 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694-project-member] [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e42d44c3-e2b4-45a5-b707-7d81a18b96b0, please check neutron logs for more information. [ 1886.297290] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] Traceback (most recent call last): [ 1886.297290] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] File "/opt/stack/nova/nova/compute/manager.py", line 2648, in _build_and_run_instance [ 1886.297290] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] self.driver.spawn(context, instance, image_meta, [ 1886.297290] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 1886.297290] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 1886.297290] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 1886.297290] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] vm_ref = self.build_virtual_machine(instance, [ 1886.297290] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 1886.297290] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] vif_infos = vmwarevif.get_vif_info(self._session, [ 1886.297290] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 1886.297596] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] for vif in network_info: [ 1886.297596] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 1886.297596] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] return self._sync_wrapper(fn, *args, **kwargs) [ 1886.297596] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 1886.297596] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] self.wait() [ 1886.297596] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 1886.297596] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] self[:] = self._gt.wait() [ 1886.297596] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 1886.297596] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] return self._exit_event.wait() [ 1886.297596] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 1886.297596] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] result = hub.switch() [ 1886.297596] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 1886.297596] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] return self.greenlet.switch() [ 1886.297935] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 1886.297935] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] result = function(*args, **kwargs) [ 1886.297935] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 1886.297935] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] return func(*args, **kwargs) [ 1886.297935] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] File "/opt/stack/nova/nova/compute/manager.py", line 2018, in _allocate_network_async [ 1886.297935] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] raise e [ 1886.297935] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] File "/opt/stack/nova/nova/compute/manager.py", line 1996, in _allocate_network_async [ 1886.297935] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] nwinfo = self.network_api.allocate_for_instance( [ 1886.297935] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 1886.297935] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] created_port_ids = self._update_ports_for_instance( [ 1886.297935] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 1886.297935] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] with excutils.save_and_reraise_exception(): [ 1886.297935] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1886.298263] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] self.force_reraise() [ 1886.298263] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1886.298263] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] raise self.value [ 1886.298263] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 1886.298263] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] updated_port = self._update_port( [ 1886.298263] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 1886.298263] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] _ensure_no_port_binding_failure(port) [ 1886.298263] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 1886.298263] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] raise exception.PortBindingFailed(port_id=port['id']) [ 1886.298263] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] nova.exception.PortBindingFailed: Binding failed for port e42d44c3-e2b4-45a5-b707-7d81a18b96b0, please check neutron logs for more information. [ 1886.298263] env[62824]: ERROR nova.compute.manager [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] [ 1886.298563] env[62824]: DEBUG nova.compute.utils [None req-5cdb74c7-a27b-45c4-a89a-19c7714f04b5 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694-project-member] [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] Binding failed for port e42d44c3-e2b4-45a5-b707-7d81a18b96b0, please check neutron logs for more information. {{(pid=62824) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 1886.301415] env[62824]: DEBUG nova.network.neutron [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c] Updating instance_info_cache with network_info: [{"id": "084f765d-6af5-4867-bb76-af443b0bf834", "address": "fa:16:3e:ac:49:e4", "network": {"id": "6f888a0c-8089-4dd4-aace-14b020e11bee", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-137042394-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d041a8209e3848eba8e47251e013ca17", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5f60c972-a72d-4c5f-a250-faadfd6eafbe", "external-id": "nsx-vlan-transportzone-932", "segmentation_id": 932, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap084f765d-6a", "ovs_interfaceid": "084f765d-6af5-4867-bb76-af443b0bf834", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1886.302432] env[62824]: DEBUG oslo_concurrency.lockutils [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.062s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1886.304577] env[62824]: INFO nova.compute.claims [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] [instance: 9062e606-511f-4fe7-9621-90b3c0e51eec] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1886.308391] env[62824]: DEBUG nova.compute.manager [None req-5cdb74c7-a27b-45c4-a89a-19c7714f04b5 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694-project-member] [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] Build of instance 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c was re-scheduled: Binding failed for port e42d44c3-e2b4-45a5-b707-7d81a18b96b0, please check neutron logs for more information. {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2487}} [ 1886.310066] env[62824]: DEBUG nova.compute.manager [None req-5cdb74c7-a27b-45c4-a89a-19c7714f04b5 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694-project-member] [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] Unplugging VIFs for instance {{(pid=62824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3013}} [ 1886.310066] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5cdb74c7-a27b-45c4-a89a-19c7714f04b5 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694-project-member] Acquiring lock "refresh_cache-2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1886.310066] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5cdb74c7-a27b-45c4-a89a-19c7714f04b5 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694-project-member] Acquired lock "refresh_cache-2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1886.310066] env[62824]: DEBUG nova.network.neutron [None req-5cdb74c7-a27b-45c4-a89a-19c7714f04b5 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694-project-member] [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1886.661722] env[62824]: DEBUG oslo_vmware.api [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Task: {'id': task-2145676, 'name': PowerOnVM_Task} progress is 82%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1886.808976] env[62824]: DEBUG oslo_concurrency.lockutils [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Releasing lock "refresh_cache-9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1886.809333] env[62824]: DEBUG nova.compute.manager [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c] Instance network_info: |[{"id": "084f765d-6af5-4867-bb76-af443b0bf834", "address": "fa:16:3e:ac:49:e4", "network": {"id": "6f888a0c-8089-4dd4-aace-14b020e11bee", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-137042394-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d041a8209e3848eba8e47251e013ca17", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5f60c972-a72d-4c5f-a250-faadfd6eafbe", "external-id": "nsx-vlan-transportzone-932", "segmentation_id": 932, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap084f765d-6a", "ovs_interfaceid": "084f765d-6af5-4867-bb76-af443b0bf834", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1886.809780] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ac:49:e4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5f60c972-a72d-4c5f-a250-faadfd6eafbe', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '084f765d-6af5-4867-bb76-af443b0bf834', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1886.820165] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Creating folder: Project (d041a8209e3848eba8e47251e013ca17). Parent ref: group-v438503. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1886.825606] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c7a838a0-cbab-4799-825a-75a220dc5243 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1886.839051] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Created folder: Project (d041a8209e3848eba8e47251e013ca17) in parent group-v438503. [ 1886.839309] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Creating folder: Instances. Parent ref: group-v438523. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1886.839794] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bdbe1da6-86a9-407c-a29d-60920c95af43 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1886.850915] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Created folder: Instances in parent group-v438523. [ 1886.851217] env[62824]: DEBUG oslo.service.loopingcall [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1886.851440] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1886.851687] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-183039f9-a9d3-4155-8132-a53768ce9259 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1886.870217] env[62824]: DEBUG nova.network.neutron [None req-5cdb74c7-a27b-45c4-a89a-19c7714f04b5 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694-project-member] [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1886.873547] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1886.873547] env[62824]: value = "task-2145679" [ 1886.873547] env[62824]: _type = "Task" [ 1886.873547] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1886.881916] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145679, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1887.061301] env[62824]: DEBUG nova.compute.manager [req-8b3c568f-a414-4ed9-ab4c-03ab4367c724 req-1a6c84bf-c4b6-4c50-97aa-07d8c2036f47 service nova] [instance: 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c] Received event network-changed-084f765d-6af5-4867-bb76-af443b0bf834 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1887.061825] env[62824]: DEBUG nova.compute.manager [req-8b3c568f-a414-4ed9-ab4c-03ab4367c724 req-1a6c84bf-c4b6-4c50-97aa-07d8c2036f47 service nova] [instance: 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c] Refreshing instance network info cache due to event network-changed-084f765d-6af5-4867-bb76-af443b0bf834. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1887.061825] env[62824]: DEBUG oslo_concurrency.lockutils [req-8b3c568f-a414-4ed9-ab4c-03ab4367c724 req-1a6c84bf-c4b6-4c50-97aa-07d8c2036f47 service nova] Acquiring lock "refresh_cache-9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1887.061931] env[62824]: DEBUG oslo_concurrency.lockutils [req-8b3c568f-a414-4ed9-ab4c-03ab4367c724 req-1a6c84bf-c4b6-4c50-97aa-07d8c2036f47 service nova] Acquired lock "refresh_cache-9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1887.062682] env[62824]: DEBUG nova.network.neutron [req-8b3c568f-a414-4ed9-ab4c-03ab4367c724 req-1a6c84bf-c4b6-4c50-97aa-07d8c2036f47 service nova] [instance: 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c] Refreshing network info cache for port 084f765d-6af5-4867-bb76-af443b0bf834 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1887.067453] env[62824]: DEBUG nova.network.neutron [None req-5cdb74c7-a27b-45c4-a89a-19c7714f04b5 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694-project-member] [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1887.161930] env[62824]: DEBUG oslo_vmware.api [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Task: {'id': task-2145676, 'name': PowerOnVM_Task, 'duration_secs': 1.063546} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1887.162226] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] [instance: 82382932-7302-4441-a6f8-9aa2300ec0f6] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1887.162427] env[62824]: INFO nova.compute.manager [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] [instance: 82382932-7302-4441-a6f8-9aa2300ec0f6] Took 10.50 seconds to spawn the instance on the hypervisor. [ 1887.162606] env[62824]: DEBUG nova.compute.manager [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] [instance: 82382932-7302-4441-a6f8-9aa2300ec0f6] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1887.163548] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b654392-7059-43a1-ab76-cce2b20e233e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1887.393874] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145679, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1887.569653] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5cdb74c7-a27b-45c4-a89a-19c7714f04b5 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694-project-member] Releasing lock "refresh_cache-2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1887.569905] env[62824]: DEBUG nova.compute.manager [None req-5cdb74c7-a27b-45c4-a89a-19c7714f04b5 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3036}} [ 1887.570105] env[62824]: DEBUG nova.compute.manager [None req-5cdb74c7-a27b-45c4-a89a-19c7714f04b5 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694-project-member] [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1887.570368] env[62824]: DEBUG nova.network.neutron [None req-5cdb74c7-a27b-45c4-a89a-19c7714f04b5 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694-project-member] [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1887.619861] env[62824]: DEBUG nova.network.neutron [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] [instance: 517bf2cf-e142-4f36-bf53-79a8d2d8a530] Successfully updated port: a245bb15-a02d-4ed0-a558-e3be831df931 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1887.634055] env[62824]: DEBUG nova.network.neutron [None req-5cdb74c7-a27b-45c4-a89a-19c7714f04b5 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694-project-member] [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1887.687998] env[62824]: INFO nova.compute.manager [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] [instance: 82382932-7302-4441-a6f8-9aa2300ec0f6] Took 36.87 seconds to build instance. [ 1887.878406] env[62824]: DEBUG nova.compute.manager [req-62c3c569-ed8d-461f-93f8-cb7b57402554 req-7e933287-4f89-4385-974d-f373acf72a4c service nova] [instance: 517bf2cf-e142-4f36-bf53-79a8d2d8a530] Received event network-vif-plugged-a245bb15-a02d-4ed0-a558-e3be831df931 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1887.878406] env[62824]: DEBUG oslo_concurrency.lockutils [req-62c3c569-ed8d-461f-93f8-cb7b57402554 req-7e933287-4f89-4385-974d-f373acf72a4c service nova] Acquiring lock "517bf2cf-e142-4f36-bf53-79a8d2d8a530-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1887.878406] env[62824]: DEBUG oslo_concurrency.lockutils [req-62c3c569-ed8d-461f-93f8-cb7b57402554 req-7e933287-4f89-4385-974d-f373acf72a4c service nova] Lock "517bf2cf-e142-4f36-bf53-79a8d2d8a530-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1887.878406] env[62824]: DEBUG oslo_concurrency.lockutils [req-62c3c569-ed8d-461f-93f8-cb7b57402554 req-7e933287-4f89-4385-974d-f373acf72a4c service nova] Lock "517bf2cf-e142-4f36-bf53-79a8d2d8a530-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1887.878406] env[62824]: DEBUG nova.compute.manager [req-62c3c569-ed8d-461f-93f8-cb7b57402554 req-7e933287-4f89-4385-974d-f373acf72a4c service nova] [instance: 517bf2cf-e142-4f36-bf53-79a8d2d8a530] No waiting events found dispatching network-vif-plugged-a245bb15-a02d-4ed0-a558-e3be831df931 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1887.878657] env[62824]: WARNING nova.compute.manager [req-62c3c569-ed8d-461f-93f8-cb7b57402554 req-7e933287-4f89-4385-974d-f373acf72a4c service nova] [instance: 517bf2cf-e142-4f36-bf53-79a8d2d8a530] Received unexpected event network-vif-plugged-a245bb15-a02d-4ed0-a558-e3be831df931 for instance with vm_state building and task_state spawning. [ 1887.884162] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-399c649e-b5c7-43c9-8c11-c064b51fbbb8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1887.894116] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145679, 'name': CreateVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1887.897571] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31d99734-879a-4367-883b-c935f51d4c0f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1887.945564] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba0fbe13-5ec1-4fc1-bc91-3911cd034d7a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1887.959028] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3403e579-9aad-4d3d-91d3-a0e5434777d6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1887.984094] env[62824]: DEBUG nova.compute.provider_tree [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1888.017650] env[62824]: DEBUG nova.network.neutron [req-8b3c568f-a414-4ed9-ab4c-03ab4367c724 req-1a6c84bf-c4b6-4c50-97aa-07d8c2036f47 service nova] [instance: 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c] Updated VIF entry in instance network info cache for port 084f765d-6af5-4867-bb76-af443b0bf834. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1888.018015] env[62824]: DEBUG nova.network.neutron [req-8b3c568f-a414-4ed9-ab4c-03ab4367c724 req-1a6c84bf-c4b6-4c50-97aa-07d8c2036f47 service nova] [instance: 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c] Updating instance_info_cache with network_info: [{"id": "084f765d-6af5-4867-bb76-af443b0bf834", "address": "fa:16:3e:ac:49:e4", "network": {"id": "6f888a0c-8089-4dd4-aace-14b020e11bee", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-137042394-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d041a8209e3848eba8e47251e013ca17", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5f60c972-a72d-4c5f-a250-faadfd6eafbe", "external-id": "nsx-vlan-transportzone-932", "segmentation_id": 932, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap084f765d-6a", "ovs_interfaceid": "084f765d-6af5-4867-bb76-af443b0bf834", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1888.125945] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Acquiring lock "refresh_cache-517bf2cf-e142-4f36-bf53-79a8d2d8a530" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1888.125945] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Acquired lock "refresh_cache-517bf2cf-e142-4f36-bf53-79a8d2d8a530" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1888.125945] env[62824]: DEBUG nova.network.neutron [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] [instance: 517bf2cf-e142-4f36-bf53-79a8d2d8a530] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1888.139152] env[62824]: DEBUG nova.network.neutron [None req-5cdb74c7-a27b-45c4-a89a-19c7714f04b5 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694-project-member] [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1888.189243] env[62824]: DEBUG oslo_concurrency.lockutils [None req-18e40baf-2a54-4705-8faa-a7f47356504b tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Lock "82382932-7302-4441-a6f8-9aa2300ec0f6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 54.112s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1888.361145] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Acquiring lock "bbf1440b-8681-48c8-a178-9a83b925c695" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1888.361374] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Lock "bbf1440b-8681-48c8-a178-9a83b925c695" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1888.386228] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145679, 'name': CreateVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1888.488924] env[62824]: DEBUG nova.scheduler.client.report [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1888.520498] env[62824]: DEBUG oslo_concurrency.lockutils [req-8b3c568f-a414-4ed9-ab4c-03ab4367c724 req-1a6c84bf-c4b6-4c50-97aa-07d8c2036f47 service nova] Releasing lock "refresh_cache-9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1888.520844] env[62824]: DEBUG nova.compute.manager [req-8b3c568f-a414-4ed9-ab4c-03ab4367c724 req-1a6c84bf-c4b6-4c50-97aa-07d8c2036f47 service nova] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Received event network-changed-5a010977-7d05-47e0-8688-bbf9a0e70900 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1888.521051] env[62824]: DEBUG nova.compute.manager [req-8b3c568f-a414-4ed9-ab4c-03ab4367c724 req-1a6c84bf-c4b6-4c50-97aa-07d8c2036f47 service nova] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Refreshing instance network info cache due to event network-changed-5a010977-7d05-47e0-8688-bbf9a0e70900. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1888.521260] env[62824]: DEBUG oslo_concurrency.lockutils [req-8b3c568f-a414-4ed9-ab4c-03ab4367c724 req-1a6c84bf-c4b6-4c50-97aa-07d8c2036f47 service nova] Acquiring lock "refresh_cache-f4d63a93-23af-470c-b36a-662af81dc386" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1888.521406] env[62824]: DEBUG oslo_concurrency.lockutils [req-8b3c568f-a414-4ed9-ab4c-03ab4367c724 req-1a6c84bf-c4b6-4c50-97aa-07d8c2036f47 service nova] Acquired lock "refresh_cache-f4d63a93-23af-470c-b36a-662af81dc386" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1888.521564] env[62824]: DEBUG nova.network.neutron [req-8b3c568f-a414-4ed9-ab4c-03ab4367c724 req-1a6c84bf-c4b6-4c50-97aa-07d8c2036f47 service nova] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Refreshing network info cache for port 5a010977-7d05-47e0-8688-bbf9a0e70900 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1888.641846] env[62824]: INFO nova.compute.manager [None req-5cdb74c7-a27b-45c4-a89a-19c7714f04b5 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694-project-member] [instance: 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c] Took 1.07 seconds to deallocate network for instance. [ 1888.672649] env[62824]: DEBUG nova.network.neutron [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] [instance: 517bf2cf-e142-4f36-bf53-79a8d2d8a530] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1888.692680] env[62824]: DEBUG nova.compute.manager [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: f3740d9a-b21a-4a79-9e28-2a89ecd08bb6] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1888.800479] env[62824]: DEBUG oslo_concurrency.lockutils [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Acquiring lock "7febb8b4-f984-4d79-a888-b2829f2a9df6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1888.800722] env[62824]: DEBUG oslo_concurrency.lockutils [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Lock "7febb8b4-f984-4d79-a888-b2829f2a9df6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1888.889069] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145679, 'name': CreateVM_Task, 'duration_secs': 1.749901} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1888.889498] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1888.889944] env[62824]: DEBUG oslo_concurrency.lockutils [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1888.890128] env[62824]: DEBUG oslo_concurrency.lockutils [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1888.890443] env[62824]: DEBUG oslo_concurrency.lockutils [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1888.890661] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4725fa68-7466-4bd5-bb22-a1ec04188ae6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1888.897672] env[62824]: DEBUG oslo_vmware.api [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Waiting for the task: (returnval){ [ 1888.897672] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52b4184a-c0a1-03fa-b4ff-0ead0e00fe6a" [ 1888.897672] env[62824]: _type = "Task" [ 1888.897672] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1888.906824] env[62824]: DEBUG oslo_vmware.api [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52b4184a-c0a1-03fa-b4ff-0ead0e00fe6a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1888.943270] env[62824]: DEBUG nova.network.neutron [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] [instance: 517bf2cf-e142-4f36-bf53-79a8d2d8a530] Updating instance_info_cache with network_info: [{"id": "a245bb15-a02d-4ed0-a558-e3be831df931", "address": "fa:16:3e:4b:9d:17", "network": {"id": "b6403dc3-30ef-4c76-b521-380c179f345d", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.48", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "22c5c40c2e004babaa15896c89307303", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1002b79b-224e-41e3-a484-4245a767147a", "external-id": "nsx-vlan-transportzone-353", "segmentation_id": 353, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa245bb15-a0", "ovs_interfaceid": "a245bb15-a02d-4ed0-a558-e3be831df931", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1888.993854] env[62824]: DEBUG oslo_concurrency.lockutils [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.691s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1888.993854] env[62824]: DEBUG nova.compute.manager [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] [instance: 9062e606-511f-4fe7-9621-90b3c0e51eec] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1888.996685] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ba162a0b-a017-488a-838b-b0c0ba13e0d5 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.917s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1888.996685] env[62824]: DEBUG nova.objects.instance [None req-ba162a0b-a017-488a-838b-b0c0ba13e0d5 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Lazy-loading 'resources' on Instance uuid f4157385-43ea-4e8c-887c-0985af37abd1 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1889.229353] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1889.289920] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Acquiring lock "387b7f14-f3c1-43ed-8c4e-6de3ce9822e2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1889.291772] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Lock "387b7f14-f3c1-43ed-8c4e-6de3ce9822e2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.003s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1889.410442] env[62824]: DEBUG oslo_vmware.api [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52b4184a-c0a1-03fa-b4ff-0ead0e00fe6a, 'name': SearchDatastore_Task, 'duration_secs': 0.012208} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1889.410889] env[62824]: DEBUG oslo_concurrency.lockutils [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1889.411143] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1889.411378] env[62824]: DEBUG oslo_concurrency.lockutils [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1889.411521] env[62824]: DEBUG oslo_concurrency.lockutils [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1889.411699] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1889.412197] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6412de45-d7a3-4269-ba21-ef4dca39ff4d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1889.422213] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1889.422507] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1889.423577] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d5cf5550-8da8-4acd-8e87-e8e9efe4306c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1889.432042] env[62824]: DEBUG oslo_vmware.api [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Waiting for the task: (returnval){ [ 1889.432042] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52eae6f2-28aa-9bf9-8799-dbe38408138f" [ 1889.432042] env[62824]: _type = "Task" [ 1889.432042] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1889.442899] env[62824]: DEBUG oslo_vmware.api [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52eae6f2-28aa-9bf9-8799-dbe38408138f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1889.446675] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Releasing lock "refresh_cache-517bf2cf-e142-4f36-bf53-79a8d2d8a530" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1889.447311] env[62824]: DEBUG nova.compute.manager [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] [instance: 517bf2cf-e142-4f36-bf53-79a8d2d8a530] Instance network_info: |[{"id": "a245bb15-a02d-4ed0-a558-e3be831df931", "address": "fa:16:3e:4b:9d:17", "network": {"id": "b6403dc3-30ef-4c76-b521-380c179f345d", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.48", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "22c5c40c2e004babaa15896c89307303", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1002b79b-224e-41e3-a484-4245a767147a", "external-id": "nsx-vlan-transportzone-353", "segmentation_id": 353, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa245bb15-a0", "ovs_interfaceid": "a245bb15-a02d-4ed0-a558-e3be831df931", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1889.447507] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] [instance: 517bf2cf-e142-4f36-bf53-79a8d2d8a530] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4b:9d:17', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1002b79b-224e-41e3-a484-4245a767147a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a245bb15-a02d-4ed0-a558-e3be831df931', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1889.456368] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Creating folder: Project (dbe0dd011b0c43328d69799d09fc050c). Parent ref: group-v438503. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1889.456562] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8f4d2bcf-9c7e-4696-a0ce-70bfa6231984 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1889.469924] env[62824]: DEBUG nova.network.neutron [req-8b3c568f-a414-4ed9-ab4c-03ab4367c724 req-1a6c84bf-c4b6-4c50-97aa-07d8c2036f47 service nova] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Updated VIF entry in instance network info cache for port 5a010977-7d05-47e0-8688-bbf9a0e70900. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1889.470292] env[62824]: DEBUG nova.network.neutron [req-8b3c568f-a414-4ed9-ab4c-03ab4367c724 req-1a6c84bf-c4b6-4c50-97aa-07d8c2036f47 service nova] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Updating instance_info_cache with network_info: [{"id": "5a010977-7d05-47e0-8688-bbf9a0e70900", "address": "fa:16:3e:1e:14:06", "network": {"id": "aa1c09e5-8d72-43ad-a903-a0f5e711da80", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1643616572-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.251", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c987c5f844be42c99cc9d57a9ca84c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9bc2632-36f9-4912-8782-8bbb789f909d", "external-id": "nsx-vlan-transportzone-897", "segmentation_id": 897, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5a010977-7d", "ovs_interfaceid": "5a010977-7d05-47e0-8688-bbf9a0e70900", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1889.473035] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Created folder: Project (dbe0dd011b0c43328d69799d09fc050c) in parent group-v438503. [ 1889.473218] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Creating folder: Instances. Parent ref: group-v438526. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1889.473549] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-db8abdba-cc8f-497f-830d-b0ecee167cbd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1889.487852] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Created folder: Instances in parent group-v438526. [ 1889.488130] env[62824]: DEBUG oslo.service.loopingcall [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1889.488933] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 517bf2cf-e142-4f36-bf53-79a8d2d8a530] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1889.489248] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7acd0d5b-2082-4b4a-b581-83e7ffe85adc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1889.506498] env[62824]: DEBUG nova.compute.utils [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1889.511444] env[62824]: DEBUG nova.compute.manager [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] [instance: 9062e606-511f-4fe7-9621-90b3c0e51eec] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1889.511624] env[62824]: DEBUG nova.network.neutron [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] [instance: 9062e606-511f-4fe7-9621-90b3c0e51eec] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1889.521644] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1889.521644] env[62824]: value = "task-2145682" [ 1889.521644] env[62824]: _type = "Task" [ 1889.521644] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1889.533205] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145682, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1889.577509] env[62824]: DEBUG nova.policy [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8c916768a5f04a00b63f0cb22177e540', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '950ce635c14048ba9f7e8f0e7aafa8af', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 1889.676875] env[62824]: INFO nova.scheduler.client.report [None req-5cdb74c7-a27b-45c4-a89a-19c7714f04b5 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694-project-member] Deleted allocations for instance 2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c [ 1889.773304] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1889.773588] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager.update_available_resource {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1889.950294] env[62824]: DEBUG oslo_vmware.api [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52eae6f2-28aa-9bf9-8799-dbe38408138f, 'name': SearchDatastore_Task, 'duration_secs': 0.013338} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1889.950294] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8f5c07e5-c08c-4558-b07a-1d630851df2f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1889.957951] env[62824]: DEBUG oslo_vmware.api [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Waiting for the task: (returnval){ [ 1889.957951] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52e8c005-2b1d-692e-aafb-651aeeaafb34" [ 1889.957951] env[62824]: _type = "Task" [ 1889.957951] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1889.968904] env[62824]: DEBUG oslo_vmware.api [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52e8c005-2b1d-692e-aafb-651aeeaafb34, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1889.972576] env[62824]: DEBUG oslo_concurrency.lockutils [req-8b3c568f-a414-4ed9-ab4c-03ab4367c724 req-1a6c84bf-c4b6-4c50-97aa-07d8c2036f47 service nova] Releasing lock "refresh_cache-f4d63a93-23af-470c-b36a-662af81dc386" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1890.007888] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Acquiring lock "8bb90a4d-93c9-4f54-b15e-48fe966e6c97" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1890.008147] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Lock "8bb90a4d-93c9-4f54-b15e-48fe966e6c97" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1890.011253] env[62824]: DEBUG nova.compute.manager [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] [instance: 9062e606-511f-4fe7-9621-90b3c0e51eec] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1890.029432] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d238b80-8849-4cfd-a662-3499868575ad {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1890.039559] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145682, 'name': CreateVM_Task, 'duration_secs': 0.508605} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1890.041865] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 517bf2cf-e142-4f36-bf53-79a8d2d8a530] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1890.042867] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1890.043070] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1890.043390] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1890.044402] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbb5a230-230e-464b-ab74-a4f4b59c4e39 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1890.047982] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ee9549ed-a2eb-48c1-abcd-2891ee22b321 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1890.055789] env[62824]: DEBUG oslo_vmware.api [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Waiting for the task: (returnval){ [ 1890.055789] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]529eca1f-7b39-8ea6-3bc1-f5923e395275" [ 1890.055789] env[62824]: _type = "Task" [ 1890.055789] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1890.089279] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dbec8e3-7957-4f2c-a746-a768210c71c3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1890.104287] env[62824]: DEBUG oslo_vmware.api [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]529eca1f-7b39-8ea6-3bc1-f5923e395275, 'name': SearchDatastore_Task, 'duration_secs': 0.010259} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1890.104287] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1890.104287] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] [instance: 517bf2cf-e142-4f36-bf53-79a8d2d8a530] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1890.104287] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1890.105400] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-791d6a8f-2ed9-46a3-9cf4-8c4bed84ffcd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1890.109869] env[62824]: DEBUG nova.network.neutron [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] [instance: 9062e606-511f-4fe7-9621-90b3c0e51eec] Successfully created port: 3c79003f-579c-4415-8515-cee7f296a297 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1890.124926] env[62824]: DEBUG nova.compute.provider_tree [None req-ba162a0b-a017-488a-838b-b0c0ba13e0d5 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1890.189693] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5cdb74c7-a27b-45c4-a89a-19c7714f04b5 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694 tempest-FloatingIPsAssociationNegativeTestJSON-1669327694-project-member] Lock "2b2ccb04-80b7-47ac-9fb5-7bd52a439d7c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 59.745s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1890.277615] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1890.286991] env[62824]: DEBUG nova.compute.manager [req-c1b58e3a-77fe-4335-90a1-65390c11ddae req-a0c87c51-97ac-4229-8ad5-21f7b56bb5f1 service nova] [instance: 517bf2cf-e142-4f36-bf53-79a8d2d8a530] Received event network-changed-a245bb15-a02d-4ed0-a558-e3be831df931 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1890.287239] env[62824]: DEBUG nova.compute.manager [req-c1b58e3a-77fe-4335-90a1-65390c11ddae req-a0c87c51-97ac-4229-8ad5-21f7b56bb5f1 service nova] [instance: 517bf2cf-e142-4f36-bf53-79a8d2d8a530] Refreshing instance network info cache due to event network-changed-a245bb15-a02d-4ed0-a558-e3be831df931. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1890.287475] env[62824]: DEBUG oslo_concurrency.lockutils [req-c1b58e3a-77fe-4335-90a1-65390c11ddae req-a0c87c51-97ac-4229-8ad5-21f7b56bb5f1 service nova] Acquiring lock "refresh_cache-517bf2cf-e142-4f36-bf53-79a8d2d8a530" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1890.287612] env[62824]: DEBUG oslo_concurrency.lockutils [req-c1b58e3a-77fe-4335-90a1-65390c11ddae req-a0c87c51-97ac-4229-8ad5-21f7b56bb5f1 service nova] Acquired lock "refresh_cache-517bf2cf-e142-4f36-bf53-79a8d2d8a530" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1890.287808] env[62824]: DEBUG nova.network.neutron [req-c1b58e3a-77fe-4335-90a1-65390c11ddae req-a0c87c51-97ac-4229-8ad5-21f7b56bb5f1 service nova] [instance: 517bf2cf-e142-4f36-bf53-79a8d2d8a530] Refreshing network info cache for port a245bb15-a02d-4ed0-a558-e3be831df931 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1890.469475] env[62824]: DEBUG oslo_vmware.api [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52e8c005-2b1d-692e-aafb-651aeeaafb34, 'name': SearchDatastore_Task, 'duration_secs': 0.010259} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1890.473027] env[62824]: DEBUG oslo_concurrency.lockutils [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1890.473027] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c/9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1890.473027] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1890.473027] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1890.473297] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-678d2681-086c-4e0c-a7b9-65577b14a12e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1890.473980] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-70b42eb0-3e1a-41d9-a4d7-f65a9165a0a9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1890.481451] env[62824]: DEBUG oslo_vmware.api [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Waiting for the task: (returnval){ [ 1890.481451] env[62824]: value = "task-2145683" [ 1890.481451] env[62824]: _type = "Task" [ 1890.481451] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1890.482879] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1890.483252] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1890.489455] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b116d51a-4d22-4947-bc17-0679f9afb3c9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1890.501582] env[62824]: DEBUG oslo_vmware.api [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Task: {'id': task-2145683, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1890.502159] env[62824]: DEBUG oslo_vmware.api [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Waiting for the task: (returnval){ [ 1890.502159] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52577034-42ef-5a38-4243-3e5d3a03792a" [ 1890.502159] env[62824]: _type = "Task" [ 1890.502159] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1890.512703] env[62824]: DEBUG oslo_vmware.api [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52577034-42ef-5a38-4243-3e5d3a03792a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1890.632108] env[62824]: DEBUG nova.scheduler.client.report [None req-ba162a0b-a017-488a-838b-b0c0ba13e0d5 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1890.696178] env[62824]: DEBUG nova.compute.manager [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] [instance: ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1890.783297] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Acquiring lock "2afecec5-763d-4616-b690-41d3101cfc47" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1890.783363] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Lock "2afecec5-763d-4616-b690-41d3101cfc47" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1890.994644] env[62824]: DEBUG oslo_vmware.api [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Task: {'id': task-2145683, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.449547} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1890.994816] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c/9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1890.995685] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1890.995685] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4b077e28-053a-44f5-915e-2e6e592177e8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1891.004099] env[62824]: DEBUG oslo_vmware.api [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Waiting for the task: (returnval){ [ 1891.004099] env[62824]: value = "task-2145684" [ 1891.004099] env[62824]: _type = "Task" [ 1891.004099] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1891.017109] env[62824]: DEBUG oslo_vmware.api [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52577034-42ef-5a38-4243-3e5d3a03792a, 'name': SearchDatastore_Task, 'duration_secs': 0.021226} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1891.022048] env[62824]: DEBUG nova.compute.manager [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] [instance: 9062e606-511f-4fe7-9621-90b3c0e51eec] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1891.024450] env[62824]: DEBUG oslo_vmware.api [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Task: {'id': task-2145684, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1891.024901] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bd666bce-2b8c-425c-a870-c6e6e01e38a9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1891.031762] env[62824]: DEBUG oslo_vmware.api [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Waiting for the task: (returnval){ [ 1891.031762] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52a6b292-860e-b38f-f29c-e99f40efce39" [ 1891.031762] env[62824]: _type = "Task" [ 1891.031762] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1891.047721] env[62824]: DEBUG oslo_vmware.api [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52a6b292-860e-b38f-f29c-e99f40efce39, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1891.058232] env[62824]: DEBUG nova.virt.hardware [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1891.058232] env[62824]: DEBUG nova.virt.hardware [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1891.058232] env[62824]: DEBUG nova.virt.hardware [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1891.058408] env[62824]: DEBUG nova.virt.hardware [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1891.059023] env[62824]: DEBUG nova.virt.hardware [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1891.059023] env[62824]: DEBUG nova.virt.hardware [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1891.059023] env[62824]: DEBUG nova.virt.hardware [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1891.059023] env[62824]: DEBUG nova.virt.hardware [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1891.059226] env[62824]: DEBUG nova.virt.hardware [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1891.059360] env[62824]: DEBUG nova.virt.hardware [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1891.059535] env[62824]: DEBUG nova.virt.hardware [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1891.060407] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8126d9b3-0745-4bab-a02f-97c827ec98d9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1891.070043] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acc0e40f-4b1b-4888-8526-2e7ee27811e3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1891.135249] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ba162a0b-a017-488a-838b-b0c0ba13e0d5 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.139s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1891.138974] env[62824]: DEBUG oslo_concurrency.lockutils [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.680s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1891.143310] env[62824]: INFO nova.compute.claims [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1891.166320] env[62824]: INFO nova.scheduler.client.report [None req-ba162a0b-a017-488a-838b-b0c0ba13e0d5 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Deleted allocations for instance f4157385-43ea-4e8c-887c-0985af37abd1 [ 1891.173852] env[62824]: DEBUG nova.network.neutron [req-c1b58e3a-77fe-4335-90a1-65390c11ddae req-a0c87c51-97ac-4229-8ad5-21f7b56bb5f1 service nova] [instance: 517bf2cf-e142-4f36-bf53-79a8d2d8a530] Updated VIF entry in instance network info cache for port a245bb15-a02d-4ed0-a558-e3be831df931. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1891.174206] env[62824]: DEBUG nova.network.neutron [req-c1b58e3a-77fe-4335-90a1-65390c11ddae req-a0c87c51-97ac-4229-8ad5-21f7b56bb5f1 service nova] [instance: 517bf2cf-e142-4f36-bf53-79a8d2d8a530] Updating instance_info_cache with network_info: [{"id": "a245bb15-a02d-4ed0-a558-e3be831df931", "address": "fa:16:3e:4b:9d:17", "network": {"id": "b6403dc3-30ef-4c76-b521-380c179f345d", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.48", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "22c5c40c2e004babaa15896c89307303", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1002b79b-224e-41e3-a484-4245a767147a", "external-id": "nsx-vlan-transportzone-353", "segmentation_id": 353, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa245bb15-a0", "ovs_interfaceid": "a245bb15-a02d-4ed0-a558-e3be831df931", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1891.221956] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1891.398325] env[62824]: DEBUG oslo_concurrency.lockutils [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Acquiring lock "41f068d3-2c8d-46f0-8d84-78531d28b0dd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1891.398325] env[62824]: DEBUG oslo_concurrency.lockutils [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Lock "41f068d3-2c8d-46f0-8d84-78531d28b0dd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1891.518291] env[62824]: DEBUG oslo_vmware.api [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Task: {'id': task-2145684, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071505} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1891.518558] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1891.519389] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f1abe86-c6a5-4eae-886c-119777a34cac {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1891.541861] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c] Reconfiguring VM instance instance-00000011 to attach disk [datastore1] 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c/9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1891.545011] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e4d166a3-99b6-4045-b452-619f076bb425 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1891.564751] env[62824]: DEBUG oslo_vmware.api [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52a6b292-860e-b38f-f29c-e99f40efce39, 'name': SearchDatastore_Task, 'duration_secs': 0.013645} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1891.566033] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1891.566301] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 517bf2cf-e142-4f36-bf53-79a8d2d8a530/517bf2cf-e142-4f36-bf53-79a8d2d8a530.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1891.566606] env[62824]: DEBUG oslo_vmware.api [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Waiting for the task: (returnval){ [ 1891.566606] env[62824]: value = "task-2145685" [ 1891.566606] env[62824]: _type = "Task" [ 1891.566606] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1891.567801] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5ec050d5-711d-443c-812a-ff23cf2c30b3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1891.576442] env[62824]: DEBUG oslo_vmware.api [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Task: {'id': task-2145685, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1891.577476] env[62824]: DEBUG oslo_vmware.api [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Waiting for the task: (returnval){ [ 1891.577476] env[62824]: value = "task-2145686" [ 1891.577476] env[62824]: _type = "Task" [ 1891.577476] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1891.584666] env[62824]: DEBUG oslo_vmware.api [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Task: {'id': task-2145686, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1891.674346] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ba162a0b-a017-488a-838b-b0c0ba13e0d5 tempest-ServerDiagnosticsV248Test-1419977481 tempest-ServerDiagnosticsV248Test-1419977481-project-member] Lock "f4157385-43ea-4e8c-887c-0985af37abd1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.855s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1891.676958] env[62824]: DEBUG oslo_concurrency.lockutils [req-c1b58e3a-77fe-4335-90a1-65390c11ddae req-a0c87c51-97ac-4229-8ad5-21f7b56bb5f1 service nova] Releasing lock "refresh_cache-517bf2cf-e142-4f36-bf53-79a8d2d8a530" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1891.883796] env[62824]: DEBUG nova.compute.manager [req-6bf95e54-e633-4015-ade8-4dedda655d21 req-390b4aa1-85f7-4316-9f45-a143f4bb7f5f service nova] [instance: 9062e606-511f-4fe7-9621-90b3c0e51eec] Received event network-vif-plugged-3c79003f-579c-4415-8515-cee7f296a297 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1891.883796] env[62824]: DEBUG oslo_concurrency.lockutils [req-6bf95e54-e633-4015-ade8-4dedda655d21 req-390b4aa1-85f7-4316-9f45-a143f4bb7f5f service nova] Acquiring lock "9062e606-511f-4fe7-9621-90b3c0e51eec-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1891.883796] env[62824]: DEBUG oslo_concurrency.lockutils [req-6bf95e54-e633-4015-ade8-4dedda655d21 req-390b4aa1-85f7-4316-9f45-a143f4bb7f5f service nova] Lock "9062e606-511f-4fe7-9621-90b3c0e51eec-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1891.883796] env[62824]: DEBUG oslo_concurrency.lockutils [req-6bf95e54-e633-4015-ade8-4dedda655d21 req-390b4aa1-85f7-4316-9f45-a143f4bb7f5f service nova] Lock "9062e606-511f-4fe7-9621-90b3c0e51eec-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1891.883952] env[62824]: DEBUG nova.compute.manager [req-6bf95e54-e633-4015-ade8-4dedda655d21 req-390b4aa1-85f7-4316-9f45-a143f4bb7f5f service nova] [instance: 9062e606-511f-4fe7-9621-90b3c0e51eec] No waiting events found dispatching network-vif-plugged-3c79003f-579c-4415-8515-cee7f296a297 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1891.884894] env[62824]: WARNING nova.compute.manager [req-6bf95e54-e633-4015-ade8-4dedda655d21 req-390b4aa1-85f7-4316-9f45-a143f4bb7f5f service nova] [instance: 9062e606-511f-4fe7-9621-90b3c0e51eec] Received unexpected event network-vif-plugged-3c79003f-579c-4415-8515-cee7f296a297 for instance with vm_state building and task_state spawning. [ 1892.021022] env[62824]: DEBUG nova.network.neutron [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] [instance: 9062e606-511f-4fe7-9621-90b3c0e51eec] Successfully updated port: 3c79003f-579c-4415-8515-cee7f296a297 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1892.826717] env[62824]: DEBUG oslo_concurrency.lockutils [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Acquiring lock "refresh_cache-9062e606-511f-4fe7-9621-90b3c0e51eec" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1892.826984] env[62824]: DEBUG oslo_concurrency.lockutils [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Acquired lock "refresh_cache-9062e606-511f-4fe7-9621-90b3c0e51eec" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1892.827125] env[62824]: DEBUG nova.network.neutron [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] [instance: 9062e606-511f-4fe7-9621-90b3c0e51eec] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1892.828758] env[62824]: DEBUG oslo_vmware.api [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Task: {'id': task-2145685, 'name': ReconfigVM_Task, 'duration_secs': 0.311926} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1892.838492] env[62824]: DEBUG nova.compute.manager [req-6af5f652-67f3-4ab8-9741-2355fc46d4bc req-4a8f4691-3571-4dd6-98b3-9be09c3dff97 service nova] [instance: 82382932-7302-4441-a6f8-9aa2300ec0f6] Received event network-changed-760062c3-a8b6-4863-a557-0d44a27e3bb3 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1892.838752] env[62824]: DEBUG nova.compute.manager [req-6af5f652-67f3-4ab8-9741-2355fc46d4bc req-4a8f4691-3571-4dd6-98b3-9be09c3dff97 service nova] [instance: 82382932-7302-4441-a6f8-9aa2300ec0f6] Refreshing instance network info cache due to event network-changed-760062c3-a8b6-4863-a557-0d44a27e3bb3. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1892.839085] env[62824]: DEBUG oslo_concurrency.lockutils [req-6af5f652-67f3-4ab8-9741-2355fc46d4bc req-4a8f4691-3571-4dd6-98b3-9be09c3dff97 service nova] Acquiring lock "refresh_cache-82382932-7302-4441-a6f8-9aa2300ec0f6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1892.839265] env[62824]: DEBUG oslo_concurrency.lockutils [req-6af5f652-67f3-4ab8-9741-2355fc46d4bc req-4a8f4691-3571-4dd6-98b3-9be09c3dff97 service nova] Acquired lock "refresh_cache-82382932-7302-4441-a6f8-9aa2300ec0f6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1892.840078] env[62824]: DEBUG nova.network.neutron [req-6af5f652-67f3-4ab8-9741-2355fc46d4bc req-4a8f4691-3571-4dd6-98b3-9be09c3dff97 service nova] [instance: 82382932-7302-4441-a6f8-9aa2300ec0f6] Refreshing network info cache for port 760062c3-a8b6-4863-a557-0d44a27e3bb3 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1892.841259] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c] Reconfigured VM instance instance-00000011 to attach disk [datastore1] 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c/9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1892.843304] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6f5a183c-ff0c-435b-9bdb-f94bf334d8d3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1892.854627] env[62824]: DEBUG oslo_vmware.api [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Task: {'id': task-2145686, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.51503} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1892.858598] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 517bf2cf-e142-4f36-bf53-79a8d2d8a530/517bf2cf-e142-4f36-bf53-79a8d2d8a530.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1892.858819] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] [instance: 517bf2cf-e142-4f36-bf53-79a8d2d8a530] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1892.859580] env[62824]: DEBUG oslo_vmware.api [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Waiting for the task: (returnval){ [ 1892.859580] env[62824]: value = "task-2145687" [ 1892.859580] env[62824]: _type = "Task" [ 1892.859580] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1892.860797] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4cf60e7d-408c-4ae6-bbde-81447cef5a41 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1892.874194] env[62824]: DEBUG oslo_vmware.api [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Task: {'id': task-2145687, 'name': Rename_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1892.878434] env[62824]: DEBUG oslo_vmware.api [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Waiting for the task: (returnval){ [ 1892.878434] env[62824]: value = "task-2145688" [ 1892.878434] env[62824]: _type = "Task" [ 1892.878434] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1892.887661] env[62824]: DEBUG oslo_vmware.api [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Task: {'id': task-2145688, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1893.284417] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0725d08f-c105-4559-9618-24d428a57e84 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1893.292437] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dbaf5c1-299e-4b0f-8fc7-da58b77d86f9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1893.322664] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b6a01b3-b45c-4b44-8360-d75348cab8c3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1893.330652] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f83112d-8357-4700-b80f-80c663363dd3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1893.350817] env[62824]: DEBUG nova.compute.provider_tree [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1893.373654] env[62824]: DEBUG oslo_vmware.api [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Task: {'id': task-2145687, 'name': Rename_Task, 'duration_secs': 0.244093} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1893.373899] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1893.374141] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ce9809f0-74ad-4595-aea7-04f5698e03ac {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1893.383929] env[62824]: DEBUG oslo_vmware.api [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Waiting for the task: (returnval){ [ 1893.383929] env[62824]: value = "task-2145689" [ 1893.383929] env[62824]: _type = "Task" [ 1893.383929] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1893.390250] env[62824]: DEBUG oslo_vmware.api [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Task: {'id': task-2145688, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.243624} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1893.391361] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] [instance: 517bf2cf-e142-4f36-bf53-79a8d2d8a530] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1893.392025] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32a8c18b-8cd2-460c-b690-224b6b5fd7e5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1893.401023] env[62824]: DEBUG oslo_vmware.api [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Task: {'id': task-2145689, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1893.420295] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] [instance: 517bf2cf-e142-4f36-bf53-79a8d2d8a530] Reconfiguring VM instance instance-00000012 to attach disk [datastore1] 517bf2cf-e142-4f36-bf53-79a8d2d8a530/517bf2cf-e142-4f36-bf53-79a8d2d8a530.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1893.424163] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c41b316c-0a7d-4796-97a1-2700fd727cf1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1893.438803] env[62824]: DEBUG nova.network.neutron [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] [instance: 9062e606-511f-4fe7-9621-90b3c0e51eec] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1893.446289] env[62824]: DEBUG oslo_vmware.api [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Waiting for the task: (returnval){ [ 1893.446289] env[62824]: value = "task-2145690" [ 1893.446289] env[62824]: _type = "Task" [ 1893.446289] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1893.455844] env[62824]: DEBUG oslo_vmware.api [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Task: {'id': task-2145690, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1893.854556] env[62824]: DEBUG nova.scheduler.client.report [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1893.895197] env[62824]: DEBUG oslo_vmware.api [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Task: {'id': task-2145689, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1893.957405] env[62824]: DEBUG oslo_vmware.api [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Task: {'id': task-2145690, 'name': ReconfigVM_Task, 'duration_secs': 0.326931} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1893.957845] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] [instance: 517bf2cf-e142-4f36-bf53-79a8d2d8a530] Reconfigured VM instance instance-00000012 to attach disk [datastore1] 517bf2cf-e142-4f36-bf53-79a8d2d8a530/517bf2cf-e142-4f36-bf53-79a8d2d8a530.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1893.958391] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ac1cbd53-ed95-4a58-ac7e-26ada3e90691 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1893.967595] env[62824]: DEBUG oslo_vmware.api [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Waiting for the task: (returnval){ [ 1893.967595] env[62824]: value = "task-2145691" [ 1893.967595] env[62824]: _type = "Task" [ 1893.967595] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1893.975850] env[62824]: DEBUG oslo_vmware.api [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Task: {'id': task-2145691, 'name': Rename_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1894.125958] env[62824]: DEBUG nova.network.neutron [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] [instance: 9062e606-511f-4fe7-9621-90b3c0e51eec] Updating instance_info_cache with network_info: [{"id": "3c79003f-579c-4415-8515-cee7f296a297", "address": "fa:16:3e:1a:d4:42", "network": {"id": "7aa6d314-a33f-4c6e-a94e-c0cf1601d6f7", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-1542833912-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "950ce635c14048ba9f7e8f0e7aafa8af", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "be8bd197-4b2b-46e7-88ea-2554b0438584", "external-id": "nsx-vlan-transportzone-338", "segmentation_id": 338, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3c79003f-57", "ovs_interfaceid": "3c79003f-579c-4415-8515-cee7f296a297", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1894.275917] env[62824]: DEBUG nova.network.neutron [req-6af5f652-67f3-4ab8-9741-2355fc46d4bc req-4a8f4691-3571-4dd6-98b3-9be09c3dff97 service nova] [instance: 82382932-7302-4441-a6f8-9aa2300ec0f6] Updated VIF entry in instance network info cache for port 760062c3-a8b6-4863-a557-0d44a27e3bb3. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1894.276332] env[62824]: DEBUG nova.network.neutron [req-6af5f652-67f3-4ab8-9741-2355fc46d4bc req-4a8f4691-3571-4dd6-98b3-9be09c3dff97 service nova] [instance: 82382932-7302-4441-a6f8-9aa2300ec0f6] Updating instance_info_cache with network_info: [{"id": "760062c3-a8b6-4863-a557-0d44a27e3bb3", "address": "fa:16:3e:d4:15:ee", "network": {"id": "668ff2d5-a4e1-488b-b71f-7f45c130d0e0", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-616618532-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.169", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0cf3a345a7d54625885c2159edcf0877", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "57691231-2b8d-4d71-8f79-d4a6a1d95ec8", "external-id": "nsx-vlan-transportzone-373", "segmentation_id": 373, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap760062c3-a8", "ovs_interfaceid": "760062c3-a8b6-4863-a557-0d44a27e3bb3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1894.348682] env[62824]: DEBUG nova.compute.manager [req-4943e028-e87e-46c9-a952-f660ae0e4502 req-8ab25bc0-ff4a-4791-b5ab-280dc554f21b service nova] [instance: 9062e606-511f-4fe7-9621-90b3c0e51eec] Received event network-changed-3c79003f-579c-4415-8515-cee7f296a297 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1894.348950] env[62824]: DEBUG nova.compute.manager [req-4943e028-e87e-46c9-a952-f660ae0e4502 req-8ab25bc0-ff4a-4791-b5ab-280dc554f21b service nova] [instance: 9062e606-511f-4fe7-9621-90b3c0e51eec] Refreshing instance network info cache due to event network-changed-3c79003f-579c-4415-8515-cee7f296a297. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1894.349117] env[62824]: DEBUG oslo_concurrency.lockutils [req-4943e028-e87e-46c9-a952-f660ae0e4502 req-8ab25bc0-ff4a-4791-b5ab-280dc554f21b service nova] Acquiring lock "refresh_cache-9062e606-511f-4fe7-9621-90b3c0e51eec" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1894.359694] env[62824]: DEBUG oslo_concurrency.lockutils [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.221s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1894.360230] env[62824]: DEBUG nova.compute.manager [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1894.363051] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.763s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1894.364772] env[62824]: INFO nova.compute.claims [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] [instance: dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1894.397046] env[62824]: DEBUG oslo_vmware.api [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Task: {'id': task-2145689, 'name': PowerOnVM_Task, 'duration_secs': 0.983246} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1894.397300] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1894.397588] env[62824]: INFO nova.compute.manager [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c] Took 10.83 seconds to spawn the instance on the hypervisor. [ 1894.397773] env[62824]: DEBUG nova.compute.manager [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1894.399849] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b9c0877-e827-4a6a-854a-9efddcb7adb9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1894.477912] env[62824]: DEBUG oslo_vmware.api [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Task: {'id': task-2145691, 'name': Rename_Task, 'duration_secs': 0.166778} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1894.478210] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] [instance: 517bf2cf-e142-4f36-bf53-79a8d2d8a530] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1894.478450] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f220b916-95b6-489e-a276-c4f2b465e5c3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1894.484654] env[62824]: DEBUG oslo_vmware.api [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Waiting for the task: (returnval){ [ 1894.484654] env[62824]: value = "task-2145692" [ 1894.484654] env[62824]: _type = "Task" [ 1894.484654] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1894.492676] env[62824]: DEBUG oslo_vmware.api [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Task: {'id': task-2145692, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1894.628227] env[62824]: DEBUG oslo_concurrency.lockutils [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Releasing lock "refresh_cache-9062e606-511f-4fe7-9621-90b3c0e51eec" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1894.628632] env[62824]: DEBUG nova.compute.manager [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] [instance: 9062e606-511f-4fe7-9621-90b3c0e51eec] Instance network_info: |[{"id": "3c79003f-579c-4415-8515-cee7f296a297", "address": "fa:16:3e:1a:d4:42", "network": {"id": "7aa6d314-a33f-4c6e-a94e-c0cf1601d6f7", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-1542833912-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "950ce635c14048ba9f7e8f0e7aafa8af", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "be8bd197-4b2b-46e7-88ea-2554b0438584", "external-id": "nsx-vlan-transportzone-338", "segmentation_id": 338, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3c79003f-57", "ovs_interfaceid": "3c79003f-579c-4415-8515-cee7f296a297", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1894.629192] env[62824]: DEBUG oslo_concurrency.lockutils [req-4943e028-e87e-46c9-a952-f660ae0e4502 req-8ab25bc0-ff4a-4791-b5ab-280dc554f21b service nova] Acquired lock "refresh_cache-9062e606-511f-4fe7-9621-90b3c0e51eec" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1894.629920] env[62824]: DEBUG nova.network.neutron [req-4943e028-e87e-46c9-a952-f660ae0e4502 req-8ab25bc0-ff4a-4791-b5ab-280dc554f21b service nova] [instance: 9062e606-511f-4fe7-9621-90b3c0e51eec] Refreshing network info cache for port 3c79003f-579c-4415-8515-cee7f296a297 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1894.631970] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] [instance: 9062e606-511f-4fe7-9621-90b3c0e51eec] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1a:d4:42', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'be8bd197-4b2b-46e7-88ea-2554b0438584', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3c79003f-579c-4415-8515-cee7f296a297', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1894.639977] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Creating folder: Project (950ce635c14048ba9f7e8f0e7aafa8af). Parent ref: group-v438503. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1894.642986] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cb30f65c-732e-4dea-9f58-bfdb32c492c4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1894.653910] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Created folder: Project (950ce635c14048ba9f7e8f0e7aafa8af) in parent group-v438503. [ 1894.655567] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Creating folder: Instances. Parent ref: group-v438529. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1894.655567] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cc07f59f-5dc3-41ea-a566-effbb0ee2fc7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1894.662942] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Created folder: Instances in parent group-v438529. [ 1894.663205] env[62824]: DEBUG oslo.service.loopingcall [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1894.663401] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9062e606-511f-4fe7-9621-90b3c0e51eec] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1894.663607] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-62570b52-a0ef-4b44-ab4c-bdc2a7b6a693 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1894.686688] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1894.686688] env[62824]: value = "task-2145695" [ 1894.686688] env[62824]: _type = "Task" [ 1894.686688] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1894.697057] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145695, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1894.780081] env[62824]: DEBUG oslo_concurrency.lockutils [req-6af5f652-67f3-4ab8-9741-2355fc46d4bc req-4a8f4691-3571-4dd6-98b3-9be09c3dff97 service nova] Releasing lock "refresh_cache-82382932-7302-4441-a6f8-9aa2300ec0f6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1894.873492] env[62824]: DEBUG nova.compute.utils [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1894.878715] env[62824]: DEBUG nova.compute.manager [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Not allocating networking since 'none' was specified. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1984}} [ 1894.921854] env[62824]: INFO nova.compute.manager [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c] Took 37.43 seconds to build instance. [ 1894.988275] env[62824]: DEBUG nova.network.neutron [req-4943e028-e87e-46c9-a952-f660ae0e4502 req-8ab25bc0-ff4a-4791-b5ab-280dc554f21b service nova] [instance: 9062e606-511f-4fe7-9621-90b3c0e51eec] Updated VIF entry in instance network info cache for port 3c79003f-579c-4415-8515-cee7f296a297. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1894.988275] env[62824]: DEBUG nova.network.neutron [req-4943e028-e87e-46c9-a952-f660ae0e4502 req-8ab25bc0-ff4a-4791-b5ab-280dc554f21b service nova] [instance: 9062e606-511f-4fe7-9621-90b3c0e51eec] Updating instance_info_cache with network_info: [{"id": "3c79003f-579c-4415-8515-cee7f296a297", "address": "fa:16:3e:1a:d4:42", "network": {"id": "7aa6d314-a33f-4c6e-a94e-c0cf1601d6f7", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-1542833912-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "950ce635c14048ba9f7e8f0e7aafa8af", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "be8bd197-4b2b-46e7-88ea-2554b0438584", "external-id": "nsx-vlan-transportzone-338", "segmentation_id": 338, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3c79003f-57", "ovs_interfaceid": "3c79003f-579c-4415-8515-cee7f296a297", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1895.001349] env[62824]: DEBUG oslo_vmware.api [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Task: {'id': task-2145692, 'name': PowerOnVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1895.197378] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145695, 'name': CreateVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1895.379455] env[62824]: DEBUG nova.compute.manager [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1895.424110] env[62824]: DEBUG oslo_concurrency.lockutils [None req-555cc149-7fd3-4b61-ac55-b0257285e50a tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Lock "9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 56.012s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1895.502128] env[62824]: DEBUG oslo_concurrency.lockutils [req-4943e028-e87e-46c9-a952-f660ae0e4502 req-8ab25bc0-ff4a-4791-b5ab-280dc554f21b service nova] Releasing lock "refresh_cache-9062e606-511f-4fe7-9621-90b3c0e51eec" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1895.510787] env[62824]: DEBUG oslo_vmware.api [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Task: {'id': task-2145692, 'name': PowerOnVM_Task, 'duration_secs': 0.694869} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1895.511067] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] [instance: 517bf2cf-e142-4f36-bf53-79a8d2d8a530] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1895.511358] env[62824]: INFO nova.compute.manager [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] [instance: 517bf2cf-e142-4f36-bf53-79a8d2d8a530] Took 9.31 seconds to spawn the instance on the hypervisor. [ 1895.511538] env[62824]: DEBUG nova.compute.manager [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] [instance: 517bf2cf-e142-4f36-bf53-79a8d2d8a530] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1895.512367] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-428c2b9f-ac41-40a9-a157-5898f0bc8456 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1895.699038] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145695, 'name': CreateVM_Task, 'duration_secs': 0.838491} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1895.701779] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9062e606-511f-4fe7-9621-90b3c0e51eec] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1895.702820] env[62824]: DEBUG oslo_concurrency.lockutils [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1895.703038] env[62824]: DEBUG oslo_concurrency.lockutils [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1895.703400] env[62824]: DEBUG oslo_concurrency.lockutils [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1895.704175] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2d01df7e-f09a-468b-b6fa-0fe3f57824ec {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1895.708957] env[62824]: DEBUG oslo_vmware.api [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Waiting for the task: (returnval){ [ 1895.708957] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5214a449-92d6-4e68-369a-ca282d4e64a9" [ 1895.708957] env[62824]: _type = "Task" [ 1895.708957] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1895.722116] env[62824]: DEBUG oslo_vmware.api [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5214a449-92d6-4e68-369a-ca282d4e64a9, 'name': SearchDatastore_Task, 'duration_secs': 0.009235} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1895.722116] env[62824]: DEBUG oslo_concurrency.lockutils [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1895.722437] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] [instance: 9062e606-511f-4fe7-9621-90b3c0e51eec] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1895.722611] env[62824]: DEBUG oslo_concurrency.lockutils [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1895.722754] env[62824]: DEBUG oslo_concurrency.lockutils [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1895.723090] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1895.723493] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0fde6809-d010-4c24-8f94-ad67fffe4a51 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1895.734478] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1895.734948] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1895.735970] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-49876970-9ca5-44ee-b096-4db19468288a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1895.744817] env[62824]: DEBUG oslo_vmware.api [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Waiting for the task: (returnval){ [ 1895.744817] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52efe3b5-9d0a-41b3-6c60-4411fd1cb709" [ 1895.744817] env[62824]: _type = "Task" [ 1895.744817] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1895.753777] env[62824]: DEBUG oslo_vmware.api [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52efe3b5-9d0a-41b3-6c60-4411fd1cb709, 'name': SearchDatastore_Task, 'duration_secs': 0.008064} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1895.759341] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9e5b07ca-bd9b-40ed-869f-b954033b8dce {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1895.766360] env[62824]: DEBUG oslo_vmware.api [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Waiting for the task: (returnval){ [ 1895.766360] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]528ec198-21d4-7218-b8e6-c242ea900735" [ 1895.766360] env[62824]: _type = "Task" [ 1895.766360] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1895.777812] env[62824]: DEBUG oslo_vmware.api [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]528ec198-21d4-7218-b8e6-c242ea900735, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1895.918638] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2219d5f-87a3-4d00-8eba-b857aa9e7ad7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1895.926738] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9de960d8-8ae6-4631-b20a-0b8b8b26a59c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1895.930492] env[62824]: DEBUG nova.compute.manager [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1895.960210] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1193377c-bbfb-41c9-b37b-2d0601b4f666 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1895.970941] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-195dc1d1-bea1-463f-8abb-b62e3aeec183 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1895.990058] env[62824]: DEBUG nova.compute.provider_tree [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1895.993992] env[62824]: DEBUG nova.compute.manager [req-9ac13862-252c-4a8a-b4dc-3086a0942ed2 req-1fd947b9-c362-4ad6-a977-f42395b0090b service nova] [instance: 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c] Received event network-changed-084f765d-6af5-4867-bb76-af443b0bf834 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1895.993992] env[62824]: DEBUG nova.compute.manager [req-9ac13862-252c-4a8a-b4dc-3086a0942ed2 req-1fd947b9-c362-4ad6-a977-f42395b0090b service nova] [instance: 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c] Refreshing instance network info cache due to event network-changed-084f765d-6af5-4867-bb76-af443b0bf834. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1895.993992] env[62824]: DEBUG oslo_concurrency.lockutils [req-9ac13862-252c-4a8a-b4dc-3086a0942ed2 req-1fd947b9-c362-4ad6-a977-f42395b0090b service nova] Acquiring lock "refresh_cache-9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1895.993992] env[62824]: DEBUG oslo_concurrency.lockutils [req-9ac13862-252c-4a8a-b4dc-3086a0942ed2 req-1fd947b9-c362-4ad6-a977-f42395b0090b service nova] Acquired lock "refresh_cache-9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1895.993992] env[62824]: DEBUG nova.network.neutron [req-9ac13862-252c-4a8a-b4dc-3086a0942ed2 req-1fd947b9-c362-4ad6-a977-f42395b0090b service nova] [instance: 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c] Refreshing network info cache for port 084f765d-6af5-4867-bb76-af443b0bf834 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1896.031631] env[62824]: INFO nova.compute.manager [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] [instance: 517bf2cf-e142-4f36-bf53-79a8d2d8a530] Took 36.48 seconds to build instance. [ 1896.276988] env[62824]: DEBUG oslo_vmware.api [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]528ec198-21d4-7218-b8e6-c242ea900735, 'name': SearchDatastore_Task, 'duration_secs': 0.008874} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1896.277455] env[62824]: DEBUG oslo_concurrency.lockutils [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1896.277911] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 9062e606-511f-4fe7-9621-90b3c0e51eec/9062e606-511f-4fe7-9621-90b3c0e51eec.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1896.278082] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-024a4f29-072d-4187-8386-671a99a11d3d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1896.284820] env[62824]: DEBUG oslo_vmware.api [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Waiting for the task: (returnval){ [ 1896.284820] env[62824]: value = "task-2145696" [ 1896.284820] env[62824]: _type = "Task" [ 1896.284820] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1896.293590] env[62824]: DEBUG oslo_vmware.api [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Task: {'id': task-2145696, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1896.397604] env[62824]: DEBUG nova.compute.manager [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1896.436548] env[62824]: DEBUG nova.virt.hardware [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1896.437130] env[62824]: DEBUG nova.virt.hardware [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1896.437398] env[62824]: DEBUG nova.virt.hardware [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1896.438526] env[62824]: DEBUG nova.virt.hardware [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1896.438829] env[62824]: DEBUG nova.virt.hardware [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1896.442018] env[62824]: DEBUG nova.virt.hardware [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1896.442018] env[62824]: DEBUG nova.virt.hardware [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1896.442018] env[62824]: DEBUG nova.virt.hardware [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1896.442018] env[62824]: DEBUG nova.virt.hardware [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1896.442018] env[62824]: DEBUG nova.virt.hardware [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1896.442562] env[62824]: DEBUG nova.virt.hardware [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1896.442562] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37915f2c-9d21-4db9-81e5-0e60b623e652 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1896.457414] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-325dc156-dd36-41c1-9de7-c00548c38db5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1896.480192] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Instance VIF info [] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1896.489417] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Creating folder: Project (065d8f0b7ffe496a87e7f95e83b1a0ee). Parent ref: group-v438503. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1896.490801] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1896.491137] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-efc49db6-042a-41cf-aef6-68377616c56d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1896.495235] env[62824]: DEBUG nova.scheduler.client.report [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1896.504374] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Created folder: Project (065d8f0b7ffe496a87e7f95e83b1a0ee) in parent group-v438503. [ 1896.504589] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Creating folder: Instances. Parent ref: group-v438532. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1896.505208] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e72ddb67-55a7-49a2-86bd-3211188d72cf {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1896.516999] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Created folder: Instances in parent group-v438532. [ 1896.517311] env[62824]: DEBUG oslo.service.loopingcall [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1896.517510] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1896.519394] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3e7f1a08-1fd0-41ce-b2ef-924816d3ea80 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1896.535724] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7ad7f98e-1031-4f14-ab5e-4710316ab3a5 tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Lock "517bf2cf-e142-4f36-bf53-79a8d2d8a530" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 54.384s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1896.542482] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1896.542482] env[62824]: value = "task-2145699" [ 1896.542482] env[62824]: _type = "Task" [ 1896.542482] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1896.552539] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145699, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1896.797738] env[62824]: DEBUG oslo_vmware.api [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Task: {'id': task-2145696, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.482898} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1896.798244] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 9062e606-511f-4fe7-9621-90b3c0e51eec/9062e606-511f-4fe7-9621-90b3c0e51eec.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1896.798688] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] [instance: 9062e606-511f-4fe7-9621-90b3c0e51eec] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1896.799103] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-75c87727-f19f-401f-bb45-40331ba67cda {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1896.807148] env[62824]: DEBUG oslo_vmware.api [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Waiting for the task: (returnval){ [ 1896.807148] env[62824]: value = "task-2145700" [ 1896.807148] env[62824]: _type = "Task" [ 1896.807148] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1896.824521] env[62824]: DEBUG oslo_vmware.api [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Task: {'id': task-2145700, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1897.004365] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.641s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1897.005118] env[62824]: DEBUG nova.compute.manager [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] [instance: dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1897.008349] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.928s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1897.009822] env[62824]: INFO nova.compute.claims [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] [instance: 4dfe9f82-7f89-486b-b159-948656d2896c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1897.040133] env[62824]: DEBUG nova.compute.manager [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] [instance: b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1897.061790] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145699, 'name': CreateVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1897.150364] env[62824]: DEBUG nova.network.neutron [req-9ac13862-252c-4a8a-b4dc-3086a0942ed2 req-1fd947b9-c362-4ad6-a977-f42395b0090b service nova] [instance: 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c] Updated VIF entry in instance network info cache for port 084f765d-6af5-4867-bb76-af443b0bf834. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1897.151031] env[62824]: DEBUG nova.network.neutron [req-9ac13862-252c-4a8a-b4dc-3086a0942ed2 req-1fd947b9-c362-4ad6-a977-f42395b0090b service nova] [instance: 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c] Updating instance_info_cache with network_info: [{"id": "084f765d-6af5-4867-bb76-af443b0bf834", "address": "fa:16:3e:ac:49:e4", "network": {"id": "6f888a0c-8089-4dd4-aace-14b020e11bee", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-137042394-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.242", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d041a8209e3848eba8e47251e013ca17", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5f60c972-a72d-4c5f-a250-faadfd6eafbe", "external-id": "nsx-vlan-transportzone-932", "segmentation_id": 932, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap084f765d-6a", "ovs_interfaceid": "084f765d-6af5-4867-bb76-af443b0bf834", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1897.317544] env[62824]: DEBUG oslo_vmware.api [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Task: {'id': task-2145700, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073413} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1897.317840] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] [instance: 9062e606-511f-4fe7-9621-90b3c0e51eec] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1897.319841] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b4dcc44-30d0-43c0-8794-61780503b5be {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1897.347780] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] [instance: 9062e606-511f-4fe7-9621-90b3c0e51eec] Reconfiguring VM instance instance-00000013 to attach disk [datastore1] 9062e606-511f-4fe7-9621-90b3c0e51eec/9062e606-511f-4fe7-9621-90b3c0e51eec.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1897.348162] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8e6e3393-e40f-46da-ad6c-c0ca78c24bfe {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1897.371942] env[62824]: DEBUG oslo_vmware.api [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Waiting for the task: (returnval){ [ 1897.371942] env[62824]: value = "task-2145701" [ 1897.371942] env[62824]: _type = "Task" [ 1897.371942] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1897.385976] env[62824]: DEBUG oslo_vmware.api [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Task: {'id': task-2145701, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1897.521448] env[62824]: DEBUG nova.compute.utils [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1897.523875] env[62824]: DEBUG nova.compute.manager [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] [instance: dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1897.524148] env[62824]: DEBUG nova.network.neutron [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] [instance: dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1897.567821] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145699, 'name': CreateVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1897.579172] env[62824]: DEBUG oslo_concurrency.lockutils [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1897.609238] env[62824]: DEBUG nova.policy [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd5a4562cfaf34333ac1f60af0eecd644', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '51135bed305c4e89afaeed2b187cc7f6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 1897.655874] env[62824]: DEBUG oslo_concurrency.lockutils [req-9ac13862-252c-4a8a-b4dc-3086a0942ed2 req-1fd947b9-c362-4ad6-a977-f42395b0090b service nova] Releasing lock "refresh_cache-9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1897.887049] env[62824]: DEBUG oslo_vmware.api [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Task: {'id': task-2145701, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1898.024439] env[62824]: DEBUG nova.compute.manager [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] [instance: dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1898.066430] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145699, 'name': CreateVM_Task, 'duration_secs': 1.399057} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1898.068949] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1898.068949] env[62824]: DEBUG oslo_concurrency.lockutils [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1898.068949] env[62824]: DEBUG oslo_concurrency.lockutils [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1898.069223] env[62824]: DEBUG oslo_concurrency.lockutils [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1898.069501] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-90977a71-81c1-4648-8028-91844fec4092 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1898.080611] env[62824]: DEBUG oslo_vmware.api [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Waiting for the task: (returnval){ [ 1898.080611] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]523284b1-8dde-28b5-04e6-b6d539b901df" [ 1898.080611] env[62824]: _type = "Task" [ 1898.080611] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1898.092600] env[62824]: DEBUG oslo_vmware.api [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]523284b1-8dde-28b5-04e6-b6d539b901df, 'name': SearchDatastore_Task, 'duration_secs': 0.010336} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1898.092912] env[62824]: DEBUG oslo_concurrency.lockutils [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1898.093601] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1898.093601] env[62824]: DEBUG oslo_concurrency.lockutils [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1898.093601] env[62824]: DEBUG oslo_concurrency.lockutils [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1898.094078] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1898.094078] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9d5b064e-16be-41e2-9324-c9c9363d0f1a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1898.104669] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1898.104863] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1898.105624] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-70189478-3659-4c63-a66d-7fa8526e4337 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1898.117787] env[62824]: DEBUG oslo_vmware.api [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Waiting for the task: (returnval){ [ 1898.117787] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5218833e-6d70-82df-b7d8-df4dec4ed097" [ 1898.117787] env[62824]: _type = "Task" [ 1898.117787] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1898.130429] env[62824]: DEBUG oslo_vmware.api [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5218833e-6d70-82df-b7d8-df4dec4ed097, 'name': SearchDatastore_Task, 'duration_secs': 0.009516} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1898.131322] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-17964aa6-4fdd-4492-901f-2c5dc109c7b9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1898.140073] env[62824]: DEBUG oslo_vmware.api [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Waiting for the task: (returnval){ [ 1898.140073] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5265162c-a2bd-9297-0677-fad35a44c4d4" [ 1898.140073] env[62824]: _type = "Task" [ 1898.140073] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1898.152034] env[62824]: DEBUG oslo_vmware.api [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5265162c-a2bd-9297-0677-fad35a44c4d4, 'name': SearchDatastore_Task, 'duration_secs': 0.008679} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1898.154800] env[62824]: DEBUG oslo_concurrency.lockutils [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1898.155090] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 384a96ea-40ce-43e1-a5f9-82f50b710b0b/384a96ea-40ce-43e1-a5f9-82f50b710b0b.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1898.155682] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-deb763c0-fc83-49fd-8d98-8e0dc1d8e024 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1898.163035] env[62824]: DEBUG oslo_vmware.api [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Waiting for the task: (returnval){ [ 1898.163035] env[62824]: value = "task-2145702" [ 1898.163035] env[62824]: _type = "Task" [ 1898.163035] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1898.172941] env[62824]: DEBUG oslo_vmware.api [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Task: {'id': task-2145702, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1898.378127] env[62824]: DEBUG nova.network.neutron [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] [instance: dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda] Successfully created port: 9f1b3f36-9c69-4e99-8030-2d2ec535b324 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1898.386369] env[62824]: DEBUG oslo_vmware.api [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Task: {'id': task-2145701, 'name': ReconfigVM_Task, 'duration_secs': 0.774983} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1898.389941] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] [instance: 9062e606-511f-4fe7-9621-90b3c0e51eec] Reconfigured VM instance instance-00000013 to attach disk [datastore1] 9062e606-511f-4fe7-9621-90b3c0e51eec/9062e606-511f-4fe7-9621-90b3c0e51eec.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1898.391384] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-230d46b0-ede0-44d3-a2a7-ba8d04c7027d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1898.399892] env[62824]: DEBUG oslo_vmware.api [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Waiting for the task: (returnval){ [ 1898.399892] env[62824]: value = "task-2145703" [ 1898.399892] env[62824]: _type = "Task" [ 1898.399892] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1898.414130] env[62824]: DEBUG oslo_vmware.api [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Task: {'id': task-2145703, 'name': Rename_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1898.643938] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0aa4178e-f6f6-476a-8622-994763d3dee3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1898.654178] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7bd0b45-48e5-4a16-a8c4-dddf00fd1e4c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1898.702155] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7787321-da98-48dc-a550-47623a088d86 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1898.710508] env[62824]: DEBUG oslo_vmware.api [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Task: {'id': task-2145702, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.518301} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1898.713390] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 384a96ea-40ce-43e1-a5f9-82f50b710b0b/384a96ea-40ce-43e1-a5f9-82f50b710b0b.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1898.713633] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1898.713929] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d7c2ae08-b233-4cae-9b96-ab1bf5a7fb0a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1898.717279] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcbfaa08-4b26-480d-9447-96a0e98db714 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1898.732219] env[62824]: DEBUG nova.compute.provider_tree [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1898.735220] env[62824]: DEBUG oslo_vmware.api [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Waiting for the task: (returnval){ [ 1898.735220] env[62824]: value = "task-2145704" [ 1898.735220] env[62824]: _type = "Task" [ 1898.735220] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1898.743079] env[62824]: DEBUG oslo_vmware.api [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Task: {'id': task-2145704, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1898.910246] env[62824]: DEBUG oslo_vmware.api [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Task: {'id': task-2145703, 'name': Rename_Task, 'duration_secs': 0.190472} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1898.910511] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] [instance: 9062e606-511f-4fe7-9621-90b3c0e51eec] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1898.910773] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-05a9c093-4bd4-4830-a178-f53e8260b1b0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1898.917504] env[62824]: DEBUG oslo_vmware.api [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Waiting for the task: (returnval){ [ 1898.917504] env[62824]: value = "task-2145705" [ 1898.917504] env[62824]: _type = "Task" [ 1898.917504] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1898.920726] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7f338944-9cb1-4d55-9d22-6a91d729ee9f tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Acquiring lock "517bf2cf-e142-4f36-bf53-79a8d2d8a530" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1898.920972] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7f338944-9cb1-4d55-9d22-6a91d729ee9f tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Lock "517bf2cf-e142-4f36-bf53-79a8d2d8a530" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1898.921195] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7f338944-9cb1-4d55-9d22-6a91d729ee9f tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Acquiring lock "517bf2cf-e142-4f36-bf53-79a8d2d8a530-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1898.921384] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7f338944-9cb1-4d55-9d22-6a91d729ee9f tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Lock "517bf2cf-e142-4f36-bf53-79a8d2d8a530-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1898.921547] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7f338944-9cb1-4d55-9d22-6a91d729ee9f tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Lock "517bf2cf-e142-4f36-bf53-79a8d2d8a530-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1898.924080] env[62824]: INFO nova.compute.manager [None req-7f338944-9cb1-4d55-9d22-6a91d729ee9f tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] [instance: 517bf2cf-e142-4f36-bf53-79a8d2d8a530] Terminating instance [ 1898.928599] env[62824]: DEBUG oslo_vmware.api [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Task: {'id': task-2145705, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1899.035890] env[62824]: DEBUG oslo_concurrency.lockutils [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Acquiring lock "35f0435c-fe46-46ab-bc71-c21bb0ad3d55" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1899.036390] env[62824]: DEBUG oslo_concurrency.lockutils [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Lock "35f0435c-fe46-46ab-bc71-c21bb0ad3d55" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1899.043646] env[62824]: DEBUG nova.compute.manager [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] [instance: dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1899.076561] env[62824]: DEBUG nova.virt.hardware [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1899.076842] env[62824]: DEBUG nova.virt.hardware [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1899.077030] env[62824]: DEBUG nova.virt.hardware [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1899.077227] env[62824]: DEBUG nova.virt.hardware [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1899.077391] env[62824]: DEBUG nova.virt.hardware [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1899.077551] env[62824]: DEBUG nova.virt.hardware [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1899.079741] env[62824]: DEBUG nova.virt.hardware [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1899.079741] env[62824]: DEBUG nova.virt.hardware [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1899.079741] env[62824]: DEBUG nova.virt.hardware [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1899.079741] env[62824]: DEBUG nova.virt.hardware [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1899.079741] env[62824]: DEBUG nova.virt.hardware [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1899.079913] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2e4f590-3b58-4e84-b8ea-fa083b9d2567 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1899.087345] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-639f8d96-1eb4-40c7-9336-09ca97879e8b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1899.237208] env[62824]: DEBUG nova.scheduler.client.report [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1899.259310] env[62824]: DEBUG oslo_vmware.api [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Task: {'id': task-2145704, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.093935} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1899.259310] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1899.260568] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82318390-6e46-4fec-9985-184df0176e3f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1899.290657] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Reconfiguring VM instance instance-00000014 to attach disk [datastore2] 384a96ea-40ce-43e1-a5f9-82f50b710b0b/384a96ea-40ce-43e1-a5f9-82f50b710b0b.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1899.291659] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-216c5b4d-7b56-4c7c-a959-aa6e08d80e04 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1899.312659] env[62824]: DEBUG oslo_vmware.api [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Waiting for the task: (returnval){ [ 1899.312659] env[62824]: value = "task-2145706" [ 1899.312659] env[62824]: _type = "Task" [ 1899.312659] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1899.320597] env[62824]: DEBUG oslo_vmware.api [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Task: {'id': task-2145706, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1899.428960] env[62824]: DEBUG oslo_vmware.api [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Task: {'id': task-2145705, 'name': PowerOnVM_Task} progress is 37%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1899.433770] env[62824]: DEBUG nova.compute.manager [None req-7f338944-9cb1-4d55-9d22-6a91d729ee9f tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] [instance: 517bf2cf-e142-4f36-bf53-79a8d2d8a530] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1899.434055] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-7f338944-9cb1-4d55-9d22-6a91d729ee9f tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] [instance: 517bf2cf-e142-4f36-bf53-79a8d2d8a530] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1899.434910] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3aad079-b48a-43e3-89f6-3bcfa2b707ba {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1899.442763] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f338944-9cb1-4d55-9d22-6a91d729ee9f tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] [instance: 517bf2cf-e142-4f36-bf53-79a8d2d8a530] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1899.445692] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5146a772-91e4-4a7d-bd98-6e1c1318cc08 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1899.455023] env[62824]: DEBUG oslo_vmware.api [None req-7f338944-9cb1-4d55-9d22-6a91d729ee9f tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Waiting for the task: (returnval){ [ 1899.455023] env[62824]: value = "task-2145707" [ 1899.455023] env[62824]: _type = "Task" [ 1899.455023] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1899.461640] env[62824]: DEBUG oslo_vmware.api [None req-7f338944-9cb1-4d55-9d22-6a91d729ee9f tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Task: {'id': task-2145707, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1899.752273] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.743s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1899.752273] env[62824]: DEBUG nova.compute.manager [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] [instance: 4dfe9f82-7f89-486b-b159-948656d2896c] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1899.755896] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.093s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1899.757875] env[62824]: INFO nova.compute.claims [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1899.824622] env[62824]: DEBUG oslo_vmware.api [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Task: {'id': task-2145706, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1899.929572] env[62824]: DEBUG oslo_vmware.api [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Task: {'id': task-2145705, 'name': PowerOnVM_Task} progress is 91%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1899.963588] env[62824]: DEBUG oslo_vmware.api [None req-7f338944-9cb1-4d55-9d22-6a91d729ee9f tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Task: {'id': task-2145707, 'name': PowerOffVM_Task, 'duration_secs': 0.185948} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1899.963867] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f338944-9cb1-4d55-9d22-6a91d729ee9f tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] [instance: 517bf2cf-e142-4f36-bf53-79a8d2d8a530] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1899.964053] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-7f338944-9cb1-4d55-9d22-6a91d729ee9f tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] [instance: 517bf2cf-e142-4f36-bf53-79a8d2d8a530] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1899.964376] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6e1ac5d1-793e-4550-9bd9-2287159b31e9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1900.038585] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-7f338944-9cb1-4d55-9d22-6a91d729ee9f tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] [instance: 517bf2cf-e142-4f36-bf53-79a8d2d8a530] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1900.038585] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-7f338944-9cb1-4d55-9d22-6a91d729ee9f tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] [instance: 517bf2cf-e142-4f36-bf53-79a8d2d8a530] Deleting contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1900.038585] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-7f338944-9cb1-4d55-9d22-6a91d729ee9f tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Deleting the datastore file [datastore1] 517bf2cf-e142-4f36-bf53-79a8d2d8a530 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1900.038585] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2c4393db-b1a6-4bcf-ad40-c77fd1be4bad {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1900.048172] env[62824]: DEBUG oslo_vmware.api [None req-7f338944-9cb1-4d55-9d22-6a91d729ee9f tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Waiting for the task: (returnval){ [ 1900.048172] env[62824]: value = "task-2145709" [ 1900.048172] env[62824]: _type = "Task" [ 1900.048172] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1900.056641] env[62824]: DEBUG oslo_vmware.api [None req-7f338944-9cb1-4d55-9d22-6a91d729ee9f tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Task: {'id': task-2145709, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1900.265929] env[62824]: DEBUG nova.compute.utils [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1900.269849] env[62824]: DEBUG nova.compute.manager [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] [instance: 4dfe9f82-7f89-486b-b159-948656d2896c] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1900.269849] env[62824]: DEBUG nova.network.neutron [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] [instance: 4dfe9f82-7f89-486b-b159-948656d2896c] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1900.324874] env[62824]: DEBUG oslo_vmware.api [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Task: {'id': task-2145706, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1900.371610] env[62824]: DEBUG nova.policy [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b7adebaa2445434a8ec4532e57d4177a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a041687b6f0a423c9573c0da91d853c1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 1900.430271] env[62824]: DEBUG oslo_vmware.api [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Task: {'id': task-2145705, 'name': PowerOnVM_Task, 'duration_secs': 1.043056} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1900.430271] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] [instance: 9062e606-511f-4fe7-9621-90b3c0e51eec] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1900.430271] env[62824]: INFO nova.compute.manager [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] [instance: 9062e606-511f-4fe7-9621-90b3c0e51eec] Took 9.41 seconds to spawn the instance on the hypervisor. [ 1900.430271] env[62824]: DEBUG nova.compute.manager [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] [instance: 9062e606-511f-4fe7-9621-90b3c0e51eec] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1900.431135] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90bdb8fe-dd5c-450a-80e7-ad2c631e1f17 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1900.559543] env[62824]: DEBUG oslo_vmware.api [None req-7f338944-9cb1-4d55-9d22-6a91d729ee9f tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Task: {'id': task-2145709, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1900.776017] env[62824]: DEBUG nova.compute.manager [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] [instance: 4dfe9f82-7f89-486b-b159-948656d2896c] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1900.793274] env[62824]: DEBUG nova.network.neutron [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] [instance: dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda] Successfully updated port: 9f1b3f36-9c69-4e99-8030-2d2ec535b324 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1900.807505] env[62824]: DEBUG nova.compute.manager [req-7b0df93e-8c09-4aec-b9c2-6f5c71baa9ad req-2bd8d06a-f04f-4c42-a121-922dcf2eca42 service nova] [instance: dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda] Received event network-vif-plugged-9f1b3f36-9c69-4e99-8030-2d2ec535b324 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1900.807644] env[62824]: DEBUG oslo_concurrency.lockutils [req-7b0df93e-8c09-4aec-b9c2-6f5c71baa9ad req-2bd8d06a-f04f-4c42-a121-922dcf2eca42 service nova] Acquiring lock "dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1900.807967] env[62824]: DEBUG oslo_concurrency.lockutils [req-7b0df93e-8c09-4aec-b9c2-6f5c71baa9ad req-2bd8d06a-f04f-4c42-a121-922dcf2eca42 service nova] Lock "dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1900.807967] env[62824]: DEBUG oslo_concurrency.lockutils [req-7b0df93e-8c09-4aec-b9c2-6f5c71baa9ad req-2bd8d06a-f04f-4c42-a121-922dcf2eca42 service nova] Lock "dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1900.808159] env[62824]: DEBUG nova.compute.manager [req-7b0df93e-8c09-4aec-b9c2-6f5c71baa9ad req-2bd8d06a-f04f-4c42-a121-922dcf2eca42 service nova] [instance: dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda] No waiting events found dispatching network-vif-plugged-9f1b3f36-9c69-4e99-8030-2d2ec535b324 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1900.808348] env[62824]: WARNING nova.compute.manager [req-7b0df93e-8c09-4aec-b9c2-6f5c71baa9ad req-2bd8d06a-f04f-4c42-a121-922dcf2eca42 service nova] [instance: dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda] Received unexpected event network-vif-plugged-9f1b3f36-9c69-4e99-8030-2d2ec535b324 for instance with vm_state building and task_state spawning. [ 1900.831048] env[62824]: DEBUG oslo_vmware.api [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Task: {'id': task-2145706, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1900.962391] env[62824]: INFO nova.compute.manager [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] [instance: 9062e606-511f-4fe7-9621-90b3c0e51eec] Took 36.74 seconds to build instance. [ 1900.978874] env[62824]: DEBUG nova.network.neutron [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] [instance: 4dfe9f82-7f89-486b-b159-948656d2896c] Successfully created port: 4a48d848-c8ab-4cb1-8c18-8dc67595674e {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1901.058975] env[62824]: DEBUG oslo_vmware.api [None req-7f338944-9cb1-4d55-9d22-6a91d729ee9f tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Task: {'id': task-2145709, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.692027} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1901.059322] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-7f338944-9cb1-4d55-9d22-6a91d729ee9f tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1901.059322] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-7f338944-9cb1-4d55-9d22-6a91d729ee9f tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] [instance: 517bf2cf-e142-4f36-bf53-79a8d2d8a530] Deleted contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1901.059531] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-7f338944-9cb1-4d55-9d22-6a91d729ee9f tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] [instance: 517bf2cf-e142-4f36-bf53-79a8d2d8a530] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1901.059656] env[62824]: INFO nova.compute.manager [None req-7f338944-9cb1-4d55-9d22-6a91d729ee9f tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] [instance: 517bf2cf-e142-4f36-bf53-79a8d2d8a530] Took 1.63 seconds to destroy the instance on the hypervisor. [ 1901.059918] env[62824]: DEBUG oslo.service.loopingcall [None req-7f338944-9cb1-4d55-9d22-6a91d729ee9f tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1901.060254] env[62824]: DEBUG nova.compute.manager [-] [instance: 517bf2cf-e142-4f36-bf53-79a8d2d8a530] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1901.060392] env[62824]: DEBUG nova.network.neutron [-] [instance: 517bf2cf-e142-4f36-bf53-79a8d2d8a530] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1901.072462] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Acquiring lock "db018af0-7bdf-40c5-b129-fde15dfb8ab4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1901.072778] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Lock "db018af0-7bdf-40c5-b129-fde15dfb8ab4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1901.300633] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Acquiring lock "refresh_cache-dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1901.300633] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Acquired lock "refresh_cache-dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1901.300633] env[62824]: DEBUG nova.network.neutron [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] [instance: dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1901.309604] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68de8989-ddd2-4c04-88e6-de2a8f18d2e9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1901.323062] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-130ee7d4-fe12-459f-81ca-5bca8a68f61a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1901.337017] env[62824]: DEBUG oslo_vmware.api [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Task: {'id': task-2145706, 'name': ReconfigVM_Task, 'duration_secs': 1.528062} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1901.360114] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Reconfigured VM instance instance-00000014 to attach disk [datastore2] 384a96ea-40ce-43e1-a5f9-82f50b710b0b/384a96ea-40ce-43e1-a5f9-82f50b710b0b.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1901.364202] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9dc954f8-e4be-4369-b516-88fe72a910fb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1901.364850] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c3e0004-8783-464b-9b23-46ddfb89a999 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1901.376455] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9b40442-ffcf-4c61-b408-158e350714d3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1901.380371] env[62824]: DEBUG oslo_vmware.api [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Waiting for the task: (returnval){ [ 1901.380371] env[62824]: value = "task-2145710" [ 1901.380371] env[62824]: _type = "Task" [ 1901.380371] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1901.392431] env[62824]: DEBUG nova.compute.provider_tree [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1901.398849] env[62824]: DEBUG oslo_vmware.api [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Task: {'id': task-2145710, 'name': Rename_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1901.468076] env[62824]: DEBUG oslo_concurrency.lockutils [None req-31c37239-1922-4dfd-8158-578f2697917e tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Lock "9062e606-511f-4fe7-9621-90b3c0e51eec" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 57.492s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1901.794280] env[62824]: DEBUG nova.compute.manager [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] [instance: 4dfe9f82-7f89-486b-b159-948656d2896c] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1901.837044] env[62824]: DEBUG nova.virt.hardware [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1901.837275] env[62824]: DEBUG nova.virt.hardware [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1901.837482] env[62824]: DEBUG nova.virt.hardware [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1901.837619] env[62824]: DEBUG nova.virt.hardware [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1901.837908] env[62824]: DEBUG nova.virt.hardware [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1901.838138] env[62824]: DEBUG nova.virt.hardware [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1901.838334] env[62824]: DEBUG nova.virt.hardware [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1901.838508] env[62824]: DEBUG nova.virt.hardware [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1901.838751] env[62824]: DEBUG nova.virt.hardware [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1901.838907] env[62824]: DEBUG nova.virt.hardware [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1901.839028] env[62824]: DEBUG nova.virt.hardware [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1901.840530] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6148489d-303d-425c-9026-313c73e8b605 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1901.848515] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c75ebc6-7a9c-45e8-b3ba-736f7b0ea5b9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1901.853165] env[62824]: DEBUG nova.network.neutron [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] [instance: dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1901.890608] env[62824]: DEBUG oslo_vmware.api [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Task: {'id': task-2145710, 'name': Rename_Task, 'duration_secs': 0.145196} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1901.890608] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1901.890608] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a7386154-59ef-4579-bf7e-182879cdb24a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1901.897474] env[62824]: DEBUG oslo_vmware.api [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Waiting for the task: (returnval){ [ 1901.897474] env[62824]: value = "task-2145711" [ 1901.897474] env[62824]: _type = "Task" [ 1901.897474] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1901.906796] env[62824]: DEBUG nova.scheduler.client.report [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1901.910124] env[62824]: DEBUG oslo_vmware.api [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Task: {'id': task-2145711, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1901.970385] env[62824]: DEBUG nova.compute.manager [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1902.083475] env[62824]: DEBUG nova.network.neutron [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] [instance: dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda] Updating instance_info_cache with network_info: [{"id": "9f1b3f36-9c69-4e99-8030-2d2ec535b324", "address": "fa:16:3e:91:83:1b", "network": {"id": "a0e32d69-182a-4e0c-a984-bfe393b641d7", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-2040711326-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "51135bed305c4e89afaeed2b187cc7f6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "274afb4c-04df-4213-8ad2-8f48a10d78a8", "external-id": "nsx-vlan-transportzone-515", "segmentation_id": 515, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9f1b3f36-9c", "ovs_interfaceid": "9f1b3f36-9c69-4e99-8030-2d2ec535b324", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1902.288997] env[62824]: DEBUG nova.network.neutron [-] [instance: 517bf2cf-e142-4f36-bf53-79a8d2d8a530] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1902.377207] env[62824]: DEBUG oslo_concurrency.lockutils [None req-95c8b711-718a-4a21-abec-a4473e9dbb40 tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Acquiring lock "9062e606-511f-4fe7-9621-90b3c0e51eec" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1902.377462] env[62824]: DEBUG oslo_concurrency.lockutils [None req-95c8b711-718a-4a21-abec-a4473e9dbb40 tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Lock "9062e606-511f-4fe7-9621-90b3c0e51eec" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1902.377705] env[62824]: INFO nova.compute.manager [None req-95c8b711-718a-4a21-abec-a4473e9dbb40 tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] [instance: 9062e606-511f-4fe7-9621-90b3c0e51eec] Rebooting instance [ 1902.408461] env[62824]: DEBUG oslo_vmware.api [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Task: {'id': task-2145711, 'name': PowerOnVM_Task, 'duration_secs': 0.433667} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1902.408699] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1902.408905] env[62824]: INFO nova.compute.manager [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Took 6.01 seconds to spawn the instance on the hypervisor. [ 1902.409100] env[62824]: DEBUG nova.compute.manager [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1902.409868] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8f778d6-235b-4c00-82db-0e75519684e4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1902.412897] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.657s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1902.413383] env[62824]: DEBUG nova.compute.manager [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1902.415970] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fa6dedf3-2458-4e1e-a821-8c6b0570fd79 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.796s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1902.416204] env[62824]: DEBUG nova.objects.instance [None req-fa6dedf3-2458-4e1e-a821-8c6b0570fd79 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Lazy-loading 'resources' on Instance uuid 04ac4ade-0bc8-4469-9b1b-f288b2f0a367 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1902.493810] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1902.586712] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Releasing lock "refresh_cache-dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1902.587086] env[62824]: DEBUG nova.compute.manager [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] [instance: dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda] Instance network_info: |[{"id": "9f1b3f36-9c69-4e99-8030-2d2ec535b324", "address": "fa:16:3e:91:83:1b", "network": {"id": "a0e32d69-182a-4e0c-a984-bfe393b641d7", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-2040711326-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "51135bed305c4e89afaeed2b187cc7f6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "274afb4c-04df-4213-8ad2-8f48a10d78a8", "external-id": "nsx-vlan-transportzone-515", "segmentation_id": 515, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9f1b3f36-9c", "ovs_interfaceid": "9f1b3f36-9c69-4e99-8030-2d2ec535b324", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1902.587597] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] [instance: dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:91:83:1b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '274afb4c-04df-4213-8ad2-8f48a10d78a8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9f1b3f36-9c69-4e99-8030-2d2ec535b324', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1902.595738] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Creating folder: Project (51135bed305c4e89afaeed2b187cc7f6). Parent ref: group-v438503. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1902.596051] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-dc195318-5318-49ac-bf85-f7d935fab417 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1902.607713] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Created folder: Project (51135bed305c4e89afaeed2b187cc7f6) in parent group-v438503. [ 1902.607934] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Creating folder: Instances. Parent ref: group-v438535. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1902.608202] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-67125a8d-3819-41f6-8f30-0d695158bb7c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1902.618050] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Created folder: Instances in parent group-v438535. [ 1902.618316] env[62824]: DEBUG oslo.service.loopingcall [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1902.618662] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1902.618785] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-34eff4da-b40c-4f60-9c43-ef28501f3233 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1902.638051] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1902.638051] env[62824]: value = "task-2145714" [ 1902.638051] env[62824]: _type = "Task" [ 1902.638051] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1902.646455] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145714, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1902.792284] env[62824]: INFO nova.compute.manager [-] [instance: 517bf2cf-e142-4f36-bf53-79a8d2d8a530] Took 1.73 seconds to deallocate network for instance. [ 1902.835585] env[62824]: DEBUG nova.compute.manager [req-a918e527-9b21-4392-8337-fef6422404cf req-bd11ab9c-8617-4959-91d7-70dc6bf4803e service nova] [instance: dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda] Received event network-changed-9f1b3f36-9c69-4e99-8030-2d2ec535b324 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1902.836139] env[62824]: DEBUG nova.compute.manager [req-a918e527-9b21-4392-8337-fef6422404cf req-bd11ab9c-8617-4959-91d7-70dc6bf4803e service nova] [instance: dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda] Refreshing instance network info cache due to event network-changed-9f1b3f36-9c69-4e99-8030-2d2ec535b324. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1902.836387] env[62824]: DEBUG oslo_concurrency.lockutils [req-a918e527-9b21-4392-8337-fef6422404cf req-bd11ab9c-8617-4959-91d7-70dc6bf4803e service nova] Acquiring lock "refresh_cache-dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1902.836693] env[62824]: DEBUG oslo_concurrency.lockutils [req-a918e527-9b21-4392-8337-fef6422404cf req-bd11ab9c-8617-4959-91d7-70dc6bf4803e service nova] Acquired lock "refresh_cache-dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1902.836959] env[62824]: DEBUG nova.network.neutron [req-a918e527-9b21-4392-8337-fef6422404cf req-bd11ab9c-8617-4959-91d7-70dc6bf4803e service nova] [instance: dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda] Refreshing network info cache for port 9f1b3f36-9c69-4e99-8030-2d2ec535b324 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1902.910897] env[62824]: DEBUG oslo_concurrency.lockutils [None req-95c8b711-718a-4a21-abec-a4473e9dbb40 tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Acquiring lock "refresh_cache-9062e606-511f-4fe7-9621-90b3c0e51eec" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1902.911114] env[62824]: DEBUG oslo_concurrency.lockutils [None req-95c8b711-718a-4a21-abec-a4473e9dbb40 tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Acquired lock "refresh_cache-9062e606-511f-4fe7-9621-90b3c0e51eec" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1902.911298] env[62824]: DEBUG nova.network.neutron [None req-95c8b711-718a-4a21-abec-a4473e9dbb40 tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] [instance: 9062e606-511f-4fe7-9621-90b3c0e51eec] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1902.920277] env[62824]: DEBUG nova.compute.utils [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1902.926378] env[62824]: DEBUG nova.compute.manager [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1902.926543] env[62824]: DEBUG nova.network.neutron [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1902.935058] env[62824]: INFO nova.compute.manager [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Took 27.50 seconds to build instance. [ 1903.047356] env[62824]: DEBUG nova.policy [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2cbf8eb928484d54a5371365386e73f6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b2f9084931bc4b94b0ae85e3e4b0ba90', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 1903.151384] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145714, 'name': CreateVM_Task, 'duration_secs': 0.309951} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1903.151692] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1903.152400] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1903.152546] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1903.152876] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1903.153134] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ce2ef37a-afcb-47aa-b271-628270634ed6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1903.161303] env[62824]: DEBUG oslo_vmware.api [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Waiting for the task: (returnval){ [ 1903.161303] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5270999b-8324-6319-47a6-d0899d090664" [ 1903.161303] env[62824]: _type = "Task" [ 1903.161303] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1903.169682] env[62824]: DEBUG oslo_vmware.api [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5270999b-8324-6319-47a6-d0899d090664, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1903.299183] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7f338944-9cb1-4d55-9d22-6a91d729ee9f tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1903.427729] env[62824]: DEBUG nova.compute.manager [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1903.437067] env[62824]: DEBUG oslo_concurrency.lockutils [None req-59daab71-b9e7-49e4-bfe5-f6f843ea7a1f tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Lock "384a96ea-40ce-43e1-a5f9-82f50b710b0b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 56.321s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1903.440740] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f4f0f6b-2333-4d9b-9ece-fd85668b3080 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1903.449090] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca8ecf56-5349-4d83-954d-e3bf4e906221 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1903.482652] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84c44806-149c-4117-a122-bcae33a4c3e3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1903.492971] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13d8889e-f117-4c23-bd1c-029613be47ee {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1903.508724] env[62824]: DEBUG nova.compute.provider_tree [None req-fa6dedf3-2458-4e1e-a821-8c6b0570fd79 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1903.537352] env[62824]: DEBUG nova.network.neutron [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] [instance: 4dfe9f82-7f89-486b-b159-948656d2896c] Successfully updated port: 4a48d848-c8ab-4cb1-8c18-8dc67595674e {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1903.649833] env[62824]: DEBUG nova.network.neutron [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Successfully created port: 02e90c28-7f36-4d83-84bb-fec194a04dab {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1903.671482] env[62824]: DEBUG oslo_vmware.api [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5270999b-8324-6319-47a6-d0899d090664, 'name': SearchDatastore_Task, 'duration_secs': 0.046317} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1903.674569] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1903.674809] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] [instance: dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1903.675685] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1903.675892] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1903.676094] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1903.676368] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-40277a7a-8a6d-488e-8178-49582c4419d9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1903.687441] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1903.687590] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1903.688353] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fe79d8f8-d918-4356-a192-20d75ed63e3f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1903.694380] env[62824]: DEBUG oslo_vmware.api [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Waiting for the task: (returnval){ [ 1903.694380] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]524464cf-debd-fde5-f70c-27017038b0de" [ 1903.694380] env[62824]: _type = "Task" [ 1903.694380] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1903.703763] env[62824]: DEBUG oslo_vmware.api [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]524464cf-debd-fde5-f70c-27017038b0de, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1903.752657] env[62824]: DEBUG nova.network.neutron [req-a918e527-9b21-4392-8337-fef6422404cf req-bd11ab9c-8617-4959-91d7-70dc6bf4803e service nova] [instance: dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda] Updated VIF entry in instance network info cache for port 9f1b3f36-9c69-4e99-8030-2d2ec535b324. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1903.753144] env[62824]: DEBUG nova.network.neutron [req-a918e527-9b21-4392-8337-fef6422404cf req-bd11ab9c-8617-4959-91d7-70dc6bf4803e service nova] [instance: dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda] Updating instance_info_cache with network_info: [{"id": "9f1b3f36-9c69-4e99-8030-2d2ec535b324", "address": "fa:16:3e:91:83:1b", "network": {"id": "a0e32d69-182a-4e0c-a984-bfe393b641d7", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-2040711326-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "51135bed305c4e89afaeed2b187cc7f6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "274afb4c-04df-4213-8ad2-8f48a10d78a8", "external-id": "nsx-vlan-transportzone-515", "segmentation_id": 515, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9f1b3f36-9c", "ovs_interfaceid": "9f1b3f36-9c69-4e99-8030-2d2ec535b324", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1903.791553] env[62824]: DEBUG nova.network.neutron [None req-95c8b711-718a-4a21-abec-a4473e9dbb40 tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] [instance: 9062e606-511f-4fe7-9621-90b3c0e51eec] Updating instance_info_cache with network_info: [{"id": "3c79003f-579c-4415-8515-cee7f296a297", "address": "fa:16:3e:1a:d4:42", "network": {"id": "7aa6d314-a33f-4c6e-a94e-c0cf1601d6f7", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-1542833912-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "950ce635c14048ba9f7e8f0e7aafa8af", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "be8bd197-4b2b-46e7-88ea-2554b0438584", "external-id": "nsx-vlan-transportzone-338", "segmentation_id": 338, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3c79003f-57", "ovs_interfaceid": "3c79003f-579c-4415-8515-cee7f296a297", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1903.939985] env[62824]: DEBUG nova.compute.manager [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] [instance: c36ece43-3d70-4e67-a740-9057f413c722] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1904.012162] env[62824]: DEBUG nova.scheduler.client.report [None req-fa6dedf3-2458-4e1e-a821-8c6b0570fd79 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1904.040421] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Acquiring lock "refresh_cache-4dfe9f82-7f89-486b-b159-948656d2896c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1904.040421] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Acquired lock "refresh_cache-4dfe9f82-7f89-486b-b159-948656d2896c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1904.040421] env[62824]: DEBUG nova.network.neutron [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] [instance: 4dfe9f82-7f89-486b-b159-948656d2896c] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1904.204650] env[62824]: DEBUG oslo_vmware.api [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]524464cf-debd-fde5-f70c-27017038b0de, 'name': SearchDatastore_Task, 'duration_secs': 0.01947} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1904.205451] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8cd229de-1381-493e-8bac-5dfed17f6d7c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1904.210759] env[62824]: DEBUG oslo_vmware.api [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Waiting for the task: (returnval){ [ 1904.210759] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5295d8c0-4379-0628-2481-f0f2adfc6546" [ 1904.210759] env[62824]: _type = "Task" [ 1904.210759] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1904.218416] env[62824]: DEBUG oslo_vmware.api [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5295d8c0-4379-0628-2481-f0f2adfc6546, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1904.257584] env[62824]: DEBUG oslo_concurrency.lockutils [req-a918e527-9b21-4392-8337-fef6422404cf req-bd11ab9c-8617-4959-91d7-70dc6bf4803e service nova] Releasing lock "refresh_cache-dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1904.257584] env[62824]: DEBUG nova.compute.manager [req-a918e527-9b21-4392-8337-fef6422404cf req-bd11ab9c-8617-4959-91d7-70dc6bf4803e service nova] [instance: 517bf2cf-e142-4f36-bf53-79a8d2d8a530] Received event network-vif-deleted-a245bb15-a02d-4ed0-a558-e3be831df931 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1904.268214] env[62824]: INFO nova.compute.manager [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Rebuilding instance [ 1904.295248] env[62824]: DEBUG oslo_concurrency.lockutils [None req-95c8b711-718a-4a21-abec-a4473e9dbb40 tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Releasing lock "refresh_cache-9062e606-511f-4fe7-9621-90b3c0e51eec" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1904.320670] env[62824]: DEBUG nova.compute.manager [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1904.321750] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a656e2a-541b-48d2-b17f-cbcd59917bc1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1904.437727] env[62824]: DEBUG nova.compute.manager [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1904.460796] env[62824]: DEBUG nova.virt.hardware [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:50:59Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='1bdf5fc7-1e8e-4668-88fe-6c47a0097f11',id=29,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-1536379991',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1904.461050] env[62824]: DEBUG nova.virt.hardware [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1904.461288] env[62824]: DEBUG nova.virt.hardware [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1904.461581] env[62824]: DEBUG nova.virt.hardware [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1904.461823] env[62824]: DEBUG nova.virt.hardware [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1904.462087] env[62824]: DEBUG nova.virt.hardware [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1904.462401] env[62824]: DEBUG nova.virt.hardware [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1904.462591] env[62824]: DEBUG nova.virt.hardware [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1904.462770] env[62824]: DEBUG nova.virt.hardware [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1904.462938] env[62824]: DEBUG nova.virt.hardware [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1904.463125] env[62824]: DEBUG nova.virt.hardware [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1904.463975] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54c8e41c-f000-4076-bc22-7cf0ce0a2d8c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1904.467244] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1904.472493] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91f46d6d-a5d6-4273-85f6-60a0aaae4efe {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1904.520999] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fa6dedf3-2458-4e1e-a821-8c6b0570fd79 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.105s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1904.523515] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.508s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1904.524955] env[62824]: INFO nova.compute.claims [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1904.541594] env[62824]: INFO nova.scheduler.client.report [None req-fa6dedf3-2458-4e1e-a821-8c6b0570fd79 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Deleted allocations for instance 04ac4ade-0bc8-4469-9b1b-f288b2f0a367 [ 1904.578296] env[62824]: DEBUG nova.network.neutron [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] [instance: 4dfe9f82-7f89-486b-b159-948656d2896c] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1904.721608] env[62824]: DEBUG oslo_vmware.api [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5295d8c0-4379-0628-2481-f0f2adfc6546, 'name': SearchDatastore_Task, 'duration_secs': 0.044426} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1904.721847] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1904.722116] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda/dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1904.722390] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-635f41c0-0045-47ff-8930-84e19196b77b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1904.729968] env[62824]: DEBUG oslo_vmware.api [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Waiting for the task: (returnval){ [ 1904.729968] env[62824]: value = "task-2145715" [ 1904.729968] env[62824]: _type = "Task" [ 1904.729968] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1904.733938] env[62824]: DEBUG nova.network.neutron [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] [instance: 4dfe9f82-7f89-486b-b159-948656d2896c] Updating instance_info_cache with network_info: [{"id": "4a48d848-c8ab-4cb1-8c18-8dc67595674e", "address": "fa:16:3e:70:ef:af", "network": {"id": "f8a50884-dea5-4ba0-8484-553e2c83aaa9", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-1131986537-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a041687b6f0a423c9573c0da91d853c1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74e6f6e0-95e6-4531-99e9-0e78350fb655", "external-id": "nsx-vlan-transportzone-896", "segmentation_id": 896, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4a48d848-c8", "ovs_interfaceid": "4a48d848-c8ab-4cb1-8c18-8dc67595674e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1904.740014] env[62824]: DEBUG oslo_vmware.api [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Task: {'id': task-2145715, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1904.800730] env[62824]: DEBUG nova.compute.manager [None req-95c8b711-718a-4a21-abec-a4473e9dbb40 tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] [instance: 9062e606-511f-4fe7-9621-90b3c0e51eec] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1904.802027] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4de4f255-e95a-43c9-a6df-f049f1afb36b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1904.862296] env[62824]: DEBUG nova.compute.manager [req-1beb9071-cb8e-4998-8e14-eda2b5031ed3 req-2cbf581d-4a7d-42ec-afb6-5df197959a1f service nova] [instance: 4dfe9f82-7f89-486b-b159-948656d2896c] Received event network-vif-plugged-4a48d848-c8ab-4cb1-8c18-8dc67595674e {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1904.862581] env[62824]: DEBUG oslo_concurrency.lockutils [req-1beb9071-cb8e-4998-8e14-eda2b5031ed3 req-2cbf581d-4a7d-42ec-afb6-5df197959a1f service nova] Acquiring lock "4dfe9f82-7f89-486b-b159-948656d2896c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1904.862859] env[62824]: DEBUG oslo_concurrency.lockutils [req-1beb9071-cb8e-4998-8e14-eda2b5031ed3 req-2cbf581d-4a7d-42ec-afb6-5df197959a1f service nova] Lock "4dfe9f82-7f89-486b-b159-948656d2896c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1904.863079] env[62824]: DEBUG oslo_concurrency.lockutils [req-1beb9071-cb8e-4998-8e14-eda2b5031ed3 req-2cbf581d-4a7d-42ec-afb6-5df197959a1f service nova] Lock "4dfe9f82-7f89-486b-b159-948656d2896c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1904.863488] env[62824]: DEBUG nova.compute.manager [req-1beb9071-cb8e-4998-8e14-eda2b5031ed3 req-2cbf581d-4a7d-42ec-afb6-5df197959a1f service nova] [instance: 4dfe9f82-7f89-486b-b159-948656d2896c] No waiting events found dispatching network-vif-plugged-4a48d848-c8ab-4cb1-8c18-8dc67595674e {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1904.863569] env[62824]: WARNING nova.compute.manager [req-1beb9071-cb8e-4998-8e14-eda2b5031ed3 req-2cbf581d-4a7d-42ec-afb6-5df197959a1f service nova] [instance: 4dfe9f82-7f89-486b-b159-948656d2896c] Received unexpected event network-vif-plugged-4a48d848-c8ab-4cb1-8c18-8dc67595674e for instance with vm_state building and task_state spawning. [ 1904.863836] env[62824]: DEBUG nova.compute.manager [req-1beb9071-cb8e-4998-8e14-eda2b5031ed3 req-2cbf581d-4a7d-42ec-afb6-5df197959a1f service nova] [instance: 4dfe9f82-7f89-486b-b159-948656d2896c] Received event network-changed-4a48d848-c8ab-4cb1-8c18-8dc67595674e {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1904.863959] env[62824]: DEBUG nova.compute.manager [req-1beb9071-cb8e-4998-8e14-eda2b5031ed3 req-2cbf581d-4a7d-42ec-afb6-5df197959a1f service nova] [instance: 4dfe9f82-7f89-486b-b159-948656d2896c] Refreshing instance network info cache due to event network-changed-4a48d848-c8ab-4cb1-8c18-8dc67595674e. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1904.864160] env[62824]: DEBUG oslo_concurrency.lockutils [req-1beb9071-cb8e-4998-8e14-eda2b5031ed3 req-2cbf581d-4a7d-42ec-afb6-5df197959a1f service nova] Acquiring lock "refresh_cache-4dfe9f82-7f89-486b-b159-948656d2896c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1905.055178] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fa6dedf3-2458-4e1e-a821-8c6b0570fd79 tempest-ServerDiagnosticsNegativeTest-1804916441 tempest-ServerDiagnosticsNegativeTest-1804916441-project-member] Lock "04ac4ade-0bc8-4469-9b1b-f288b2f0a367" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.099s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1905.237449] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Releasing lock "refresh_cache-4dfe9f82-7f89-486b-b159-948656d2896c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1905.238726] env[62824]: DEBUG nova.compute.manager [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] [instance: 4dfe9f82-7f89-486b-b159-948656d2896c] Instance network_info: |[{"id": "4a48d848-c8ab-4cb1-8c18-8dc67595674e", "address": "fa:16:3e:70:ef:af", "network": {"id": "f8a50884-dea5-4ba0-8484-553e2c83aaa9", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-1131986537-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a041687b6f0a423c9573c0da91d853c1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74e6f6e0-95e6-4531-99e9-0e78350fb655", "external-id": "nsx-vlan-transportzone-896", "segmentation_id": 896, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4a48d848-c8", "ovs_interfaceid": "4a48d848-c8ab-4cb1-8c18-8dc67595674e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1905.238726] env[62824]: DEBUG oslo_concurrency.lockutils [req-1beb9071-cb8e-4998-8e14-eda2b5031ed3 req-2cbf581d-4a7d-42ec-afb6-5df197959a1f service nova] Acquired lock "refresh_cache-4dfe9f82-7f89-486b-b159-948656d2896c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1905.238881] env[62824]: DEBUG nova.network.neutron [req-1beb9071-cb8e-4998-8e14-eda2b5031ed3 req-2cbf581d-4a7d-42ec-afb6-5df197959a1f service nova] [instance: 4dfe9f82-7f89-486b-b159-948656d2896c] Refreshing network info cache for port 4a48d848-c8ab-4cb1-8c18-8dc67595674e {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1905.240528] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] [instance: 4dfe9f82-7f89-486b-b159-948656d2896c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:70:ef:af', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '74e6f6e0-95e6-4531-99e9-0e78350fb655', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4a48d848-c8ab-4cb1-8c18-8dc67595674e', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1905.254827] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Creating folder: Project (a041687b6f0a423c9573c0da91d853c1). Parent ref: group-v438503. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1905.261147] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e51300a9-5765-4f6c-a3a9-9902c92a9358 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1905.269092] env[62824]: DEBUG oslo_vmware.api [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Task: {'id': task-2145715, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1905.278415] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Created folder: Project (a041687b6f0a423c9573c0da91d853c1) in parent group-v438503. [ 1905.278663] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Creating folder: Instances. Parent ref: group-v438538. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1905.278973] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ae7cf4a7-c8e0-493c-a9bc-535693115903 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1905.288855] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Created folder: Instances in parent group-v438538. [ 1905.289143] env[62824]: DEBUG oslo.service.loopingcall [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1905.289306] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4dfe9f82-7f89-486b-b159-948656d2896c] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1905.289513] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-48ca2f8a-6788-47cb-95f4-ccb3817d786b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1905.317186] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1905.317186] env[62824]: value = "task-2145718" [ 1905.317186] env[62824]: _type = "Task" [ 1905.317186] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1905.327254] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145718, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1905.335239] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1905.335607] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a67f8a95-4788-49c5-9671-7bd48a838e87 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1905.342968] env[62824]: DEBUG oslo_vmware.api [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Waiting for the task: (returnval){ [ 1905.342968] env[62824]: value = "task-2145719" [ 1905.342968] env[62824]: _type = "Task" [ 1905.342968] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1905.358270] env[62824]: DEBUG oslo_vmware.api [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Task: {'id': task-2145719, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1905.554573] env[62824]: DEBUG nova.network.neutron [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Successfully updated port: 02e90c28-7f36-4d83-84bb-fec194a04dab {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1905.564337] env[62824]: DEBUG nova.network.neutron [req-1beb9071-cb8e-4998-8e14-eda2b5031ed3 req-2cbf581d-4a7d-42ec-afb6-5df197959a1f service nova] [instance: 4dfe9f82-7f89-486b-b159-948656d2896c] Updated VIF entry in instance network info cache for port 4a48d848-c8ab-4cb1-8c18-8dc67595674e. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1905.564905] env[62824]: DEBUG nova.network.neutron [req-1beb9071-cb8e-4998-8e14-eda2b5031ed3 req-2cbf581d-4a7d-42ec-afb6-5df197959a1f service nova] [instance: 4dfe9f82-7f89-486b-b159-948656d2896c] Updating instance_info_cache with network_info: [{"id": "4a48d848-c8ab-4cb1-8c18-8dc67595674e", "address": "fa:16:3e:70:ef:af", "network": {"id": "f8a50884-dea5-4ba0-8484-553e2c83aaa9", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-1131986537-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a041687b6f0a423c9573c0da91d853c1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "74e6f6e0-95e6-4531-99e9-0e78350fb655", "external-id": "nsx-vlan-transportzone-896", "segmentation_id": 896, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4a48d848-c8", "ovs_interfaceid": "4a48d848-c8ab-4cb1-8c18-8dc67595674e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1905.746740] env[62824]: DEBUG oslo_vmware.api [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Task: {'id': task-2145715, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1905.819637] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ab2cd6d-0e6e-4e0a-a7d7-072509ba9e88 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1905.834509] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145718, 'name': CreateVM_Task, 'duration_secs': 0.432054} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1905.836755] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4dfe9f82-7f89-486b-b159-948656d2896c] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1905.837041] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-95c8b711-718a-4a21-abec-a4473e9dbb40 tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] [instance: 9062e606-511f-4fe7-9621-90b3c0e51eec] Doing hard reboot of VM {{(pid=62824) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 1905.837704] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1905.837867] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1905.838256] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1905.838683] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-c4f2429c-1c9c-4b2e-ab32-25bf85243ba1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1905.842245] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a78ba7a8-87a1-4210-a58d-7f78d1748690 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1905.849544] env[62824]: DEBUG oslo_vmware.api [None req-95c8b711-718a-4a21-abec-a4473e9dbb40 tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Waiting for the task: (returnval){ [ 1905.849544] env[62824]: value = "task-2145720" [ 1905.849544] env[62824]: _type = "Task" [ 1905.849544] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1905.850736] env[62824]: DEBUG oslo_vmware.api [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Waiting for the task: (returnval){ [ 1905.850736] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52cb5e9c-c551-ffe1-50b0-c5685fa65623" [ 1905.850736] env[62824]: _type = "Task" [ 1905.850736] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1905.857246] env[62824]: DEBUG oslo_vmware.api [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Task: {'id': task-2145719, 'name': PowerOffVM_Task, 'duration_secs': 0.132169} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1905.862689] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1905.862930] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1905.864202] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9cda67a-f601-43e8-aa92-aefff6b17acd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1905.871503] env[62824]: DEBUG oslo_vmware.api [None req-95c8b711-718a-4a21-abec-a4473e9dbb40 tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Task: {'id': task-2145720, 'name': ResetVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1905.878068] env[62824]: DEBUG oslo_vmware.api [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52cb5e9c-c551-ffe1-50b0-c5685fa65623, 'name': SearchDatastore_Task, 'duration_secs': 0.01953} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1905.884346] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1905.884652] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] [instance: 4dfe9f82-7f89-486b-b159-948656d2896c] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1905.884965] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1905.885189] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1905.885449] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1905.885821] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1905.886374] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-637f860d-d90d-42d7-8ef5-73c7180be97c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1905.888627] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7a3089b7-5feb-45c7-8ad3-7813c9e4067e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1905.899185] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1905.899381] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1905.900112] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-adfd2d74-3bd8-40fa-b9c4-2fb4e37711b0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1905.907147] env[62824]: DEBUG oslo_vmware.api [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Waiting for the task: (returnval){ [ 1905.907147] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52a38f31-c28c-9108-e2a3-8bb8fd55ebe2" [ 1905.907147] env[62824]: _type = "Task" [ 1905.907147] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1905.916145] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1905.916344] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Deleting contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1905.916514] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Deleting the datastore file [datastore2] 384a96ea-40ce-43e1-a5f9-82f50b710b0b {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1905.916747] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-da750098-58b5-45ad-a20a-aee8ef03b988 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1905.921407] env[62824]: DEBUG oslo_vmware.api [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52a38f31-c28c-9108-e2a3-8bb8fd55ebe2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1905.927776] env[62824]: DEBUG oslo_vmware.api [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Waiting for the task: (returnval){ [ 1905.927776] env[62824]: value = "task-2145722" [ 1905.927776] env[62824]: _type = "Task" [ 1905.927776] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1905.935051] env[62824]: DEBUG oslo_vmware.api [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Task: {'id': task-2145722, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1906.029198] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f8f5b92-5cb8-4977-b947-ab73170c0e29 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1906.036701] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3369b64-3a22-45a3-bda2-b867143aea42 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1906.066595] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Acquiring lock "refresh_cache-e158b5d4-c120-4e6c-89c4-7668e097926d" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1906.066728] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Acquired lock "refresh_cache-e158b5d4-c120-4e6c-89c4-7668e097926d" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1906.066869] env[62824]: DEBUG nova.network.neutron [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1906.068611] env[62824]: DEBUG oslo_concurrency.lockutils [req-1beb9071-cb8e-4998-8e14-eda2b5031ed3 req-2cbf581d-4a7d-42ec-afb6-5df197959a1f service nova] Releasing lock "refresh_cache-4dfe9f82-7f89-486b-b159-948656d2896c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1906.070256] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60e70a39-5395-4f42-a232-c61b92da66e0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1906.078480] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8090413d-56e7-4a6e-a2c8-03df1866d919 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1906.093195] env[62824]: DEBUG nova.compute.provider_tree [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1906.241180] env[62824]: DEBUG oslo_vmware.api [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Task: {'id': task-2145715, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.461216} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1906.241443] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda/dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1906.241650] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] [instance: dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1906.241901] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-07ffd05d-f50e-42c8-ae28-ee940f341254 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1906.248697] env[62824]: DEBUG oslo_vmware.api [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Waiting for the task: (returnval){ [ 1906.248697] env[62824]: value = "task-2145723" [ 1906.248697] env[62824]: _type = "Task" [ 1906.248697] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1906.256625] env[62824]: DEBUG oslo_vmware.api [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Task: {'id': task-2145723, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1906.359848] env[62824]: DEBUG oslo_vmware.api [None req-95c8b711-718a-4a21-abec-a4473e9dbb40 tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Task: {'id': task-2145720, 'name': ResetVM_Task, 'duration_secs': 0.440958} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1906.360581] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-95c8b711-718a-4a21-abec-a4473e9dbb40 tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] [instance: 9062e606-511f-4fe7-9621-90b3c0e51eec] Did hard reboot of VM {{(pid=62824) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 1906.360581] env[62824]: DEBUG nova.compute.manager [None req-95c8b711-718a-4a21-abec-a4473e9dbb40 tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] [instance: 9062e606-511f-4fe7-9621-90b3c0e51eec] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1906.361263] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d229be83-4bcc-4803-a96b-488009356a3b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1906.418614] env[62824]: DEBUG oslo_vmware.api [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52a38f31-c28c-9108-e2a3-8bb8fd55ebe2, 'name': SearchDatastore_Task, 'duration_secs': 0.033189} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1906.419473] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-31f6761d-21bb-4770-b454-f015f07adc79 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1906.425622] env[62824]: DEBUG oslo_vmware.api [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Waiting for the task: (returnval){ [ 1906.425622] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5211270b-0cf8-4662-1cc7-39c6f176468b" [ 1906.425622] env[62824]: _type = "Task" [ 1906.425622] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1906.438283] env[62824]: DEBUG oslo_vmware.api [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Task: {'id': task-2145722, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.36037} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1906.442582] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1906.443110] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Deleted contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1906.443508] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1906.446814] env[62824]: DEBUG oslo_vmware.api [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5211270b-0cf8-4662-1cc7-39c6f176468b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1906.597222] env[62824]: DEBUG nova.scheduler.client.report [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1906.612561] env[62824]: DEBUG nova.network.neutron [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1906.759105] env[62824]: DEBUG oslo_vmware.api [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Task: {'id': task-2145723, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06836} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1906.760530] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] [instance: dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1906.760530] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48c91b9d-2d5e-4ef0-9151-3c50c7155402 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1906.782810] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] [instance: dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda] Reconfiguring VM instance instance-00000015 to attach disk [datastore2] dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda/dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1906.783843] env[62824]: DEBUG nova.network.neutron [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Updating instance_info_cache with network_info: [{"id": "02e90c28-7f36-4d83-84bb-fec194a04dab", "address": "fa:16:3e:53:1a:4c", "network": {"id": "b6403dc3-30ef-4c76-b521-380c179f345d", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.19", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "22c5c40c2e004babaa15896c89307303", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1002b79b-224e-41e3-a484-4245a767147a", "external-id": "nsx-vlan-transportzone-353", "segmentation_id": 353, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap02e90c28-7f", "ovs_interfaceid": "02e90c28-7f36-4d83-84bb-fec194a04dab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1906.789016] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-00975164-555d-4727-ae2b-776224e18a8a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1906.808957] env[62824]: DEBUG oslo_vmware.api [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Waiting for the task: (returnval){ [ 1906.808957] env[62824]: value = "task-2145724" [ 1906.808957] env[62824]: _type = "Task" [ 1906.808957] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1906.817555] env[62824]: DEBUG oslo_vmware.api [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Task: {'id': task-2145724, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1906.876963] env[62824]: DEBUG oslo_concurrency.lockutils [None req-95c8b711-718a-4a21-abec-a4473e9dbb40 tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Lock "9062e606-511f-4fe7-9621-90b3c0e51eec" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 4.499s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1906.892364] env[62824]: DEBUG nova.compute.manager [req-00adf689-aac8-41fd-a772-81447817cfc1 req-a54f2052-18e4-42c4-9d83-44c5be02f9aa service nova] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Received event network-vif-plugged-02e90c28-7f36-4d83-84bb-fec194a04dab {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1906.892622] env[62824]: DEBUG oslo_concurrency.lockutils [req-00adf689-aac8-41fd-a772-81447817cfc1 req-a54f2052-18e4-42c4-9d83-44c5be02f9aa service nova] Acquiring lock "e158b5d4-c120-4e6c-89c4-7668e097926d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1906.892829] env[62824]: DEBUG oslo_concurrency.lockutils [req-00adf689-aac8-41fd-a772-81447817cfc1 req-a54f2052-18e4-42c4-9d83-44c5be02f9aa service nova] Lock "e158b5d4-c120-4e6c-89c4-7668e097926d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1906.892997] env[62824]: DEBUG oslo_concurrency.lockutils [req-00adf689-aac8-41fd-a772-81447817cfc1 req-a54f2052-18e4-42c4-9d83-44c5be02f9aa service nova] Lock "e158b5d4-c120-4e6c-89c4-7668e097926d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1906.893174] env[62824]: DEBUG nova.compute.manager [req-00adf689-aac8-41fd-a772-81447817cfc1 req-a54f2052-18e4-42c4-9d83-44c5be02f9aa service nova] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] No waiting events found dispatching network-vif-plugged-02e90c28-7f36-4d83-84bb-fec194a04dab {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1906.893336] env[62824]: WARNING nova.compute.manager [req-00adf689-aac8-41fd-a772-81447817cfc1 req-a54f2052-18e4-42c4-9d83-44c5be02f9aa service nova] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Received unexpected event network-vif-plugged-02e90c28-7f36-4d83-84bb-fec194a04dab for instance with vm_state building and task_state spawning. [ 1906.893496] env[62824]: DEBUG nova.compute.manager [req-00adf689-aac8-41fd-a772-81447817cfc1 req-a54f2052-18e4-42c4-9d83-44c5be02f9aa service nova] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Received event network-changed-02e90c28-7f36-4d83-84bb-fec194a04dab {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1906.893647] env[62824]: DEBUG nova.compute.manager [req-00adf689-aac8-41fd-a772-81447817cfc1 req-a54f2052-18e4-42c4-9d83-44c5be02f9aa service nova] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Refreshing instance network info cache due to event network-changed-02e90c28-7f36-4d83-84bb-fec194a04dab. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1906.893805] env[62824]: DEBUG oslo_concurrency.lockutils [req-00adf689-aac8-41fd-a772-81447817cfc1 req-a54f2052-18e4-42c4-9d83-44c5be02f9aa service nova] Acquiring lock "refresh_cache-e158b5d4-c120-4e6c-89c4-7668e097926d" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1906.937417] env[62824]: DEBUG oslo_vmware.api [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5211270b-0cf8-4662-1cc7-39c6f176468b, 'name': SearchDatastore_Task, 'duration_secs': 0.021043} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1906.937945] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1906.938426] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 4dfe9f82-7f89-486b-b159-948656d2896c/4dfe9f82-7f89-486b-b159-948656d2896c.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1906.938899] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f4a27b49-1cec-4ce6-8fc5-a8b53d34542c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1906.945671] env[62824]: DEBUG oslo_vmware.api [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Waiting for the task: (returnval){ [ 1906.945671] env[62824]: value = "task-2145725" [ 1906.945671] env[62824]: _type = "Task" [ 1906.945671] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1906.957308] env[62824]: DEBUG oslo_vmware.api [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Task: {'id': task-2145725, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1907.103056] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.580s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1907.103607] env[62824]: DEBUG nova.compute.manager [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1907.107008] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.878s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1907.108532] env[62824]: INFO nova.compute.claims [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: f3740d9a-b21a-4a79-9e28-2a89ecd08bb6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1907.289019] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Releasing lock "refresh_cache-e158b5d4-c120-4e6c-89c4-7668e097926d" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1907.289019] env[62824]: DEBUG nova.compute.manager [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Instance network_info: |[{"id": "02e90c28-7f36-4d83-84bb-fec194a04dab", "address": "fa:16:3e:53:1a:4c", "network": {"id": "b6403dc3-30ef-4c76-b521-380c179f345d", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.19", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "22c5c40c2e004babaa15896c89307303", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1002b79b-224e-41e3-a484-4245a767147a", "external-id": "nsx-vlan-transportzone-353", "segmentation_id": 353, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap02e90c28-7f", "ovs_interfaceid": "02e90c28-7f36-4d83-84bb-fec194a04dab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1907.289451] env[62824]: DEBUG oslo_concurrency.lockutils [req-00adf689-aac8-41fd-a772-81447817cfc1 req-a54f2052-18e4-42c4-9d83-44c5be02f9aa service nova] Acquired lock "refresh_cache-e158b5d4-c120-4e6c-89c4-7668e097926d" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1907.289451] env[62824]: DEBUG nova.network.neutron [req-00adf689-aac8-41fd-a772-81447817cfc1 req-a54f2052-18e4-42c4-9d83-44c5be02f9aa service nova] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Refreshing network info cache for port 02e90c28-7f36-4d83-84bb-fec194a04dab {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1907.289451] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:53:1a:4c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1002b79b-224e-41e3-a484-4245a767147a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '02e90c28-7f36-4d83-84bb-fec194a04dab', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1907.300076] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Creating folder: Project (b2f9084931bc4b94b0ae85e3e4b0ba90). Parent ref: group-v438503. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1907.301027] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-515b9939-7ec4-4fe6-9ee2-dfd581f94ea8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1907.316128] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Created folder: Project (b2f9084931bc4b94b0ae85e3e4b0ba90) in parent group-v438503. [ 1907.316128] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Creating folder: Instances. Parent ref: group-v438541. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1907.316372] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d4ba8c3c-16e8-4a1e-a545-490b72271296 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1907.322323] env[62824]: DEBUG oslo_vmware.api [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Task: {'id': task-2145724, 'name': ReconfigVM_Task, 'duration_secs': 0.298927} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1907.323333] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] [instance: dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda] Reconfigured VM instance instance-00000015 to attach disk [datastore2] dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda/dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1907.324130] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4c16624a-a2c3-412b-a752-0b0be4257f77 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1907.330511] env[62824]: DEBUG oslo_vmware.api [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Waiting for the task: (returnval){ [ 1907.330511] env[62824]: value = "task-2145728" [ 1907.330511] env[62824]: _type = "Task" [ 1907.330511] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1907.331958] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Created folder: Instances in parent group-v438541. [ 1907.332295] env[62824]: DEBUG oslo.service.loopingcall [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1907.335732] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1907.336082] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ed67b6dd-1df2-43ca-9c9a-874a5c016e20 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1907.361989] env[62824]: DEBUG oslo_vmware.api [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Task: {'id': task-2145728, 'name': Rename_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1907.362184] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1907.362184] env[62824]: value = "task-2145729" [ 1907.362184] env[62824]: _type = "Task" [ 1907.362184] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1907.371545] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145729, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1907.461832] env[62824]: DEBUG oslo_vmware.api [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Task: {'id': task-2145725, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1907.498309] env[62824]: DEBUG nova.virt.hardware [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1907.498610] env[62824]: DEBUG nova.virt.hardware [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1907.498781] env[62824]: DEBUG nova.virt.hardware [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1907.499014] env[62824]: DEBUG nova.virt.hardware [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1907.499218] env[62824]: DEBUG nova.virt.hardware [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1907.499416] env[62824]: DEBUG nova.virt.hardware [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1907.499721] env[62824]: DEBUG nova.virt.hardware [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1907.499840] env[62824]: DEBUG nova.virt.hardware [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1907.500098] env[62824]: DEBUG nova.virt.hardware [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1907.500356] env[62824]: DEBUG nova.virt.hardware [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1907.500600] env[62824]: DEBUG nova.virt.hardware [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1907.502207] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63284ec2-8875-4d59-bdb5-d8846c231116 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1907.511385] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce227149-edaa-4a07-847e-17ef82d4fc0b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1907.528881] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Instance VIF info [] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1907.535116] env[62824]: DEBUG oslo.service.loopingcall [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1907.535612] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1a251766-5a96-4ca3-afcd-ba6cbf0f0b78 tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Acquiring lock "9062e606-511f-4fe7-9621-90b3c0e51eec" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1907.535847] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1a251766-5a96-4ca3-afcd-ba6cbf0f0b78 tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Lock "9062e606-511f-4fe7-9621-90b3c0e51eec" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1907.536115] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1a251766-5a96-4ca3-afcd-ba6cbf0f0b78 tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Acquiring lock "9062e606-511f-4fe7-9621-90b3c0e51eec-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1907.536309] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1a251766-5a96-4ca3-afcd-ba6cbf0f0b78 tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Lock "9062e606-511f-4fe7-9621-90b3c0e51eec-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1907.536511] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1a251766-5a96-4ca3-afcd-ba6cbf0f0b78 tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Lock "9062e606-511f-4fe7-9621-90b3c0e51eec-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1907.538330] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1907.538867] env[62824]: INFO nova.compute.manager [None req-1a251766-5a96-4ca3-afcd-ba6cbf0f0b78 tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] [instance: 9062e606-511f-4fe7-9621-90b3c0e51eec] Terminating instance [ 1907.540537] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-748ec21f-9c18-4870-bccd-868d905e93c3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1907.561895] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1907.561895] env[62824]: value = "task-2145730" [ 1907.561895] env[62824]: _type = "Task" [ 1907.561895] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1907.571439] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145730, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1907.609707] env[62824]: DEBUG nova.compute.utils [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1907.615101] env[62824]: DEBUG nova.compute.manager [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1907.615591] env[62824]: DEBUG nova.network.neutron [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1907.679234] env[62824]: DEBUG nova.policy [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a6a63328f3534c7c912384254ccda1f0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '248af2bc3e7b41d2a74b0c6074b359e1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 1907.842499] env[62824]: DEBUG oslo_vmware.api [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Task: {'id': task-2145728, 'name': Rename_Task, 'duration_secs': 0.261371} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1907.843043] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] [instance: dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1907.843725] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-20ee342c-e335-477d-90c7-4a4144515583 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1907.853021] env[62824]: DEBUG oslo_vmware.api [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Waiting for the task: (returnval){ [ 1907.853021] env[62824]: value = "task-2145731" [ 1907.853021] env[62824]: _type = "Task" [ 1907.853021] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1907.863382] env[62824]: DEBUG oslo_vmware.api [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Task: {'id': task-2145731, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1907.872678] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145729, 'name': CreateVM_Task, 'duration_secs': 0.458752} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1907.873107] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1907.874132] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1907.874367] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1907.874906] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1907.875212] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7c43996d-eb9d-459a-8d7b-f1b6cef91b17 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1907.881703] env[62824]: DEBUG oslo_vmware.api [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Waiting for the task: (returnval){ [ 1907.881703] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52a5f945-7623-e582-8b0a-c0caf56edd26" [ 1907.881703] env[62824]: _type = "Task" [ 1907.881703] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1907.892862] env[62824]: DEBUG oslo_vmware.api [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52a5f945-7623-e582-8b0a-c0caf56edd26, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1907.958143] env[62824]: DEBUG oslo_vmware.api [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Task: {'id': task-2145725, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.735265} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1907.958524] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 4dfe9f82-7f89-486b-b159-948656d2896c/4dfe9f82-7f89-486b-b159-948656d2896c.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1907.958840] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] [instance: 4dfe9f82-7f89-486b-b159-948656d2896c] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1907.959172] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-28974444-c76d-4dc4-af6d-37ab2ca5228a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1907.967462] env[62824]: DEBUG oslo_vmware.api [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Waiting for the task: (returnval){ [ 1907.967462] env[62824]: value = "task-2145732" [ 1907.967462] env[62824]: _type = "Task" [ 1907.967462] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1907.975915] env[62824]: DEBUG oslo_vmware.api [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Task: {'id': task-2145732, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1907.990183] env[62824]: DEBUG nova.network.neutron [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Successfully created port: 309b4346-1935-41f1-a374-d322c4a8c016 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1908.061021] env[62824]: DEBUG nova.compute.manager [None req-1a251766-5a96-4ca3-afcd-ba6cbf0f0b78 tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] [instance: 9062e606-511f-4fe7-9621-90b3c0e51eec] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1908.061021] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-1a251766-5a96-4ca3-afcd-ba6cbf0f0b78 tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] [instance: 9062e606-511f-4fe7-9621-90b3c0e51eec] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1908.061021] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-612cf607-2451-441f-a00d-f6d20646a49d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1908.073365] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145730, 'name': CreateVM_Task, 'duration_secs': 0.387827} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1908.075670] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1908.076163] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a251766-5a96-4ca3-afcd-ba6cbf0f0b78 tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] [instance: 9062e606-511f-4fe7-9621-90b3c0e51eec] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1908.076671] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1908.077027] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e15f001e-a99b-4075-af83-64bf38a36bb5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1908.082957] env[62824]: DEBUG oslo_vmware.api [None req-1a251766-5a96-4ca3-afcd-ba6cbf0f0b78 tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Waiting for the task: (returnval){ [ 1908.082957] env[62824]: value = "task-2145733" [ 1908.082957] env[62824]: _type = "Task" [ 1908.082957] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1908.094386] env[62824]: DEBUG oslo_vmware.api [None req-1a251766-5a96-4ca3-afcd-ba6cbf0f0b78 tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Task: {'id': task-2145733, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1908.121827] env[62824]: DEBUG nova.compute.manager [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1908.362121] env[62824]: DEBUG oslo_vmware.api [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Task: {'id': task-2145731, 'name': PowerOnVM_Task} progress is 37%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1908.391781] env[62824]: DEBUG oslo_vmware.api [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52a5f945-7623-e582-8b0a-c0caf56edd26, 'name': SearchDatastore_Task, 'duration_secs': 0.022094} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1908.394378] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1908.394654] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1908.394895] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1908.395054] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1908.395239] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1908.395758] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1908.396078] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1908.396302] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6d79a930-6470-43cf-8b37-e6970ac64835 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1908.398062] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e84cfaad-121d-405a-9dfd-8900dd8d89d5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1908.403082] env[62824]: DEBUG oslo_vmware.api [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Waiting for the task: (returnval){ [ 1908.403082] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5221ddb9-49c1-0fb5-4670-e45165b3e154" [ 1908.403082] env[62824]: _type = "Task" [ 1908.403082] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1908.408529] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1908.408713] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1908.414465] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0a3239b2-251d-41f2-b109-9744b1347f28 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1908.417138] env[62824]: DEBUG oslo_vmware.api [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5221ddb9-49c1-0fb5-4670-e45165b3e154, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1908.420355] env[62824]: DEBUG oslo_vmware.api [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Waiting for the task: (returnval){ [ 1908.420355] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52fe06ed-546c-9e2a-86e2-c2f1b409c8b0" [ 1908.420355] env[62824]: _type = "Task" [ 1908.420355] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1908.428771] env[62824]: DEBUG oslo_vmware.api [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52fe06ed-546c-9e2a-86e2-c2f1b409c8b0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1908.479943] env[62824]: DEBUG oslo_vmware.api [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Task: {'id': task-2145732, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078993} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1908.481089] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] [instance: 4dfe9f82-7f89-486b-b159-948656d2896c] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1908.481506] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c565cd89-d30a-43d9-aa26-cbd907710219 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1908.506290] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] [instance: 4dfe9f82-7f89-486b-b159-948656d2896c] Reconfiguring VM instance instance-00000016 to attach disk [datastore2] 4dfe9f82-7f89-486b-b159-948656d2896c/4dfe9f82-7f89-486b-b159-948656d2896c.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1908.509304] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4be269d3-bda3-4e41-968f-8821156ca1f0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1908.532506] env[62824]: DEBUG oslo_vmware.api [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Waiting for the task: (returnval){ [ 1908.532506] env[62824]: value = "task-2145734" [ 1908.532506] env[62824]: _type = "Task" [ 1908.532506] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1908.538999] env[62824]: DEBUG nova.network.neutron [req-00adf689-aac8-41fd-a772-81447817cfc1 req-a54f2052-18e4-42c4-9d83-44c5be02f9aa service nova] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Updated VIF entry in instance network info cache for port 02e90c28-7f36-4d83-84bb-fec194a04dab. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1908.539415] env[62824]: DEBUG nova.network.neutron [req-00adf689-aac8-41fd-a772-81447817cfc1 req-a54f2052-18e4-42c4-9d83-44c5be02f9aa service nova] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Updating instance_info_cache with network_info: [{"id": "02e90c28-7f36-4d83-84bb-fec194a04dab", "address": "fa:16:3e:53:1a:4c", "network": {"id": "b6403dc3-30ef-4c76-b521-380c179f345d", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.19", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "22c5c40c2e004babaa15896c89307303", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1002b79b-224e-41e3-a484-4245a767147a", "external-id": "nsx-vlan-transportzone-353", "segmentation_id": 353, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap02e90c28-7f", "ovs_interfaceid": "02e90c28-7f36-4d83-84bb-fec194a04dab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1908.547612] env[62824]: DEBUG oslo_vmware.api [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Task: {'id': task-2145734, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1908.597014] env[62824]: DEBUG oslo_vmware.api [None req-1a251766-5a96-4ca3-afcd-ba6cbf0f0b78 tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Task: {'id': task-2145733, 'name': PowerOffVM_Task, 'duration_secs': 0.180694} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1908.597433] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a251766-5a96-4ca3-afcd-ba6cbf0f0b78 tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] [instance: 9062e606-511f-4fe7-9621-90b3c0e51eec] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1908.597482] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-1a251766-5a96-4ca3-afcd-ba6cbf0f0b78 tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] [instance: 9062e606-511f-4fe7-9621-90b3c0e51eec] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1908.597744] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c3364764-d470-41e3-9742-1bc34a5385b7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1908.711744] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-1a251766-5a96-4ca3-afcd-ba6cbf0f0b78 tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] [instance: 9062e606-511f-4fe7-9621-90b3c0e51eec] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1908.712034] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-1a251766-5a96-4ca3-afcd-ba6cbf0f0b78 tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] [instance: 9062e606-511f-4fe7-9621-90b3c0e51eec] Deleting contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1908.712224] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-1a251766-5a96-4ca3-afcd-ba6cbf0f0b78 tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Deleting the datastore file [datastore1] 9062e606-511f-4fe7-9621-90b3c0e51eec {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1908.715180] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-241608c7-2c37-46a1-aa8b-3a118fd06686 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1908.726185] env[62824]: DEBUG oslo_vmware.api [None req-1a251766-5a96-4ca3-afcd-ba6cbf0f0b78 tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Waiting for the task: (returnval){ [ 1908.726185] env[62824]: value = "task-2145736" [ 1908.726185] env[62824]: _type = "Task" [ 1908.726185] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1908.733626] env[62824]: DEBUG oslo_vmware.api [None req-1a251766-5a96-4ca3-afcd-ba6cbf0f0b78 tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Task: {'id': task-2145736, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1908.738743] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32dbe513-8b18-419b-a8ca-6e078f444719 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1908.746747] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad1aa50d-c9f6-475f-99d7-4c3d6d4563cc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1908.782283] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f21aa31-02a0-4bae-904d-34263ed31d7c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1908.792109] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-104fd622-73b7-4f0b-9cdb-dbe1644b75bc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1908.809787] env[62824]: DEBUG nova.compute.provider_tree [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1908.869238] env[62824]: DEBUG oslo_vmware.api [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Task: {'id': task-2145731, 'name': PowerOnVM_Task, 'duration_secs': 0.962684} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1908.869511] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] [instance: dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1908.869713] env[62824]: INFO nova.compute.manager [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] [instance: dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda] Took 9.83 seconds to spawn the instance on the hypervisor. [ 1908.869892] env[62824]: DEBUG nova.compute.manager [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] [instance: dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1908.870676] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-553bd0a7-fd06-4d21-965a-f8a3e49591f7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1908.913512] env[62824]: DEBUG oslo_vmware.api [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5221ddb9-49c1-0fb5-4670-e45165b3e154, 'name': SearchDatastore_Task, 'duration_secs': 0.013155} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1908.913828] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1908.914090] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1908.914316] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1908.932026] env[62824]: DEBUG oslo_vmware.api [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52fe06ed-546c-9e2a-86e2-c2f1b409c8b0, 'name': SearchDatastore_Task, 'duration_secs': 0.008973} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1908.932439] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f8247aaa-e1a1-48bb-ab7a-f5ec1b09c07d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1908.937711] env[62824]: DEBUG oslo_vmware.api [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Waiting for the task: (returnval){ [ 1908.937711] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5294a00b-f609-b9e1-f871-7b738e35d88e" [ 1908.937711] env[62824]: _type = "Task" [ 1908.937711] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1908.945766] env[62824]: DEBUG oslo_vmware.api [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5294a00b-f609-b9e1-f871-7b738e35d88e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1909.043314] env[62824]: DEBUG oslo_concurrency.lockutils [req-00adf689-aac8-41fd-a772-81447817cfc1 req-a54f2052-18e4-42c4-9d83-44c5be02f9aa service nova] Releasing lock "refresh_cache-e158b5d4-c120-4e6c-89c4-7668e097926d" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1909.043741] env[62824]: DEBUG oslo_vmware.api [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Task: {'id': task-2145734, 'name': ReconfigVM_Task, 'duration_secs': 0.302034} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1909.044032] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] [instance: 4dfe9f82-7f89-486b-b159-948656d2896c] Reconfigured VM instance instance-00000016 to attach disk [datastore2] 4dfe9f82-7f89-486b-b159-948656d2896c/4dfe9f82-7f89-486b-b159-948656d2896c.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1909.044720] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6cd4173b-5a0c-4175-a478-21d8c2594b16 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1909.051926] env[62824]: DEBUG oslo_vmware.api [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Waiting for the task: (returnval){ [ 1909.051926] env[62824]: value = "task-2145737" [ 1909.051926] env[62824]: _type = "Task" [ 1909.051926] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1909.060552] env[62824]: DEBUG oslo_vmware.api [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Task: {'id': task-2145737, 'name': Rename_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1909.136583] env[62824]: DEBUG nova.compute.manager [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1909.163933] env[62824]: DEBUG nova.virt.hardware [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1909.164236] env[62824]: DEBUG nova.virt.hardware [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1909.164518] env[62824]: DEBUG nova.virt.hardware [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1909.164723] env[62824]: DEBUG nova.virt.hardware [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1909.164875] env[62824]: DEBUG nova.virt.hardware [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1909.165036] env[62824]: DEBUG nova.virt.hardware [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1909.165255] env[62824]: DEBUG nova.virt.hardware [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1909.165416] env[62824]: DEBUG nova.virt.hardware [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1909.165594] env[62824]: DEBUG nova.virt.hardware [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1909.165797] env[62824]: DEBUG nova.virt.hardware [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1909.165991] env[62824]: DEBUG nova.virt.hardware [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1909.166887] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-802362e5-14eb-4f2e-8f83-96d3e59fa21d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1909.174796] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcc4f1e6-995c-43f3-8ee4-617b704cb9f3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1909.232180] env[62824]: DEBUG oslo_vmware.api [None req-1a251766-5a96-4ca3-afcd-ba6cbf0f0b78 tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Task: {'id': task-2145736, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.26907} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1909.232448] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-1a251766-5a96-4ca3-afcd-ba6cbf0f0b78 tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1909.232621] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-1a251766-5a96-4ca3-afcd-ba6cbf0f0b78 tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] [instance: 9062e606-511f-4fe7-9621-90b3c0e51eec] Deleted contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1909.232800] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-1a251766-5a96-4ca3-afcd-ba6cbf0f0b78 tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] [instance: 9062e606-511f-4fe7-9621-90b3c0e51eec] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1909.232978] env[62824]: INFO nova.compute.manager [None req-1a251766-5a96-4ca3-afcd-ba6cbf0f0b78 tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] [instance: 9062e606-511f-4fe7-9621-90b3c0e51eec] Took 1.17 seconds to destroy the instance on the hypervisor. [ 1909.233258] env[62824]: DEBUG oslo.service.loopingcall [None req-1a251766-5a96-4ca3-afcd-ba6cbf0f0b78 tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1909.233464] env[62824]: DEBUG nova.compute.manager [-] [instance: 9062e606-511f-4fe7-9621-90b3c0e51eec] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1909.233562] env[62824]: DEBUG nova.network.neutron [-] [instance: 9062e606-511f-4fe7-9621-90b3c0e51eec] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1909.312873] env[62824]: DEBUG nova.scheduler.client.report [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1909.391802] env[62824]: INFO nova.compute.manager [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] [instance: dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda] Took 27.82 seconds to build instance. [ 1909.449760] env[62824]: DEBUG oslo_vmware.api [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5294a00b-f609-b9e1-f871-7b738e35d88e, 'name': SearchDatastore_Task, 'duration_secs': 0.009365} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1909.449760] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1909.449760] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] e158b5d4-c120-4e6c-89c4-7668e097926d/e158b5d4-c120-4e6c-89c4-7668e097926d.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1909.449760] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1909.450014] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1909.450014] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-77db680f-1d0c-4c6d-843a-fa855d739925 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1909.451796] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7cbfc94d-33a6-4778-bdd4-7e95d4b5320a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1909.458865] env[62824]: DEBUG oslo_vmware.api [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Waiting for the task: (returnval){ [ 1909.458865] env[62824]: value = "task-2145738" [ 1909.458865] env[62824]: _type = "Task" [ 1909.458865] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1909.463296] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1909.463296] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1909.463813] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-26a31ea2-24f0-4820-97be-558b464d00f1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1909.470692] env[62824]: DEBUG oslo_vmware.api [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Task: {'id': task-2145738, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1909.471720] env[62824]: DEBUG nova.compute.manager [req-4170b03c-44e4-47ec-b054-1f3b1782c065 req-ac6c4db5-66b2-462e-9a98-2848f9bd0d85 service nova] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Received event network-vif-plugged-309b4346-1935-41f1-a374-d322c4a8c016 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1909.471919] env[62824]: DEBUG oslo_concurrency.lockutils [req-4170b03c-44e4-47ec-b054-1f3b1782c065 req-ac6c4db5-66b2-462e-9a98-2848f9bd0d85 service nova] Acquiring lock "07ce931f-45ef-409b-b714-9f1cd47a3a88-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1909.472142] env[62824]: DEBUG oslo_concurrency.lockutils [req-4170b03c-44e4-47ec-b054-1f3b1782c065 req-ac6c4db5-66b2-462e-9a98-2848f9bd0d85 service nova] Lock "07ce931f-45ef-409b-b714-9f1cd47a3a88-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1909.472311] env[62824]: DEBUG oslo_concurrency.lockutils [req-4170b03c-44e4-47ec-b054-1f3b1782c065 req-ac6c4db5-66b2-462e-9a98-2848f9bd0d85 service nova] Lock "07ce931f-45ef-409b-b714-9f1cd47a3a88-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1909.472526] env[62824]: DEBUG nova.compute.manager [req-4170b03c-44e4-47ec-b054-1f3b1782c065 req-ac6c4db5-66b2-462e-9a98-2848f9bd0d85 service nova] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] No waiting events found dispatching network-vif-plugged-309b4346-1935-41f1-a374-d322c4a8c016 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1909.472690] env[62824]: WARNING nova.compute.manager [req-4170b03c-44e4-47ec-b054-1f3b1782c065 req-ac6c4db5-66b2-462e-9a98-2848f9bd0d85 service nova] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Received unexpected event network-vif-plugged-309b4346-1935-41f1-a374-d322c4a8c016 for instance with vm_state building and task_state spawning. [ 1909.476285] env[62824]: DEBUG oslo_vmware.api [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Waiting for the task: (returnval){ [ 1909.476285] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52b6edda-c546-a941-78fc-bd42f61cb96c" [ 1909.476285] env[62824]: _type = "Task" [ 1909.476285] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1909.487216] env[62824]: DEBUG oslo_vmware.api [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52b6edda-c546-a941-78fc-bd42f61cb96c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1909.565236] env[62824]: DEBUG oslo_vmware.api [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Task: {'id': task-2145737, 'name': Rename_Task, 'duration_secs': 0.147502} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1909.566180] env[62824]: DEBUG nova.network.neutron [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Successfully updated port: 309b4346-1935-41f1-a374-d322c4a8c016 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1909.567463] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] [instance: 4dfe9f82-7f89-486b-b159-948656d2896c] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1909.567736] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a214c690-1e77-4d80-83b3-ae54a496116b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1909.576510] env[62824]: DEBUG oslo_vmware.api [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Waiting for the task: (returnval){ [ 1909.576510] env[62824]: value = "task-2145739" [ 1909.576510] env[62824]: _type = "Task" [ 1909.576510] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1909.586329] env[62824]: DEBUG oslo_vmware.api [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Task: {'id': task-2145739, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1909.724745] env[62824]: DEBUG nova.compute.manager [req-5fe8f18d-ecc5-4f54-a8cc-f6a0e12d3ff5 req-b35d0867-27e3-4f9b-8b8f-157873589b67 service nova] [instance: 9062e606-511f-4fe7-9621-90b3c0e51eec] Received event network-vif-deleted-3c79003f-579c-4415-8515-cee7f296a297 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1909.724745] env[62824]: INFO nova.compute.manager [req-5fe8f18d-ecc5-4f54-a8cc-f6a0e12d3ff5 req-b35d0867-27e3-4f9b-8b8f-157873589b67 service nova] [instance: 9062e606-511f-4fe7-9621-90b3c0e51eec] Neutron deleted interface 3c79003f-579c-4415-8515-cee7f296a297; detaching it from the instance and deleting it from the info cache [ 1909.728993] env[62824]: DEBUG nova.network.neutron [req-5fe8f18d-ecc5-4f54-a8cc-f6a0e12d3ff5 req-b35d0867-27e3-4f9b-8b8f-157873589b67 service nova] [instance: 9062e606-511f-4fe7-9621-90b3c0e51eec] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1909.820089] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.712s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1909.820774] env[62824]: DEBUG nova.compute.manager [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: f3740d9a-b21a-4a79-9e28-2a89ecd08bb6] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1909.826048] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 19.548s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1909.826048] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1909.826048] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62824) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1909.826372] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.605s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1909.829830] env[62824]: INFO nova.compute.claims [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] [instance: ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1909.837341] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-642a5c0c-e5e5-4c79-82ac-a4ce88b51d6e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1909.848258] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92eb6ade-cfa0-4455-af37-59ce637a69e5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1909.869921] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-290d133a-be13-41ba-adda-08ad79bc86ff {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1909.878154] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f555c81a-147d-49aa-b0ed-515ab6e20b1f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1909.916139] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d1ff4193-6632-43f7-bfcf-085fd71be835 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Lock "dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 58.983s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1909.916139] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181116MB free_disk=175GB free_vcpus=48 pci_devices=None {{(pid=62824) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1909.916384] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1909.969794] env[62824]: DEBUG oslo_vmware.api [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Task: {'id': task-2145738, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.469605} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1909.970099] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] e158b5d4-c120-4e6c-89c4-7668e097926d/e158b5d4-c120-4e6c-89c4-7668e097926d.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1909.970318] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1909.970567] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0bb45a56-2308-4c06-a6b9-e1846a44e112 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1909.978211] env[62824]: DEBUG oslo_vmware.api [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Waiting for the task: (returnval){ [ 1909.978211] env[62824]: value = "task-2145740" [ 1909.978211] env[62824]: _type = "Task" [ 1909.978211] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1909.992770] env[62824]: DEBUG oslo_vmware.api [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52b6edda-c546-a941-78fc-bd42f61cb96c, 'name': SearchDatastore_Task, 'duration_secs': 0.007944} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1909.996666] env[62824]: DEBUG oslo_vmware.api [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Task: {'id': task-2145740, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1909.997308] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a618b4b6-404f-471a-8959-99ebc0828e9c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1910.003339] env[62824]: DEBUG oslo_vmware.api [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Waiting for the task: (returnval){ [ 1910.003339] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52db6917-a8d2-9b36-bd9a-3b20d79b7303" [ 1910.003339] env[62824]: _type = "Task" [ 1910.003339] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1910.012252] env[62824]: DEBUG oslo_vmware.api [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52db6917-a8d2-9b36-bd9a-3b20d79b7303, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1910.077890] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Acquiring lock "refresh_cache-07ce931f-45ef-409b-b714-9f1cd47a3a88" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1910.077890] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Acquired lock "refresh_cache-07ce931f-45ef-409b-b714-9f1cd47a3a88" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1910.077890] env[62824]: DEBUG nova.network.neutron [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1910.087044] env[62824]: DEBUG oslo_vmware.api [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Task: {'id': task-2145739, 'name': PowerOnVM_Task} progress is 1%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1910.132519] env[62824]: DEBUG nova.network.neutron [-] [instance: 9062e606-511f-4fe7-9621-90b3c0e51eec] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1910.232179] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cfb6ee02-8d0d-4568-b391-44643b3f0bdd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1910.241278] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec9929c2-6a40-4b99-ae2b-b268f4af56b4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1910.272589] env[62824]: DEBUG nova.compute.manager [req-5fe8f18d-ecc5-4f54-a8cc-f6a0e12d3ff5 req-b35d0867-27e3-4f9b-8b8f-157873589b67 service nova] [instance: 9062e606-511f-4fe7-9621-90b3c0e51eec] Detach interface failed, port_id=3c79003f-579c-4415-8515-cee7f296a297, reason: Instance 9062e606-511f-4fe7-9621-90b3c0e51eec could not be found. {{(pid=62824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11482}} [ 1910.337510] env[62824]: DEBUG nova.compute.utils [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1910.338919] env[62824]: DEBUG nova.compute.manager [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: f3740d9a-b21a-4a79-9e28-2a89ecd08bb6] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1910.340070] env[62824]: DEBUG nova.network.neutron [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: f3740d9a-b21a-4a79-9e28-2a89ecd08bb6] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1910.419394] env[62824]: DEBUG nova.compute.manager [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] [instance: b96d1351-f3a7-4bac-998c-a34ab2606041] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1910.489393] env[62824]: DEBUG nova.policy [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a6a63328f3534c7c912384254ccda1f0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '248af2bc3e7b41d2a74b0c6074b359e1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 1910.494720] env[62824]: DEBUG oslo_vmware.api [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Task: {'id': task-2145740, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.08825} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1910.494994] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1910.496286] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d9b9e19-ee64-4315-a739-c7c474419dcd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1910.525500] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Reconfiguring VM instance instance-00000017 to attach disk [datastore2] e158b5d4-c120-4e6c-89c4-7668e097926d/e158b5d4-c120-4e6c-89c4-7668e097926d.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1910.531086] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bf4fd59a-758b-4eec-a1f6-d6e37ec3e17d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1910.553843] env[62824]: DEBUG oslo_vmware.api [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52db6917-a8d2-9b36-bd9a-3b20d79b7303, 'name': SearchDatastore_Task, 'duration_secs': 0.009522} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1910.555489] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1910.555782] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 384a96ea-40ce-43e1-a5f9-82f50b710b0b/384a96ea-40ce-43e1-a5f9-82f50b710b0b.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1910.556134] env[62824]: DEBUG oslo_vmware.api [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Waiting for the task: (returnval){ [ 1910.556134] env[62824]: value = "task-2145741" [ 1910.556134] env[62824]: _type = "Task" [ 1910.556134] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1910.556335] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6825289c-e2ad-4545-8e90-b98c4b2b0621 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1910.567506] env[62824]: DEBUG oslo_vmware.api [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Task: {'id': task-2145741, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1910.569197] env[62824]: DEBUG oslo_vmware.api [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Waiting for the task: (returnval){ [ 1910.569197] env[62824]: value = "task-2145742" [ 1910.569197] env[62824]: _type = "Task" [ 1910.569197] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1910.577662] env[62824]: DEBUG oslo_vmware.api [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Task: {'id': task-2145742, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1910.589081] env[62824]: DEBUG oslo_vmware.api [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Task: {'id': task-2145739, 'name': PowerOnVM_Task} progress is 82%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1910.635136] env[62824]: INFO nova.compute.manager [-] [instance: 9062e606-511f-4fe7-9621-90b3c0e51eec] Took 1.40 seconds to deallocate network for instance. [ 1910.676250] env[62824]: DEBUG nova.network.neutron [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1910.844972] env[62824]: DEBUG nova.compute.manager [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: f3740d9a-b21a-4a79-9e28-2a89ecd08bb6] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1910.942512] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1910.958532] env[62824]: DEBUG nova.network.neutron [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Updating instance_info_cache with network_info: [{"id": "309b4346-1935-41f1-a374-d322c4a8c016", "address": "fa:16:3e:07:e9:a0", "network": {"id": "fc381ca5-198f-4afc-9ebf-bc6e3de68c35", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-425455451-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "248af2bc3e7b41d2a74b0c6074b359e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4c9a12d2-469f-4199-bfaa-f791d765deac", "external-id": "nsx-vlan-transportzone-96", "segmentation_id": 96, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap309b4346-19", "ovs_interfaceid": "309b4346-1935-41f1-a374-d322c4a8c016", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1911.071396] env[62824]: DEBUG oslo_vmware.api [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Task: {'id': task-2145741, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1911.088960] env[62824]: DEBUG oslo_vmware.api [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Task: {'id': task-2145742, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1911.097174] env[62824]: DEBUG oslo_vmware.api [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Task: {'id': task-2145739, 'name': PowerOnVM_Task, 'duration_secs': 1.322211} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1911.098038] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] [instance: 4dfe9f82-7f89-486b-b159-948656d2896c] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1911.098323] env[62824]: INFO nova.compute.manager [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] [instance: 4dfe9f82-7f89-486b-b159-948656d2896c] Took 9.30 seconds to spawn the instance on the hypervisor. [ 1911.098695] env[62824]: DEBUG nova.compute.manager [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] [instance: 4dfe9f82-7f89-486b-b159-948656d2896c] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1911.100272] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d56fcda-88b5-4614-b235-7bad0f5b1188 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1911.144153] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1a251766-5a96-4ca3-afcd-ba6cbf0f0b78 tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1911.223121] env[62824]: DEBUG nova.network.neutron [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: f3740d9a-b21a-4a79-9e28-2a89ecd08bb6] Successfully created port: f2fb594d-5cb0-4c90-9787-3a362864f81a {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1911.433895] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1dd59d99-7ed2-416a-903a-89329cf35131 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1911.442142] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3b0a11c-624e-4a01-b7b9-fa735b87a768 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1911.477244] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Releasing lock "refresh_cache-07ce931f-45ef-409b-b714-9f1cd47a3a88" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1911.477244] env[62824]: DEBUG nova.compute.manager [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Instance network_info: |[{"id": "309b4346-1935-41f1-a374-d322c4a8c016", "address": "fa:16:3e:07:e9:a0", "network": {"id": "fc381ca5-198f-4afc-9ebf-bc6e3de68c35", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-425455451-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "248af2bc3e7b41d2a74b0c6074b359e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4c9a12d2-469f-4199-bfaa-f791d765deac", "external-id": "nsx-vlan-transportzone-96", "segmentation_id": 96, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap309b4346-19", "ovs_interfaceid": "309b4346-1935-41f1-a374-d322c4a8c016", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1911.478082] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:07:e9:a0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4c9a12d2-469f-4199-bfaa-f791d765deac', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '309b4346-1935-41f1-a374-d322c4a8c016', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1911.486599] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Creating folder: Project (248af2bc3e7b41d2a74b0c6074b359e1). Parent ref: group-v438503. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1911.487508] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce437a00-2e06-49b1-9425-8885d2f0565f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1911.490402] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d096d8ef-b18c-416f-be64-bdc6aca031ea {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1911.497430] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85f0b978-cebb-462f-9706-2ceb58dc75c3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1911.502257] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Created folder: Project (248af2bc3e7b41d2a74b0c6074b359e1) in parent group-v438503. [ 1911.502459] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Creating folder: Instances. Parent ref: group-v438545. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1911.502999] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bdeb3001-4a2c-447a-a2a4-c554714c2336 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1911.512599] env[62824]: DEBUG nova.compute.provider_tree [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1911.520906] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Created folder: Instances in parent group-v438545. [ 1911.520906] env[62824]: DEBUG oslo.service.loopingcall [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1911.520906] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1911.520906] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a74e65ea-01f4-49b5-b527-7d2bebb1809a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1911.541701] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1911.541701] env[62824]: value = "task-2145745" [ 1911.541701] env[62824]: _type = "Task" [ 1911.541701] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1911.550610] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145745, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1911.570051] env[62824]: DEBUG oslo_vmware.api [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Task: {'id': task-2145741, 'name': ReconfigVM_Task, 'duration_secs': 0.731499} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1911.570338] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Reconfigured VM instance instance-00000017 to attach disk [datastore2] e158b5d4-c120-4e6c-89c4-7668e097926d/e158b5d4-c120-4e6c-89c4-7668e097926d.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1911.571111] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ce9df94b-4654-4fd4-b7d8-743eedaccb53 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1911.582024] env[62824]: DEBUG oslo_vmware.api [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Task: {'id': task-2145742, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.757925} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1911.583453] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 384a96ea-40ce-43e1-a5f9-82f50b710b0b/384a96ea-40ce-43e1-a5f9-82f50b710b0b.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1911.583665] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1911.584147] env[62824]: DEBUG oslo_vmware.api [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Waiting for the task: (returnval){ [ 1911.584147] env[62824]: value = "task-2145746" [ 1911.584147] env[62824]: _type = "Task" [ 1911.584147] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1911.584358] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fe3c1dd5-d57d-489d-8323-c57594b20f55 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1911.598678] env[62824]: DEBUG oslo_vmware.api [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Task: {'id': task-2145746, 'name': Rename_Task} progress is 10%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1911.600501] env[62824]: DEBUG oslo_vmware.api [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Waiting for the task: (returnval){ [ 1911.600501] env[62824]: value = "task-2145747" [ 1911.600501] env[62824]: _type = "Task" [ 1911.600501] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1911.610658] env[62824]: DEBUG oslo_vmware.api [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Task: {'id': task-2145747, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1911.630054] env[62824]: INFO nova.compute.manager [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] [instance: 4dfe9f82-7f89-486b-b159-948656d2896c] Took 29.58 seconds to build instance. [ 1911.857574] env[62824]: DEBUG nova.compute.manager [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: f3740d9a-b21a-4a79-9e28-2a89ecd08bb6] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1911.889105] env[62824]: DEBUG nova.virt.hardware [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1911.889415] env[62824]: DEBUG nova.virt.hardware [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1911.889854] env[62824]: DEBUG nova.virt.hardware [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1911.890101] env[62824]: DEBUG nova.virt.hardware [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1911.890258] env[62824]: DEBUG nova.virt.hardware [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1911.890408] env[62824]: DEBUG nova.virt.hardware [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1911.890659] env[62824]: DEBUG nova.virt.hardware [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1911.890773] env[62824]: DEBUG nova.virt.hardware [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1911.890984] env[62824]: DEBUG nova.virt.hardware [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1911.891164] env[62824]: DEBUG nova.virt.hardware [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1911.891345] env[62824]: DEBUG nova.virt.hardware [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1911.892710] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f88740c3-4d6e-4d61-a62d-28c4fdb4dd22 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1911.900927] env[62824]: DEBUG nova.compute.manager [req-4990e6f7-7747-4451-acf2-e63bcdeb0b3f req-d59247dd-5243-46c4-841d-39af065a48bf service nova] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Received event network-changed-309b4346-1935-41f1-a374-d322c4a8c016 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1911.900927] env[62824]: DEBUG nova.compute.manager [req-4990e6f7-7747-4451-acf2-e63bcdeb0b3f req-d59247dd-5243-46c4-841d-39af065a48bf service nova] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Refreshing instance network info cache due to event network-changed-309b4346-1935-41f1-a374-d322c4a8c016. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1911.901136] env[62824]: DEBUG oslo_concurrency.lockutils [req-4990e6f7-7747-4451-acf2-e63bcdeb0b3f req-d59247dd-5243-46c4-841d-39af065a48bf service nova] Acquiring lock "refresh_cache-07ce931f-45ef-409b-b714-9f1cd47a3a88" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1911.901364] env[62824]: DEBUG oslo_concurrency.lockutils [req-4990e6f7-7747-4451-acf2-e63bcdeb0b3f req-d59247dd-5243-46c4-841d-39af065a48bf service nova] Acquired lock "refresh_cache-07ce931f-45ef-409b-b714-9f1cd47a3a88" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1911.901516] env[62824]: DEBUG nova.network.neutron [req-4990e6f7-7747-4451-acf2-e63bcdeb0b3f req-d59247dd-5243-46c4-841d-39af065a48bf service nova] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Refreshing network info cache for port 309b4346-1935-41f1-a374-d322c4a8c016 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1911.906596] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1c10a40-b416-4d2a-9104-63b427d37214 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1912.017474] env[62824]: DEBUG nova.scheduler.client.report [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1912.053050] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145745, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1912.096569] env[62824]: DEBUG oslo_vmware.api [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Task: {'id': task-2145746, 'name': Rename_Task, 'duration_secs': 0.189975} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1912.096848] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1912.097125] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-34e87eaf-32ef-4fad-bbd5-f0079cd543d3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1912.107302] env[62824]: DEBUG oslo_vmware.api [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Waiting for the task: (returnval){ [ 1912.107302] env[62824]: value = "task-2145748" [ 1912.107302] env[62824]: _type = "Task" [ 1912.107302] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1912.113807] env[62824]: DEBUG oslo_vmware.api [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Task: {'id': task-2145747, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.091383} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1912.114375] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1912.115185] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89ec4644-0f0c-4e74-b7c9-5591b8c56122 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1912.120923] env[62824]: DEBUG oslo_vmware.api [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Task: {'id': task-2145748, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1912.140016] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Reconfiguring VM instance instance-00000014 to attach disk [datastore2] 384a96ea-40ce-43e1-a5f9-82f50b710b0b/384a96ea-40ce-43e1-a5f9-82f50b710b0b.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1912.140488] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c5a8bd67-7270-40df-93d3-dea4923f4b1b tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Lock "4dfe9f82-7f89-486b-b159-948656d2896c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 58.641s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1912.140759] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1ea1179c-b27f-4bff-9a75-01fcce35ad7f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1912.161951] env[62824]: DEBUG oslo_vmware.api [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Waiting for the task: (returnval){ [ 1912.161951] env[62824]: value = "task-2145749" [ 1912.161951] env[62824]: _type = "Task" [ 1912.161951] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1912.173414] env[62824]: DEBUG oslo_vmware.api [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Task: {'id': task-2145749, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1912.526182] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.700s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1912.527196] env[62824]: DEBUG nova.compute.manager [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] [instance: ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1912.531856] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.039s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1912.532099] env[62824]: INFO nova.compute.claims [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1912.555667] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145745, 'name': CreateVM_Task, 'duration_secs': 0.694244} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1912.559364] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1912.560773] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1912.561101] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1912.561654] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1912.562641] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-26a93e13-e4cf-4277-bd51-1f62994f88e8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1912.569875] env[62824]: DEBUG oslo_vmware.api [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for the task: (returnval){ [ 1912.569875] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5298fe7e-535b-6637-9f4c-50ac78f3fb69" [ 1912.569875] env[62824]: _type = "Task" [ 1912.569875] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1912.580738] env[62824]: DEBUG oslo_vmware.api [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5298fe7e-535b-6637-9f4c-50ac78f3fb69, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1912.617068] env[62824]: DEBUG oslo_vmware.api [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Task: {'id': task-2145748, 'name': PowerOnVM_Task, 'duration_secs': 0.468637} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1912.617879] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1912.617879] env[62824]: INFO nova.compute.manager [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Took 8.18 seconds to spawn the instance on the hypervisor. [ 1912.617879] env[62824]: DEBUG nova.compute.manager [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1912.618502] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1e3cfc5-8c69-47e3-ba90-d424fd35dba0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1912.661112] env[62824]: DEBUG nova.compute.manager [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1912.677759] env[62824]: DEBUG oslo_vmware.api [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Task: {'id': task-2145749, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1912.774484] env[62824]: DEBUG nova.network.neutron [req-4990e6f7-7747-4451-acf2-e63bcdeb0b3f req-d59247dd-5243-46c4-841d-39af065a48bf service nova] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Updated VIF entry in instance network info cache for port 309b4346-1935-41f1-a374-d322c4a8c016. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1912.774895] env[62824]: DEBUG nova.network.neutron [req-4990e6f7-7747-4451-acf2-e63bcdeb0b3f req-d59247dd-5243-46c4-841d-39af065a48bf service nova] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Updating instance_info_cache with network_info: [{"id": "309b4346-1935-41f1-a374-d322c4a8c016", "address": "fa:16:3e:07:e9:a0", "network": {"id": "fc381ca5-198f-4afc-9ebf-bc6e3de68c35", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-425455451-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "248af2bc3e7b41d2a74b0c6074b359e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4c9a12d2-469f-4199-bfaa-f791d765deac", "external-id": "nsx-vlan-transportzone-96", "segmentation_id": 96, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap309b4346-19", "ovs_interfaceid": "309b4346-1935-41f1-a374-d322c4a8c016", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1913.040095] env[62824]: DEBUG nova.compute.utils [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1913.041419] env[62824]: DEBUG nova.compute.manager [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] [instance: ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1913.041604] env[62824]: DEBUG nova.network.neutron [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] [instance: ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1913.083178] env[62824]: DEBUG oslo_vmware.api [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5298fe7e-535b-6637-9f4c-50ac78f3fb69, 'name': SearchDatastore_Task, 'duration_secs': 0.012233} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1913.083178] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1913.083178] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1913.083178] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1913.083317] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1913.083317] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1913.083317] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1210628c-b6c4-485e-8488-f0d97011a483 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1913.092705] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1913.092874] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1913.093842] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0296758c-7a7c-4bea-b1ce-321a3de6df45 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1913.098962] env[62824]: DEBUG oslo_vmware.api [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for the task: (returnval){ [ 1913.098962] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52b07d00-f825-fd11-cb3d-25a2027ee108" [ 1913.098962] env[62824]: _type = "Task" [ 1913.098962] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1913.106826] env[62824]: DEBUG oslo_vmware.api [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52b07d00-f825-fd11-cb3d-25a2027ee108, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1913.115348] env[62824]: DEBUG nova.policy [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '71e13a6edb884a2aaacbbbcb6c9e8ae5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '97529775bace4449b194dcef41bea531', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 1913.139235] env[62824]: INFO nova.compute.manager [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Took 29.50 seconds to build instance. [ 1913.181492] env[62824]: DEBUG oslo_vmware.api [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Task: {'id': task-2145749, 'name': ReconfigVM_Task, 'duration_secs': 0.703185} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1913.182900] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Reconfigured VM instance instance-00000014 to attach disk [datastore2] 384a96ea-40ce-43e1-a5f9-82f50b710b0b/384a96ea-40ce-43e1-a5f9-82f50b710b0b.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1913.188361] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-db908012-7951-4ac9-bc87-2454771a08fe {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1913.195763] env[62824]: DEBUG oslo_vmware.api [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Waiting for the task: (returnval){ [ 1913.195763] env[62824]: value = "task-2145750" [ 1913.195763] env[62824]: _type = "Task" [ 1913.195763] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1913.207151] env[62824]: DEBUG oslo_vmware.api [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Task: {'id': task-2145750, 'name': Rename_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1913.208611] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1913.280861] env[62824]: DEBUG oslo_concurrency.lockutils [req-4990e6f7-7747-4451-acf2-e63bcdeb0b3f req-d59247dd-5243-46c4-841d-39af065a48bf service nova] Releasing lock "refresh_cache-07ce931f-45ef-409b-b714-9f1cd47a3a88" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1913.280861] env[62824]: DEBUG nova.compute.manager [req-4990e6f7-7747-4451-acf2-e63bcdeb0b3f req-d59247dd-5243-46c4-841d-39af065a48bf service nova] [instance: dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda] Received event network-changed-9f1b3f36-9c69-4e99-8030-2d2ec535b324 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1913.280861] env[62824]: DEBUG nova.compute.manager [req-4990e6f7-7747-4451-acf2-e63bcdeb0b3f req-d59247dd-5243-46c4-841d-39af065a48bf service nova] [instance: dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda] Refreshing instance network info cache due to event network-changed-9f1b3f36-9c69-4e99-8030-2d2ec535b324. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1913.280861] env[62824]: DEBUG oslo_concurrency.lockutils [req-4990e6f7-7747-4451-acf2-e63bcdeb0b3f req-d59247dd-5243-46c4-841d-39af065a48bf service nova] Acquiring lock "refresh_cache-dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1913.281065] env[62824]: DEBUG oslo_concurrency.lockutils [req-4990e6f7-7747-4451-acf2-e63bcdeb0b3f req-d59247dd-5243-46c4-841d-39af065a48bf service nova] Acquired lock "refresh_cache-dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1913.281955] env[62824]: DEBUG nova.network.neutron [req-4990e6f7-7747-4451-acf2-e63bcdeb0b3f req-d59247dd-5243-46c4-841d-39af065a48bf service nova] [instance: dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda] Refreshing network info cache for port 9f1b3f36-9c69-4e99-8030-2d2ec535b324 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1913.465041] env[62824]: DEBUG nova.network.neutron [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] [instance: ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5] Successfully created port: e8535330-ccec-4e36-8f78-cf45232f2a44 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1913.553937] env[62824]: DEBUG nova.compute.manager [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] [instance: ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1913.615231] env[62824]: DEBUG oslo_vmware.api [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52b07d00-f825-fd11-cb3d-25a2027ee108, 'name': SearchDatastore_Task, 'duration_secs': 0.033543} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1913.616154] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9af08d55-5eef-422f-913e-89e902e8a79d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1913.625292] env[62824]: DEBUG oslo_vmware.api [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for the task: (returnval){ [ 1913.625292] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]528945c8-6a82-28dd-7163-bb48457a083c" [ 1913.625292] env[62824]: _type = "Task" [ 1913.625292] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1913.635598] env[62824]: DEBUG oslo_vmware.api [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]528945c8-6a82-28dd-7163-bb48457a083c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1913.641419] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bfc7096d-2bcc-4dcf-9da0-a5b68635ced0 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Lock "e158b5d4-c120-4e6c-89c4-7668e097926d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 54.164s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1913.675180] env[62824]: DEBUG nova.network.neutron [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: f3740d9a-b21a-4a79-9e28-2a89ecd08bb6] Successfully updated port: f2fb594d-5cb0-4c90-9787-3a362864f81a {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1913.708541] env[62824]: DEBUG oslo_vmware.api [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Task: {'id': task-2145750, 'name': Rename_Task, 'duration_secs': 0.16886} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1913.710154] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1913.710655] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-82266772-fff7-47b8-9779-fb06e0eb3229 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1913.719543] env[62824]: DEBUG oslo_vmware.api [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Waiting for the task: (returnval){ [ 1913.719543] env[62824]: value = "task-2145751" [ 1913.719543] env[62824]: _type = "Task" [ 1913.719543] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1913.729017] env[62824]: DEBUG oslo_vmware.api [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Task: {'id': task-2145751, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1913.775896] env[62824]: DEBUG nova.compute.manager [req-9077b675-f40d-4bb1-a287-098ad454608e req-6286eeb6-4eb1-452d-b3c6-6af13b4c0572 service nova] [instance: f3740d9a-b21a-4a79-9e28-2a89ecd08bb6] Received event network-vif-plugged-f2fb594d-5cb0-4c90-9787-3a362864f81a {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1913.776132] env[62824]: DEBUG oslo_concurrency.lockutils [req-9077b675-f40d-4bb1-a287-098ad454608e req-6286eeb6-4eb1-452d-b3c6-6af13b4c0572 service nova] Acquiring lock "f3740d9a-b21a-4a79-9e28-2a89ecd08bb6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1913.776399] env[62824]: DEBUG oslo_concurrency.lockutils [req-9077b675-f40d-4bb1-a287-098ad454608e req-6286eeb6-4eb1-452d-b3c6-6af13b4c0572 service nova] Lock "f3740d9a-b21a-4a79-9e28-2a89ecd08bb6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1913.776489] env[62824]: DEBUG oslo_concurrency.lockutils [req-9077b675-f40d-4bb1-a287-098ad454608e req-6286eeb6-4eb1-452d-b3c6-6af13b4c0572 service nova] Lock "f3740d9a-b21a-4a79-9e28-2a89ecd08bb6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1913.776649] env[62824]: DEBUG nova.compute.manager [req-9077b675-f40d-4bb1-a287-098ad454608e req-6286eeb6-4eb1-452d-b3c6-6af13b4c0572 service nova] [instance: f3740d9a-b21a-4a79-9e28-2a89ecd08bb6] No waiting events found dispatching network-vif-plugged-f2fb594d-5cb0-4c90-9787-3a362864f81a {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1913.777441] env[62824]: WARNING nova.compute.manager [req-9077b675-f40d-4bb1-a287-098ad454608e req-6286eeb6-4eb1-452d-b3c6-6af13b4c0572 service nova] [instance: f3740d9a-b21a-4a79-9e28-2a89ecd08bb6] Received unexpected event network-vif-plugged-f2fb594d-5cb0-4c90-9787-3a362864f81a for instance with vm_state building and task_state spawning. [ 1913.874795] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0b8e688c-e818-48ea-b84f-ab6b2e8d5cfe tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Acquiring lock "4dfe9f82-7f89-486b-b159-948656d2896c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1913.874795] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0b8e688c-e818-48ea-b84f-ab6b2e8d5cfe tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Lock "4dfe9f82-7f89-486b-b159-948656d2896c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1913.875285] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0b8e688c-e818-48ea-b84f-ab6b2e8d5cfe tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Acquiring lock "4dfe9f82-7f89-486b-b159-948656d2896c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1913.875360] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0b8e688c-e818-48ea-b84f-ab6b2e8d5cfe tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Lock "4dfe9f82-7f89-486b-b159-948656d2896c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1913.875557] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0b8e688c-e818-48ea-b84f-ab6b2e8d5cfe tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Lock "4dfe9f82-7f89-486b-b159-948656d2896c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1913.877914] env[62824]: INFO nova.compute.manager [None req-0b8e688c-e818-48ea-b84f-ab6b2e8d5cfe tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] [instance: 4dfe9f82-7f89-486b-b159-948656d2896c] Terminating instance [ 1914.144697] env[62824]: DEBUG oslo_vmware.api [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]528945c8-6a82-28dd-7163-bb48457a083c, 'name': SearchDatastore_Task, 'duration_secs': 0.011841} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1914.145296] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1914.145591] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 07ce931f-45ef-409b-b714-9f1cd47a3a88/07ce931f-45ef-409b-b714-9f1cd47a3a88.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1914.145845] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-10ccc391-304e-4ac5-84f8-39f47f7986a1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1914.148097] env[62824]: DEBUG nova.compute.manager [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: 6d80ec06-8559-4964-8577-a2512aa366ed] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1914.159111] env[62824]: DEBUG oslo_vmware.api [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for the task: (returnval){ [ 1914.159111] env[62824]: value = "task-2145752" [ 1914.159111] env[62824]: _type = "Task" [ 1914.159111] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1914.167855] env[62824]: DEBUG oslo_vmware.api [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2145752, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1914.184691] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Acquiring lock "refresh_cache-f3740d9a-b21a-4a79-9e28-2a89ecd08bb6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1914.184691] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Acquired lock "refresh_cache-f3740d9a-b21a-4a79-9e28-2a89ecd08bb6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1914.184691] env[62824]: DEBUG nova.network.neutron [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: f3740d9a-b21a-4a79-9e28-2a89ecd08bb6] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1914.194019] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b164f42-d864-4fb7-9f32-86f226069a6e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1914.204026] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f557729-b535-49b6-9c44-8adf301cfc63 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1914.247274] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf76db4f-a1b1-426d-b81b-54dff4b96d3c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1914.257081] env[62824]: DEBUG oslo_vmware.api [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Task: {'id': task-2145751, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1914.260668] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcd404ba-1454-4d5b-b467-333710b09e4a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1914.274726] env[62824]: DEBUG nova.compute.provider_tree [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1914.333924] env[62824]: DEBUG nova.network.neutron [req-4990e6f7-7747-4451-acf2-e63bcdeb0b3f req-d59247dd-5243-46c4-841d-39af065a48bf service nova] [instance: dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda] Updated VIF entry in instance network info cache for port 9f1b3f36-9c69-4e99-8030-2d2ec535b324. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1914.334371] env[62824]: DEBUG nova.network.neutron [req-4990e6f7-7747-4451-acf2-e63bcdeb0b3f req-d59247dd-5243-46c4-841d-39af065a48bf service nova] [instance: dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda] Updating instance_info_cache with network_info: [{"id": "9f1b3f36-9c69-4e99-8030-2d2ec535b324", "address": "fa:16:3e:91:83:1b", "network": {"id": "a0e32d69-182a-4e0c-a984-bfe393b641d7", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-2040711326-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.178", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "51135bed305c4e89afaeed2b187cc7f6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "274afb4c-04df-4213-8ad2-8f48a10d78a8", "external-id": "nsx-vlan-transportzone-515", "segmentation_id": 515, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9f1b3f36-9c", "ovs_interfaceid": "9f1b3f36-9c69-4e99-8030-2d2ec535b324", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1914.389726] env[62824]: DEBUG nova.compute.manager [None req-0b8e688c-e818-48ea-b84f-ab6b2e8d5cfe tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] [instance: 4dfe9f82-7f89-486b-b159-948656d2896c] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1914.389956] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-0b8e688c-e818-48ea-b84f-ab6b2e8d5cfe tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] [instance: 4dfe9f82-7f89-486b-b159-948656d2896c] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1914.391095] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae7cc7db-d38e-42fd-9a79-37dbb2def1ea {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1914.399014] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b8e688c-e818-48ea-b84f-ab6b2e8d5cfe tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] [instance: 4dfe9f82-7f89-486b-b159-948656d2896c] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1914.399217] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9d7c60a5-d664-4bc2-ae05-db8249d497f8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1914.406674] env[62824]: DEBUG oslo_vmware.api [None req-0b8e688c-e818-48ea-b84f-ab6b2e8d5cfe tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Waiting for the task: (returnval){ [ 1914.406674] env[62824]: value = "task-2145753" [ 1914.406674] env[62824]: _type = "Task" [ 1914.406674] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1914.416845] env[62824]: DEBUG oslo_vmware.api [None req-0b8e688c-e818-48ea-b84f-ab6b2e8d5cfe tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Task: {'id': task-2145753, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1914.563300] env[62824]: DEBUG nova.compute.manager [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] [instance: ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1914.600612] env[62824]: DEBUG nova.virt.hardware [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1914.600956] env[62824]: DEBUG nova.virt.hardware [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1914.601123] env[62824]: DEBUG nova.virt.hardware [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1914.601372] env[62824]: DEBUG nova.virt.hardware [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1914.601685] env[62824]: DEBUG nova.virt.hardware [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1914.601806] env[62824]: DEBUG nova.virt.hardware [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1914.602181] env[62824]: DEBUG nova.virt.hardware [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1914.602340] env[62824]: DEBUG nova.virt.hardware [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1914.602652] env[62824]: DEBUG nova.virt.hardware [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1914.602896] env[62824]: DEBUG nova.virt.hardware [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1914.603151] env[62824]: DEBUG nova.virt.hardware [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1914.604302] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcc05eb1-985d-481c-8563-ed6425688dbc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1914.614904] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35b60bab-7e0c-4e11-853f-588841ea5fb8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1914.675107] env[62824]: DEBUG oslo_vmware.api [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2145752, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1914.677404] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1914.750131] env[62824]: DEBUG oslo_vmware.api [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Task: {'id': task-2145751, 'name': PowerOnVM_Task, 'duration_secs': 0.69781} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1914.751330] env[62824]: DEBUG nova.network.neutron [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: f3740d9a-b21a-4a79-9e28-2a89ecd08bb6] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1914.753500] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1914.753857] env[62824]: DEBUG nova.compute.manager [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1914.754885] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f6dbff0-04dd-4cba-90b2-83a6a1101612 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1914.779504] env[62824]: DEBUG nova.scheduler.client.report [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1914.837758] env[62824]: DEBUG oslo_concurrency.lockutils [req-4990e6f7-7747-4451-acf2-e63bcdeb0b3f req-d59247dd-5243-46c4-841d-39af065a48bf service nova] Releasing lock "refresh_cache-dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1914.916888] env[62824]: DEBUG oslo_vmware.api [None req-0b8e688c-e818-48ea-b84f-ab6b2e8d5cfe tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Task: {'id': task-2145753, 'name': PowerOffVM_Task, 'duration_secs': 0.358003} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1914.917301] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b8e688c-e818-48ea-b84f-ab6b2e8d5cfe tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] [instance: 4dfe9f82-7f89-486b-b159-948656d2896c] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1914.917549] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-0b8e688c-e818-48ea-b84f-ab6b2e8d5cfe tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] [instance: 4dfe9f82-7f89-486b-b159-948656d2896c] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1914.917776] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ca76845d-3aec-4f1f-8293-b63a1d8e395c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1915.008187] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-0b8e688c-e818-48ea-b84f-ab6b2e8d5cfe tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] [instance: 4dfe9f82-7f89-486b-b159-948656d2896c] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1915.008560] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-0b8e688c-e818-48ea-b84f-ab6b2e8d5cfe tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] [instance: 4dfe9f82-7f89-486b-b159-948656d2896c] Deleting contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1915.008759] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b8e688c-e818-48ea-b84f-ab6b2e8d5cfe tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Deleting the datastore file [datastore2] 4dfe9f82-7f89-486b-b159-948656d2896c {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1915.010167] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-62507ad9-1476-4c3b-a2fd-508b8cdc5f8c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1915.017410] env[62824]: DEBUG oslo_vmware.api [None req-0b8e688c-e818-48ea-b84f-ab6b2e8d5cfe tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Waiting for the task: (returnval){ [ 1915.017410] env[62824]: value = "task-2145755" [ 1915.017410] env[62824]: _type = "Task" [ 1915.017410] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1915.037322] env[62824]: DEBUG oslo_vmware.api [None req-0b8e688c-e818-48ea-b84f-ab6b2e8d5cfe tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Task: {'id': task-2145755, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1915.090660] env[62824]: DEBUG nova.compute.manager [req-d9fbbe7a-945d-4600-9ced-30e43930677d req-73cf4747-0d0e-4621-a9d9-96cd84d5406a service nova] [instance: ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5] Received event network-vif-plugged-e8535330-ccec-4e36-8f78-cf45232f2a44 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1915.091421] env[62824]: DEBUG oslo_concurrency.lockutils [req-d9fbbe7a-945d-4600-9ced-30e43930677d req-73cf4747-0d0e-4621-a9d9-96cd84d5406a service nova] Acquiring lock "ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1915.091647] env[62824]: DEBUG oslo_concurrency.lockutils [req-d9fbbe7a-945d-4600-9ced-30e43930677d req-73cf4747-0d0e-4621-a9d9-96cd84d5406a service nova] Lock "ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1915.091859] env[62824]: DEBUG oslo_concurrency.lockutils [req-d9fbbe7a-945d-4600-9ced-30e43930677d req-73cf4747-0d0e-4621-a9d9-96cd84d5406a service nova] Lock "ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1915.092027] env[62824]: DEBUG nova.compute.manager [req-d9fbbe7a-945d-4600-9ced-30e43930677d req-73cf4747-0d0e-4621-a9d9-96cd84d5406a service nova] [instance: ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5] No waiting events found dispatching network-vif-plugged-e8535330-ccec-4e36-8f78-cf45232f2a44 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1915.092199] env[62824]: WARNING nova.compute.manager [req-d9fbbe7a-945d-4600-9ced-30e43930677d req-73cf4747-0d0e-4621-a9d9-96cd84d5406a service nova] [instance: ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5] Received unexpected event network-vif-plugged-e8535330-ccec-4e36-8f78-cf45232f2a44 for instance with vm_state building and task_state spawning. [ 1915.093410] env[62824]: DEBUG nova.network.neutron [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: f3740d9a-b21a-4a79-9e28-2a89ecd08bb6] Updating instance_info_cache with network_info: [{"id": "f2fb594d-5cb0-4c90-9787-3a362864f81a", "address": "fa:16:3e:e9:f6:8e", "network": {"id": "fc381ca5-198f-4afc-9ebf-bc6e3de68c35", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-425455451-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "248af2bc3e7b41d2a74b0c6074b359e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4c9a12d2-469f-4199-bfaa-f791d765deac", "external-id": "nsx-vlan-transportzone-96", "segmentation_id": 96, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf2fb594d-5c", "ovs_interfaceid": "f2fb594d-5cb0-4c90-9787-3a362864f81a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1915.136931] env[62824]: DEBUG nova.network.neutron [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] [instance: ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5] Successfully updated port: e8535330-ccec-4e36-8f78-cf45232f2a44 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1915.173960] env[62824]: DEBUG oslo_vmware.api [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2145752, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.5337} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1915.174249] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 07ce931f-45ef-409b-b714-9f1cd47a3a88/07ce931f-45ef-409b-b714-9f1cd47a3a88.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1915.174494] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1915.174743] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-34dca04d-8331-4f28-90d2-f79e656d0a21 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1915.181339] env[62824]: DEBUG oslo_vmware.api [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for the task: (returnval){ [ 1915.181339] env[62824]: value = "task-2145756" [ 1915.181339] env[62824]: _type = "Task" [ 1915.181339] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1915.190265] env[62824]: DEBUG oslo_vmware.api [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2145756, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1915.276196] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1915.291873] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.762s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1915.292501] env[62824]: DEBUG nova.compute.manager [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1915.295843] env[62824]: DEBUG oslo_concurrency.lockutils [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.717s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1915.301295] env[62824]: INFO nova.compute.claims [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] [instance: b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1915.530121] env[62824]: DEBUG oslo_vmware.api [None req-0b8e688c-e818-48ea-b84f-ab6b2e8d5cfe tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Task: {'id': task-2145755, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.211451} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1915.530788] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b8e688c-e818-48ea-b84f-ab6b2e8d5cfe tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1915.530986] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-0b8e688c-e818-48ea-b84f-ab6b2e8d5cfe tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] [instance: 4dfe9f82-7f89-486b-b159-948656d2896c] Deleted contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1915.531172] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-0b8e688c-e818-48ea-b84f-ab6b2e8d5cfe tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] [instance: 4dfe9f82-7f89-486b-b159-948656d2896c] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1915.531346] env[62824]: INFO nova.compute.manager [None req-0b8e688c-e818-48ea-b84f-ab6b2e8d5cfe tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] [instance: 4dfe9f82-7f89-486b-b159-948656d2896c] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1915.531583] env[62824]: DEBUG oslo.service.loopingcall [None req-0b8e688c-e818-48ea-b84f-ab6b2e8d5cfe tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1915.531775] env[62824]: DEBUG nova.compute.manager [-] [instance: 4dfe9f82-7f89-486b-b159-948656d2896c] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1915.531965] env[62824]: DEBUG nova.network.neutron [-] [instance: 4dfe9f82-7f89-486b-b159-948656d2896c] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1915.596738] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Releasing lock "refresh_cache-f3740d9a-b21a-4a79-9e28-2a89ecd08bb6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1915.599798] env[62824]: DEBUG nova.compute.manager [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: f3740d9a-b21a-4a79-9e28-2a89ecd08bb6] Instance network_info: |[{"id": "f2fb594d-5cb0-4c90-9787-3a362864f81a", "address": "fa:16:3e:e9:f6:8e", "network": {"id": "fc381ca5-198f-4afc-9ebf-bc6e3de68c35", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-425455451-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "248af2bc3e7b41d2a74b0c6074b359e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4c9a12d2-469f-4199-bfaa-f791d765deac", "external-id": "nsx-vlan-transportzone-96", "segmentation_id": 96, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf2fb594d-5c", "ovs_interfaceid": "f2fb594d-5cb0-4c90-9787-3a362864f81a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1915.599904] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: f3740d9a-b21a-4a79-9e28-2a89ecd08bb6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e9:f6:8e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4c9a12d2-469f-4199-bfaa-f791d765deac', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f2fb594d-5cb0-4c90-9787-3a362864f81a', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1915.611649] env[62824]: DEBUG oslo.service.loopingcall [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1915.613056] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f3740d9a-b21a-4a79-9e28-2a89ecd08bb6] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1915.613056] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2955bb83-521c-494d-8798-738d7b3bc78a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1915.643061] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Acquiring lock "refresh_cache-ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1915.643061] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Acquired lock "refresh_cache-ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1915.643061] env[62824]: DEBUG nova.network.neutron [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] [instance: ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1915.643524] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1915.643524] env[62824]: value = "task-2145757" [ 1915.643524] env[62824]: _type = "Task" [ 1915.643524] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1915.654069] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145757, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1915.691313] env[62824]: DEBUG oslo_vmware.api [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2145756, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068782} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1915.691682] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1915.692704] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1d9d2ee-bd94-43cb-a10a-c321bed960cf {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1915.724026] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Reconfiguring VM instance instance-00000018 to attach disk [datastore1] 07ce931f-45ef-409b-b714-9f1cd47a3a88/07ce931f-45ef-409b-b714-9f1cd47a3a88.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1915.725742] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e7edb801-96cc-490b-a34a-ac8bccafde43 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1915.750154] env[62824]: DEBUG oslo_vmware.api [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for the task: (returnval){ [ 1915.750154] env[62824]: value = "task-2145758" [ 1915.750154] env[62824]: _type = "Task" [ 1915.750154] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1915.759919] env[62824]: DEBUG oslo_vmware.api [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2145758, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1915.807161] env[62824]: DEBUG nova.compute.utils [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1915.812025] env[62824]: DEBUG nova.compute.manager [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1915.812025] env[62824]: DEBUG nova.network.neutron [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1915.880010] env[62824]: DEBUG nova.policy [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4c914343569a462d9b8564f3376c2f54', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '10ee5fe93d584725ace90e76605a1b2e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 1916.021901] env[62824]: INFO nova.compute.manager [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Rebuilding instance [ 1916.048436] env[62824]: DEBUG nova.compute.manager [req-524509d6-75d1-4d50-9b1a-d81882f462b0 req-aa518129-0011-46eb-9f74-7a2cc90cad7d service nova] [instance: f3740d9a-b21a-4a79-9e28-2a89ecd08bb6] Received event network-changed-f2fb594d-5cb0-4c90-9787-3a362864f81a {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1916.049103] env[62824]: DEBUG nova.compute.manager [req-524509d6-75d1-4d50-9b1a-d81882f462b0 req-aa518129-0011-46eb-9f74-7a2cc90cad7d service nova] [instance: f3740d9a-b21a-4a79-9e28-2a89ecd08bb6] Refreshing instance network info cache due to event network-changed-f2fb594d-5cb0-4c90-9787-3a362864f81a. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1916.049353] env[62824]: DEBUG oslo_concurrency.lockutils [req-524509d6-75d1-4d50-9b1a-d81882f462b0 req-aa518129-0011-46eb-9f74-7a2cc90cad7d service nova] Acquiring lock "refresh_cache-f3740d9a-b21a-4a79-9e28-2a89ecd08bb6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1916.049446] env[62824]: DEBUG oslo_concurrency.lockutils [req-524509d6-75d1-4d50-9b1a-d81882f462b0 req-aa518129-0011-46eb-9f74-7a2cc90cad7d service nova] Acquired lock "refresh_cache-f3740d9a-b21a-4a79-9e28-2a89ecd08bb6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1916.049610] env[62824]: DEBUG nova.network.neutron [req-524509d6-75d1-4d50-9b1a-d81882f462b0 req-aa518129-0011-46eb-9f74-7a2cc90cad7d service nova] [instance: f3740d9a-b21a-4a79-9e28-2a89ecd08bb6] Refreshing network info cache for port f2fb594d-5cb0-4c90-9787-3a362864f81a {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1916.077331] env[62824]: DEBUG nova.compute.manager [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1916.077910] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43cf3701-83af-4cee-8695-ad88483baaec {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1916.162042] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145757, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1916.176055] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Acquiring lock "b9f30f87-3594-4468-9d29-70890d8761e3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1916.176290] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Lock "b9f30f87-3594-4468-9d29-70890d8761e3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1916.194672] env[62824]: DEBUG nova.network.neutron [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] [instance: ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1916.253371] env[62824]: DEBUG nova.compute.manager [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Stashing vm_state: active {{(pid=62824) _prep_resize /opt/stack/nova/nova/compute/manager.py:6136}} [ 1916.268884] env[62824]: DEBUG oslo_vmware.api [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2145758, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1916.312197] env[62824]: DEBUG nova.compute.manager [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1916.375783] env[62824]: DEBUG nova.network.neutron [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] [instance: ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5] Updating instance_info_cache with network_info: [{"id": "e8535330-ccec-4e36-8f78-cf45232f2a44", "address": "fa:16:3e:d8:16:53", "network": {"id": "b70367ae-07d9-4a3e-bfbc-4512721997ec", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-1050095803-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "97529775bace4449b194dcef41bea531", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "863474bc-a24a-4823-828c-580a187829e3", "external-id": "nsx-vlan-transportzone-617", "segmentation_id": 617, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape8535330-cc", "ovs_interfaceid": "e8535330-ccec-4e36-8f78-cf45232f2a44", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1916.392459] env[62824]: DEBUG nova.network.neutron [-] [instance: 4dfe9f82-7f89-486b-b159-948656d2896c] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1916.601328] env[62824]: DEBUG nova.network.neutron [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] Successfully created port: e0e9725d-7e5a-452f-bd1e-960934705a7f {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1916.659271] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145757, 'name': CreateVM_Task, 'duration_secs': 0.770075} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1916.660557] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f3740d9a-b21a-4a79-9e28-2a89ecd08bb6] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1916.660557] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1916.660557] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1916.661980] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1916.661980] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1860aeb2-4131-4b83-8bd6-fff9258d6665 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1916.668056] env[62824]: DEBUG oslo_vmware.api [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for the task: (returnval){ [ 1916.668056] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52973cae-2b52-e014-da37-bd501a549e4b" [ 1916.668056] env[62824]: _type = "Task" [ 1916.668056] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1916.677223] env[62824]: DEBUG oslo_vmware.api [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52973cae-2b52-e014-da37-bd501a549e4b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1916.767334] env[62824]: DEBUG oslo_vmware.api [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2145758, 'name': ReconfigVM_Task, 'duration_secs': 0.722092} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1916.769927] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Reconfigured VM instance instance-00000018 to attach disk [datastore1] 07ce931f-45ef-409b-b714-9f1cd47a3a88/07ce931f-45ef-409b-b714-9f1cd47a3a88.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1916.770991] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5cb852e1-dc1e-4b30-8397-2e1e16aff17f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1916.778892] env[62824]: DEBUG oslo_vmware.api [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for the task: (returnval){ [ 1916.778892] env[62824]: value = "task-2145759" [ 1916.778892] env[62824]: _type = "Task" [ 1916.778892] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1916.785658] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1916.788593] env[62824]: DEBUG oslo_vmware.api [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2145759, 'name': Rename_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1916.820420] env[62824]: DEBUG nova.network.neutron [req-524509d6-75d1-4d50-9b1a-d81882f462b0 req-aa518129-0011-46eb-9f74-7a2cc90cad7d service nova] [instance: f3740d9a-b21a-4a79-9e28-2a89ecd08bb6] Updated VIF entry in instance network info cache for port f2fb594d-5cb0-4c90-9787-3a362864f81a. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1916.820819] env[62824]: DEBUG nova.network.neutron [req-524509d6-75d1-4d50-9b1a-d81882f462b0 req-aa518129-0011-46eb-9f74-7a2cc90cad7d service nova] [instance: f3740d9a-b21a-4a79-9e28-2a89ecd08bb6] Updating instance_info_cache with network_info: [{"id": "f2fb594d-5cb0-4c90-9787-3a362864f81a", "address": "fa:16:3e:e9:f6:8e", "network": {"id": "fc381ca5-198f-4afc-9ebf-bc6e3de68c35", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-425455451-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "248af2bc3e7b41d2a74b0c6074b359e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4c9a12d2-469f-4199-bfaa-f791d765deac", "external-id": "nsx-vlan-transportzone-96", "segmentation_id": 96, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf2fb594d-5c", "ovs_interfaceid": "f2fb594d-5cb0-4c90-9787-3a362864f81a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1916.878192] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Releasing lock "refresh_cache-ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1916.878558] env[62824]: DEBUG nova.compute.manager [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] [instance: ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5] Instance network_info: |[{"id": "e8535330-ccec-4e36-8f78-cf45232f2a44", "address": "fa:16:3e:d8:16:53", "network": {"id": "b70367ae-07d9-4a3e-bfbc-4512721997ec", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-1050095803-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "97529775bace4449b194dcef41bea531", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "863474bc-a24a-4823-828c-580a187829e3", "external-id": "nsx-vlan-transportzone-617", "segmentation_id": 617, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape8535330-cc", "ovs_interfaceid": "e8535330-ccec-4e36-8f78-cf45232f2a44", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1916.882213] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] [instance: ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d8:16:53', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '863474bc-a24a-4823-828c-580a187829e3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e8535330-ccec-4e36-8f78-cf45232f2a44', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1916.890212] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Creating folder: Project (97529775bace4449b194dcef41bea531). Parent ref: group-v438503. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1916.892540] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-eceed49b-81d4-48b6-ae29-39875f7c3bd0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1916.896671] env[62824]: INFO nova.compute.manager [-] [instance: 4dfe9f82-7f89-486b-b159-948656d2896c] Took 1.36 seconds to deallocate network for instance. [ 1916.902667] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Created folder: Project (97529775bace4449b194dcef41bea531) in parent group-v438503. [ 1916.902839] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Creating folder: Instances. Parent ref: group-v438549. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1916.903160] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5f26051d-7d5e-41cb-ad16-a6f2bbb42a80 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1916.906998] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12e3a2cb-e4fa-4ecc-b19b-0fec8e26f41f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1916.914469] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a37f878-d3d5-4b4a-acae-79025fcc1792 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1916.918875] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Created folder: Instances in parent group-v438549. [ 1916.919136] env[62824]: DEBUG oslo.service.loopingcall [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1916.919640] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1916.919852] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-11f485c6-614e-49cb-bb8b-54942d66ba87 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1916.962033] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c54c77e-bb4a-4b4d-9973-92fd4f36c795 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1916.965604] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1916.965604] env[62824]: value = "task-2145762" [ 1916.965604] env[62824]: _type = "Task" [ 1916.965604] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1916.972987] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4a70149-489f-4ccd-9801-20b60ff96b8a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1916.980673] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145762, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1916.993021] env[62824]: DEBUG nova.compute.provider_tree [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1917.097143] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1917.097143] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-25b68713-a97a-4e7b-9c1b-e1b16ea27bb4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1917.103091] env[62824]: DEBUG oslo_vmware.api [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] Waiting for the task: (returnval){ [ 1917.103091] env[62824]: value = "task-2145763" [ 1917.103091] env[62824]: _type = "Task" [ 1917.103091] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1917.115476] env[62824]: DEBUG oslo_vmware.api [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] Task: {'id': task-2145763, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1917.179892] env[62824]: DEBUG oslo_vmware.api [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52973cae-2b52-e014-da37-bd501a549e4b, 'name': SearchDatastore_Task, 'duration_secs': 0.019973} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1917.180223] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1917.180459] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: f3740d9a-b21a-4a79-9e28-2a89ecd08bb6] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1917.180696] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1917.180867] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1917.181026] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1917.181290] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-36206c45-b06b-4560-a65b-77d6d6ffced0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1917.194019] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1917.194019] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1917.194019] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7dbaeccf-5417-489d-bc80-4fdb43da9029 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1917.199525] env[62824]: DEBUG oslo_vmware.api [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for the task: (returnval){ [ 1917.199525] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52ae85e7-c6c2-a5f6-6e73-908d5fb087fa" [ 1917.199525] env[62824]: _type = "Task" [ 1917.199525] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1917.208914] env[62824]: DEBUG oslo_vmware.api [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52ae85e7-c6c2-a5f6-6e73-908d5fb087fa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1917.258746] env[62824]: DEBUG nova.compute.manager [req-e287864c-285a-43ce-8dc7-c7752e123803 req-d08fe2c3-cb98-40f6-baef-df47617e110b service nova] [instance: ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5] Received event network-changed-e8535330-ccec-4e36-8f78-cf45232f2a44 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1917.259034] env[62824]: DEBUG nova.compute.manager [req-e287864c-285a-43ce-8dc7-c7752e123803 req-d08fe2c3-cb98-40f6-baef-df47617e110b service nova] [instance: ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5] Refreshing instance network info cache due to event network-changed-e8535330-ccec-4e36-8f78-cf45232f2a44. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1917.259389] env[62824]: DEBUG oslo_concurrency.lockutils [req-e287864c-285a-43ce-8dc7-c7752e123803 req-d08fe2c3-cb98-40f6-baef-df47617e110b service nova] Acquiring lock "refresh_cache-ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1917.259615] env[62824]: DEBUG oslo_concurrency.lockutils [req-e287864c-285a-43ce-8dc7-c7752e123803 req-d08fe2c3-cb98-40f6-baef-df47617e110b service nova] Acquired lock "refresh_cache-ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1917.259900] env[62824]: DEBUG nova.network.neutron [req-e287864c-285a-43ce-8dc7-c7752e123803 req-d08fe2c3-cb98-40f6-baef-df47617e110b service nova] [instance: ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5] Refreshing network info cache for port e8535330-ccec-4e36-8f78-cf45232f2a44 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1917.295443] env[62824]: DEBUG oslo_vmware.api [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2145759, 'name': Rename_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1917.330616] env[62824]: DEBUG oslo_concurrency.lockutils [req-524509d6-75d1-4d50-9b1a-d81882f462b0 req-aa518129-0011-46eb-9f74-7a2cc90cad7d service nova] Releasing lock "refresh_cache-f3740d9a-b21a-4a79-9e28-2a89ecd08bb6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1917.331896] env[62824]: DEBUG nova.compute.manager [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1917.357313] env[62824]: DEBUG nova.virt.hardware [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1917.357482] env[62824]: DEBUG nova.virt.hardware [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1917.357643] env[62824]: DEBUG nova.virt.hardware [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1917.357826] env[62824]: DEBUG nova.virt.hardware [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1917.357982] env[62824]: DEBUG nova.virt.hardware [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1917.358193] env[62824]: DEBUG nova.virt.hardware [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1917.358407] env[62824]: DEBUG nova.virt.hardware [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1917.358600] env[62824]: DEBUG nova.virt.hardware [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1917.358729] env[62824]: DEBUG nova.virt.hardware [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1917.358871] env[62824]: DEBUG nova.virt.hardware [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1917.359072] env[62824]: DEBUG nova.virt.hardware [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1917.359963] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-655c4dd9-2cf6-46fa-86fe-68ed3f5123ea {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1917.368681] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42152404-4da1-4427-9d50-e3aa1177c648 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1917.405549] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0b8e688c-e818-48ea-b84f-ab6b2e8d5cfe tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1917.475644] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145762, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1917.496054] env[62824]: DEBUG nova.scheduler.client.report [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1917.614065] env[62824]: DEBUG oslo_vmware.api [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] Task: {'id': task-2145763, 'name': PowerOffVM_Task, 'duration_secs': 0.156419} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1917.614397] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1917.614689] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1917.615715] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f374014a-b381-405c-8c67-a593415a4d18 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1917.623837] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1917.624183] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a62085d9-5b55-43fa-9dc0-4cafe4e9eda5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1917.650595] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1917.650595] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Deleting contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1917.650799] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] Deleting the datastore file [datastore2] 384a96ea-40ce-43e1-a5f9-82f50b710b0b {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1917.651069] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-11d18e3d-27b6-4805-b62d-124062ad96e4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1917.657356] env[62824]: DEBUG oslo_vmware.api [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] Waiting for the task: (returnval){ [ 1917.657356] env[62824]: value = "task-2145765" [ 1917.657356] env[62824]: _type = "Task" [ 1917.657356] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1917.665676] env[62824]: DEBUG oslo_vmware.api [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] Task: {'id': task-2145765, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1917.712862] env[62824]: DEBUG oslo_vmware.api [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52ae85e7-c6c2-a5f6-6e73-908d5fb087fa, 'name': SearchDatastore_Task, 'duration_secs': 0.012001} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1917.714161] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-52925cec-a422-44cd-bb6b-9a74e1741dad {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1917.720471] env[62824]: DEBUG oslo_vmware.api [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for the task: (returnval){ [ 1917.720471] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52de4528-6a9b-ff0b-d977-3bc1987b414a" [ 1917.720471] env[62824]: _type = "Task" [ 1917.720471] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1917.729013] env[62824]: DEBUG oslo_vmware.api [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52de4528-6a9b-ff0b-d977-3bc1987b414a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1917.791080] env[62824]: DEBUG oslo_vmware.api [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2145759, 'name': Rename_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1917.979289] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145762, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1917.997576] env[62824]: DEBUG nova.network.neutron [req-e287864c-285a-43ce-8dc7-c7752e123803 req-d08fe2c3-cb98-40f6-baef-df47617e110b service nova] [instance: ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5] Updated VIF entry in instance network info cache for port e8535330-ccec-4e36-8f78-cf45232f2a44. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1917.997999] env[62824]: DEBUG nova.network.neutron [req-e287864c-285a-43ce-8dc7-c7752e123803 req-d08fe2c3-cb98-40f6-baef-df47617e110b service nova] [instance: ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5] Updating instance_info_cache with network_info: [{"id": "e8535330-ccec-4e36-8f78-cf45232f2a44", "address": "fa:16:3e:d8:16:53", "network": {"id": "b70367ae-07d9-4a3e-bfbc-4512721997ec", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-1050095803-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "97529775bace4449b194dcef41bea531", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "863474bc-a24a-4823-828c-580a187829e3", "external-id": "nsx-vlan-transportzone-617", "segmentation_id": 617, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape8535330-cc", "ovs_interfaceid": "e8535330-ccec-4e36-8f78-cf45232f2a44", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1918.001632] env[62824]: DEBUG oslo_concurrency.lockutils [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.706s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1918.002136] env[62824]: DEBUG nova.compute.manager [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] [instance: b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1918.004907] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.511s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1918.006430] env[62824]: INFO nova.compute.claims [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1918.172684] env[62824]: DEBUG oslo_vmware.api [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] Task: {'id': task-2145765, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.277903} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1918.173131] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1918.173466] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Deleted contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1918.173812] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1918.232667] env[62824]: DEBUG oslo_vmware.api [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52de4528-6a9b-ff0b-d977-3bc1987b414a, 'name': SearchDatastore_Task, 'duration_secs': 0.046826} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1918.233621] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1918.233883] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] f3740d9a-b21a-4a79-9e28-2a89ecd08bb6/f3740d9a-b21a-4a79-9e28-2a89ecd08bb6.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1918.234156] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-956d8a64-c3be-418b-8af3-0b7668bc12d4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1918.241876] env[62824]: DEBUG oslo_vmware.api [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for the task: (returnval){ [ 1918.241876] env[62824]: value = "task-2145766" [ 1918.241876] env[62824]: _type = "Task" [ 1918.241876] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1918.250780] env[62824]: DEBUG oslo_vmware.api [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2145766, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1918.291777] env[62824]: DEBUG oslo_vmware.api [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2145759, 'name': Rename_Task, 'duration_secs': 1.151589} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1918.292165] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1918.292451] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-13be8696-6585-4ff9-995c-d7af3e918ffa {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1918.300136] env[62824]: DEBUG oslo_vmware.api [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for the task: (returnval){ [ 1918.300136] env[62824]: value = "task-2145767" [ 1918.300136] env[62824]: _type = "Task" [ 1918.300136] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1918.309435] env[62824]: DEBUG oslo_vmware.api [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2145767, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1918.401139] env[62824]: DEBUG nova.compute.manager [req-44059b6d-6c2e-4cdf-8e91-04d3b17e4f46 req-430d7faa-17d1-4118-8096-52a75fe42b97 service nova] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] Received event network-vif-plugged-e0e9725d-7e5a-452f-bd1e-960934705a7f {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1918.401458] env[62824]: DEBUG oslo_concurrency.lockutils [req-44059b6d-6c2e-4cdf-8e91-04d3b17e4f46 req-430d7faa-17d1-4118-8096-52a75fe42b97 service nova] Acquiring lock "16b466c3-c749-4f96-a82c-32dad31138ec-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1918.401762] env[62824]: DEBUG oslo_concurrency.lockutils [req-44059b6d-6c2e-4cdf-8e91-04d3b17e4f46 req-430d7faa-17d1-4118-8096-52a75fe42b97 service nova] Lock "16b466c3-c749-4f96-a82c-32dad31138ec-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1918.402018] env[62824]: DEBUG oslo_concurrency.lockutils [req-44059b6d-6c2e-4cdf-8e91-04d3b17e4f46 req-430d7faa-17d1-4118-8096-52a75fe42b97 service nova] Lock "16b466c3-c749-4f96-a82c-32dad31138ec-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1918.402298] env[62824]: DEBUG nova.compute.manager [req-44059b6d-6c2e-4cdf-8e91-04d3b17e4f46 req-430d7faa-17d1-4118-8096-52a75fe42b97 service nova] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] No waiting events found dispatching network-vif-plugged-e0e9725d-7e5a-452f-bd1e-960934705a7f {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1918.402772] env[62824]: WARNING nova.compute.manager [req-44059b6d-6c2e-4cdf-8e91-04d3b17e4f46 req-430d7faa-17d1-4118-8096-52a75fe42b97 service nova] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] Received unexpected event network-vif-plugged-e0e9725d-7e5a-452f-bd1e-960934705a7f for instance with vm_state building and task_state spawning. [ 1918.478806] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145762, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1918.506335] env[62824]: DEBUG oslo_concurrency.lockutils [req-e287864c-285a-43ce-8dc7-c7752e123803 req-d08fe2c3-cb98-40f6-baef-df47617e110b service nova] Releasing lock "refresh_cache-ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1918.506657] env[62824]: DEBUG nova.compute.manager [req-e287864c-285a-43ce-8dc7-c7752e123803 req-d08fe2c3-cb98-40f6-baef-df47617e110b service nova] [instance: 4dfe9f82-7f89-486b-b159-948656d2896c] Received event network-vif-deleted-4a48d848-c8ab-4cb1-8c18-8dc67595674e {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1918.517737] env[62824]: DEBUG nova.compute.utils [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1918.521204] env[62824]: DEBUG nova.network.neutron [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] Successfully updated port: e0e9725d-7e5a-452f-bd1e-960934705a7f {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1918.522532] env[62824]: DEBUG nova.compute.manager [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] [instance: b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1918.522723] env[62824]: DEBUG nova.network.neutron [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] [instance: b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1918.626551] env[62824]: DEBUG nova.policy [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7f5cbd63b2b34aedb327c1062b688fb3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a871dcf63cf7415a8a2b29d6f5d1136c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 1918.753860] env[62824]: DEBUG oslo_vmware.api [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2145766, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1918.811588] env[62824]: DEBUG oslo_vmware.api [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2145767, 'name': PowerOnVM_Task} progress is 1%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1918.979238] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145762, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1919.020576] env[62824]: DEBUG nova.compute.manager [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] [instance: b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1919.027923] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Acquiring lock "refresh_cache-16b466c3-c749-4f96-a82c-32dad31138ec" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1919.028105] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Acquired lock "refresh_cache-16b466c3-c749-4f96-a82c-32dad31138ec" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1919.028270] env[62824]: DEBUG nova.network.neutron [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1919.229349] env[62824]: DEBUG nova.virt.hardware [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1919.229768] env[62824]: DEBUG nova.virt.hardware [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1919.229768] env[62824]: DEBUG nova.virt.hardware [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1919.229956] env[62824]: DEBUG nova.virt.hardware [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1919.230119] env[62824]: DEBUG nova.virt.hardware [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1919.230265] env[62824]: DEBUG nova.virt.hardware [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1919.230545] env[62824]: DEBUG nova.virt.hardware [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1919.231210] env[62824]: DEBUG nova.virt.hardware [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1919.231356] env[62824]: DEBUG nova.virt.hardware [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1919.231531] env[62824]: DEBUG nova.virt.hardware [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1919.231711] env[62824]: DEBUG nova.virt.hardware [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1919.232623] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68c9843d-0e5f-4f43-bae9-71c7db282833 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1919.251756] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8dcb93f-7a51-4496-b6f0-82b9a425188c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1919.264360] env[62824]: DEBUG oslo_vmware.api [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2145766, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.606983} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1919.276890] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] f3740d9a-b21a-4a79-9e28-2a89ecd08bb6/f3740d9a-b21a-4a79-9e28-2a89ecd08bb6.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1919.277145] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: f3740d9a-b21a-4a79-9e28-2a89ecd08bb6] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1919.277577] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Instance VIF info [] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1919.284081] env[62824]: DEBUG oslo.service.loopingcall [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1919.287168] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b087c3ac-a248-465a-b3dd-2de4862deadf {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1919.289074] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1919.289536] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ff0e115d-4da0-4f1c-a865-20447514c2fc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1919.310010] env[62824]: DEBUG oslo_vmware.api [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for the task: (returnval){ [ 1919.310010] env[62824]: value = "task-2145768" [ 1919.310010] env[62824]: _type = "Task" [ 1919.310010] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1919.319767] env[62824]: DEBUG oslo_vmware.api [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2145767, 'name': PowerOnVM_Task} progress is 1%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1919.319996] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1919.319996] env[62824]: value = "task-2145769" [ 1919.319996] env[62824]: _type = "Task" [ 1919.319996] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1919.329058] env[62824]: DEBUG oslo_vmware.api [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2145768, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1919.329840] env[62824]: DEBUG nova.network.neutron [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] [instance: b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6] Successfully created port: c21b7203-99b2-485d-8a8f-db1f6bcdbbe0 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1919.338767] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145769, 'name': CreateVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1919.482125] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145762, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1919.582604] env[62824]: DEBUG nova.network.neutron [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1919.616928] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61e21a9b-147c-4963-a553-ff1f67f7eeb2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1919.625639] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e4ae1e9-1ac9-47e5-94a1-fdf5d567aaae {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1919.659803] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08d0e0b3-ead2-4be6-8685-b1cc8a92053a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1919.668332] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f95cd97-b1d1-4fdd-94d8-747a32083643 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1919.682575] env[62824]: DEBUG nova.compute.provider_tree [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Updating inventory in ProviderTree for provider bf95157b-a274-42de-9ccf-9851128a44a1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1919.735034] env[62824]: DEBUG nova.network.neutron [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] Updating instance_info_cache with network_info: [{"id": "e0e9725d-7e5a-452f-bd1e-960934705a7f", "address": "fa:16:3e:56:73:22", "network": {"id": "aeacc588-ae73-49ec-9aa1-68f7d92055be", "bridge": "br-int", "label": "tempest-ImagesTestJSON-314094063-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10ee5fe93d584725ace90e76605a1b2e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1e7a4976-597e-4636-990e-6062b5faadee", "external-id": "nsx-vlan-transportzone-847", "segmentation_id": 847, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape0e9725d-7e", "ovs_interfaceid": "e0e9725d-7e5a-452f-bd1e-960934705a7f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1919.812463] env[62824]: DEBUG oslo_vmware.api [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2145767, 'name': PowerOnVM_Task} progress is 1%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1919.821022] env[62824]: DEBUG oslo_vmware.api [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2145768, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.080556} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1919.821022] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: f3740d9a-b21a-4a79-9e28-2a89ecd08bb6] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1919.824160] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6856bdc0-65bc-487d-932f-0a3bd45b74a0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1919.831496] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145769, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1919.849040] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: f3740d9a-b21a-4a79-9e28-2a89ecd08bb6] Reconfiguring VM instance instance-00000019 to attach disk [datastore1] f3740d9a-b21a-4a79-9e28-2a89ecd08bb6/f3740d9a-b21a-4a79-9e28-2a89ecd08bb6.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1919.849417] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9ab8aea2-ac5e-4900-9bc7-66cbe380212d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1919.870033] env[62824]: DEBUG oslo_vmware.api [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for the task: (returnval){ [ 1919.870033] env[62824]: value = "task-2145770" [ 1919.870033] env[62824]: _type = "Task" [ 1919.870033] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1919.878488] env[62824]: DEBUG oslo_vmware.api [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2145770, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1919.982295] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145762, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1920.038719] env[62824]: DEBUG nova.compute.manager [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] [instance: b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1920.064723] env[62824]: DEBUG nova.virt.hardware [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1920.065104] env[62824]: DEBUG nova.virt.hardware [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1920.065289] env[62824]: DEBUG nova.virt.hardware [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1920.065544] env[62824]: DEBUG nova.virt.hardware [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1920.065729] env[62824]: DEBUG nova.virt.hardware [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1920.065929] env[62824]: DEBUG nova.virt.hardware [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1920.066204] env[62824]: DEBUG nova.virt.hardware [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1920.066402] env[62824]: DEBUG nova.virt.hardware [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1920.066616] env[62824]: DEBUG nova.virt.hardware [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1920.066804] env[62824]: DEBUG nova.virt.hardware [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1920.067019] env[62824]: DEBUG nova.virt.hardware [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1920.067883] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c784ddd4-7dd7-43a6-94f2-851ac2ed8f2c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1920.076394] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6e1ba7f-bed4-4701-a8fb-1f58da2664ba {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1920.205389] env[62824]: ERROR nova.scheduler.client.report [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [req-45fa80d6-50a5-4f48-8300-810343083162] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID bf95157b-a274-42de-9ccf-9851128a44a1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-45fa80d6-50a5-4f48-8300-810343083162"}]} [ 1920.224038] env[62824]: DEBUG nova.scheduler.client.report [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Refreshing inventories for resource provider bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1920.237538] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Releasing lock "refresh_cache-16b466c3-c749-4f96-a82c-32dad31138ec" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1920.237948] env[62824]: DEBUG nova.compute.manager [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] Instance network_info: |[{"id": "e0e9725d-7e5a-452f-bd1e-960934705a7f", "address": "fa:16:3e:56:73:22", "network": {"id": "aeacc588-ae73-49ec-9aa1-68f7d92055be", "bridge": "br-int", "label": "tempest-ImagesTestJSON-314094063-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10ee5fe93d584725ace90e76605a1b2e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1e7a4976-597e-4636-990e-6062b5faadee", "external-id": "nsx-vlan-transportzone-847", "segmentation_id": 847, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape0e9725d-7e", "ovs_interfaceid": "e0e9725d-7e5a-452f-bd1e-960934705a7f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1920.238398] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:56:73:22', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1e7a4976-597e-4636-990e-6062b5faadee', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e0e9725d-7e5a-452f-bd1e-960934705a7f', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1920.246576] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Creating folder: Project (10ee5fe93d584725ace90e76605a1b2e). Parent ref: group-v438503. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1920.247960] env[62824]: DEBUG nova.scheduler.client.report [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Updating ProviderTree inventory for provider bf95157b-a274-42de-9ccf-9851128a44a1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1920.248236] env[62824]: DEBUG nova.compute.provider_tree [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Updating inventory in ProviderTree for provider bf95157b-a274-42de-9ccf-9851128a44a1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1920.250543] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-85286fdc-501f-4701-827e-8ea07209a9e5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1920.262151] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Created folder: Project (10ee5fe93d584725ace90e76605a1b2e) in parent group-v438503. [ 1920.262151] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Creating folder: Instances. Parent ref: group-v438553. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1920.262471] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2d58e265-886b-4804-9439-71b0ca30dd55 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1920.265406] env[62824]: DEBUG nova.scheduler.client.report [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Refreshing aggregate associations for resource provider bf95157b-a274-42de-9ccf-9851128a44a1, aggregates: None {{(pid=62824) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1920.275656] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Created folder: Instances in parent group-v438553. [ 1920.275901] env[62824]: DEBUG oslo.service.loopingcall [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1920.276111] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1920.276325] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fa8d311c-4bae-4351-9af6-272239107c5d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1920.293955] env[62824]: DEBUG nova.scheduler.client.report [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Refreshing trait associations for resource provider bf95157b-a274-42de-9ccf-9851128a44a1, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_ISO {{(pid=62824) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1920.301419] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1920.301419] env[62824]: value = "task-2145773" [ 1920.301419] env[62824]: _type = "Task" [ 1920.301419] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1920.318237] env[62824]: DEBUG oslo_vmware.api [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2145767, 'name': PowerOnVM_Task} progress is 1%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1920.322557] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145773, 'name': CreateVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1920.333955] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145769, 'name': CreateVM_Task, 'duration_secs': 0.765429} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1920.335022] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1920.335022] env[62824]: DEBUG oslo_concurrency.lockutils [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1920.335022] env[62824]: DEBUG oslo_concurrency.lockutils [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1920.335187] env[62824]: DEBUG oslo_concurrency.lockutils [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] Acquired external semaphore "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1920.335356] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8e76b6fa-09a9-4f0f-9f4b-8d1a2ded5fc3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1920.340814] env[62824]: DEBUG oslo_vmware.api [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] Waiting for the task: (returnval){ [ 1920.340814] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5242ce5a-28e5-894c-0964-ec8c04cd841a" [ 1920.340814] env[62824]: _type = "Task" [ 1920.340814] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1920.352447] env[62824]: DEBUG oslo_vmware.api [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5242ce5a-28e5-894c-0964-ec8c04cd841a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1920.382083] env[62824]: DEBUG oslo_vmware.api [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2145770, 'name': ReconfigVM_Task, 'duration_secs': 0.50378} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1920.382582] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: f3740d9a-b21a-4a79-9e28-2a89ecd08bb6] Reconfigured VM instance instance-00000019 to attach disk [datastore1] f3740d9a-b21a-4a79-9e28-2a89ecd08bb6/f3740d9a-b21a-4a79-9e28-2a89ecd08bb6.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1920.386338] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5e1e13d9-b911-40e7-93f7-a5cd5be4f859 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1920.394248] env[62824]: DEBUG oslo_vmware.api [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for the task: (returnval){ [ 1920.394248] env[62824]: value = "task-2145774" [ 1920.394248] env[62824]: _type = "Task" [ 1920.394248] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1920.403779] env[62824]: DEBUG oslo_vmware.api [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2145774, 'name': Rename_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1920.446983] env[62824]: DEBUG nova.compute.manager [req-130d3a1d-0bcb-4828-ac79-58de9445c8f5 req-9216fd33-4c7c-4368-a5b5-85dbaf88ec41 service nova] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] Received event network-changed-e0e9725d-7e5a-452f-bd1e-960934705a7f {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1920.446983] env[62824]: DEBUG nova.compute.manager [req-130d3a1d-0bcb-4828-ac79-58de9445c8f5 req-9216fd33-4c7c-4368-a5b5-85dbaf88ec41 service nova] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] Refreshing instance network info cache due to event network-changed-e0e9725d-7e5a-452f-bd1e-960934705a7f. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1920.447210] env[62824]: DEBUG oslo_concurrency.lockutils [req-130d3a1d-0bcb-4828-ac79-58de9445c8f5 req-9216fd33-4c7c-4368-a5b5-85dbaf88ec41 service nova] Acquiring lock "refresh_cache-16b466c3-c749-4f96-a82c-32dad31138ec" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1920.447449] env[62824]: DEBUG oslo_concurrency.lockutils [req-130d3a1d-0bcb-4828-ac79-58de9445c8f5 req-9216fd33-4c7c-4368-a5b5-85dbaf88ec41 service nova] Acquired lock "refresh_cache-16b466c3-c749-4f96-a82c-32dad31138ec" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1920.447754] env[62824]: DEBUG nova.network.neutron [req-130d3a1d-0bcb-4828-ac79-58de9445c8f5 req-9216fd33-4c7c-4368-a5b5-85dbaf88ec41 service nova] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] Refreshing network info cache for port e0e9725d-7e5a-452f-bd1e-960934705a7f {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1920.485762] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145762, 'name': CreateVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1920.792395] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-109dcb72-8a0f-4391-b2da-d3e53ee3afcd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1920.800393] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b0254b1-74da-4a11-8c18-8affd9ad7c71 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1920.840536] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145773, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1920.841686] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0d6d821-7c85-4128-bad0-9a62cd82e11f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1920.850058] env[62824]: DEBUG oslo_vmware.api [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2145767, 'name': PowerOnVM_Task} progress is 1%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1920.858386] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89663cc6-c031-49df-9d6d-133d5ebf3e4b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1920.862150] env[62824]: DEBUG oslo_vmware.api [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5242ce5a-28e5-894c-0964-ec8c04cd841a, 'name': SearchDatastore_Task, 'duration_secs': 0.014426} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1920.862484] env[62824]: DEBUG oslo_concurrency.lockutils [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1920.863065] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1920.863319] env[62824]: DEBUG oslo_concurrency.lockutils [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1920.863853] env[62824]: DEBUG oslo_concurrency.lockutils [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1920.864032] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1920.864676] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c797efa1-9c89-469c-9157-81289ec4182d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1920.874608] env[62824]: DEBUG nova.compute.provider_tree [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Updating inventory in ProviderTree for provider bf95157b-a274-42de-9ccf-9851128a44a1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1920.888142] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1920.888331] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1920.889605] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d30d3ab9-cae9-4ff1-8691-9c579633b3eb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1920.894234] env[62824]: DEBUG oslo_vmware.api [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] Waiting for the task: (returnval){ [ 1920.894234] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52f580eb-9d18-cb4e-21c9-d81f5303d074" [ 1920.894234] env[62824]: _type = "Task" [ 1920.894234] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1920.904494] env[62824]: DEBUG oslo_vmware.api [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2145774, 'name': Rename_Task, 'duration_secs': 0.284132} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1920.907316] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: f3740d9a-b21a-4a79-9e28-2a89ecd08bb6] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1920.907560] env[62824]: DEBUG oslo_vmware.api [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52f580eb-9d18-cb4e-21c9-d81f5303d074, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1920.907755] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b3a4ceb0-e5de-484c-8880-47d8ca8b1e38 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1920.913556] env[62824]: DEBUG oslo_vmware.api [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for the task: (returnval){ [ 1920.913556] env[62824]: value = "task-2145775" [ 1920.913556] env[62824]: _type = "Task" [ 1920.913556] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1920.920750] env[62824]: DEBUG oslo_vmware.api [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2145775, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1920.983374] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145762, 'name': CreateVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1921.181765] env[62824]: DEBUG nova.network.neutron [req-130d3a1d-0bcb-4828-ac79-58de9445c8f5 req-9216fd33-4c7c-4368-a5b5-85dbaf88ec41 service nova] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] Updated VIF entry in instance network info cache for port e0e9725d-7e5a-452f-bd1e-960934705a7f. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1921.181765] env[62824]: DEBUG nova.network.neutron [req-130d3a1d-0bcb-4828-ac79-58de9445c8f5 req-9216fd33-4c7c-4368-a5b5-85dbaf88ec41 service nova] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] Updating instance_info_cache with network_info: [{"id": "e0e9725d-7e5a-452f-bd1e-960934705a7f", "address": "fa:16:3e:56:73:22", "network": {"id": "aeacc588-ae73-49ec-9aa1-68f7d92055be", "bridge": "br-int", "label": "tempest-ImagesTestJSON-314094063-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10ee5fe93d584725ace90e76605a1b2e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1e7a4976-597e-4636-990e-6062b5faadee", "external-id": "nsx-vlan-transportzone-847", "segmentation_id": 847, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape0e9725d-7e", "ovs_interfaceid": "e0e9725d-7e5a-452f-bd1e-960934705a7f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1921.183814] env[62824]: DEBUG nova.network.neutron [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] [instance: b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6] Successfully updated port: c21b7203-99b2-485d-8a8f-db1f6bcdbbe0 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1921.320219] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145773, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1921.320396] env[62824]: DEBUG oslo_vmware.api [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2145767, 'name': PowerOnVM_Task} progress is 1%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1921.409776] env[62824]: DEBUG oslo_vmware.api [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52f580eb-9d18-cb4e-21c9-d81f5303d074, 'name': SearchDatastore_Task, 'duration_secs': 0.025559} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1921.410900] env[62824]: DEBUG nova.scheduler.client.report [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Updated inventory for provider bf95157b-a274-42de-9ccf-9851128a44a1 with generation 55 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 1921.411218] env[62824]: DEBUG nova.compute.provider_tree [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Updating resource provider bf95157b-a274-42de-9ccf-9851128a44a1 generation from 55 to 56 during operation: update_inventory {{(pid=62824) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1921.411412] env[62824]: DEBUG nova.compute.provider_tree [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Updating inventory in ProviderTree for provider bf95157b-a274-42de-9ccf-9851128a44a1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1921.417080] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f2ddd939-447f-4530-9182-3ee0a8de5c94 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1921.427471] env[62824]: DEBUG oslo_vmware.api [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] Waiting for the task: (returnval){ [ 1921.427471] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5205ae38-1116-a367-ffe3-ce06eca86464" [ 1921.427471] env[62824]: _type = "Task" [ 1921.427471] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1921.431824] env[62824]: DEBUG oslo_vmware.api [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2145775, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1921.440818] env[62824]: DEBUG oslo_vmware.api [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5205ae38-1116-a367-ffe3-ce06eca86464, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1921.487749] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145762, 'name': CreateVM_Task, 'duration_secs': 4.490838} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1921.487749] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1921.487978] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1921.488056] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1921.488326] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1921.488588] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9c32f857-b750-41d7-9a63-562407857643 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1921.493287] env[62824]: DEBUG oslo_vmware.api [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Waiting for the task: (returnval){ [ 1921.493287] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5269052b-87d1-9ce5-d01a-c884c08709ee" [ 1921.493287] env[62824]: _type = "Task" [ 1921.493287] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1921.501419] env[62824]: DEBUG oslo_vmware.api [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5269052b-87d1-9ce5-d01a-c884c08709ee, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1921.686480] env[62824]: DEBUG oslo_concurrency.lockutils [req-130d3a1d-0bcb-4828-ac79-58de9445c8f5 req-9216fd33-4c7c-4368-a5b5-85dbaf88ec41 service nova] Releasing lock "refresh_cache-16b466c3-c749-4f96-a82c-32dad31138ec" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1921.690524] env[62824]: DEBUG oslo_concurrency.lockutils [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Acquiring lock "refresh_cache-b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1921.690724] env[62824]: DEBUG oslo_concurrency.lockutils [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Acquired lock "refresh_cache-b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1921.690943] env[62824]: DEBUG nova.network.neutron [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] [instance: b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1921.817861] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145773, 'name': CreateVM_Task, 'duration_secs': 1.251092} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1921.820954] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1921.821266] env[62824]: DEBUG oslo_vmware.api [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2145767, 'name': PowerOnVM_Task} progress is 64%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1921.821892] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1921.916917] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.912s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1921.917470] env[62824]: DEBUG nova.compute.manager [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1921.920289] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7f338944-9cb1-4d55-9d22-6a91d729ee9f tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.621s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1921.920509] env[62824]: DEBUG nova.objects.instance [None req-7f338944-9cb1-4d55-9d22-6a91d729ee9f tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Lazy-loading 'resources' on Instance uuid 517bf2cf-e142-4f36-bf53-79a8d2d8a530 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1921.931965] env[62824]: DEBUG oslo_vmware.api [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2145775, 'name': PowerOnVM_Task, 'duration_secs': 0.78412} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1921.932808] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: f3740d9a-b21a-4a79-9e28-2a89ecd08bb6] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1921.932808] env[62824]: INFO nova.compute.manager [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: f3740d9a-b21a-4a79-9e28-2a89ecd08bb6] Took 10.07 seconds to spawn the instance on the hypervisor. [ 1921.933022] env[62824]: DEBUG nova.compute.manager [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: f3740d9a-b21a-4a79-9e28-2a89ecd08bb6] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1921.937031] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f30fd94d-69d4-4491-9508-e55e7ddcceef {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1921.946381] env[62824]: DEBUG oslo_vmware.api [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5205ae38-1116-a367-ffe3-ce06eca86464, 'name': SearchDatastore_Task, 'duration_secs': 0.022576} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1921.948014] env[62824]: DEBUG oslo_concurrency.lockutils [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1921.948345] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] Copying Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 384a96ea-40ce-43e1-a5f9-82f50b710b0b/384a96ea-40ce-43e1-a5f9-82f50b710b0b.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1921.952284] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-de2d3e88-de07-48ad-ae0f-fd1ebd7739e5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1921.959053] env[62824]: DEBUG oslo_vmware.api [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] Waiting for the task: (returnval){ [ 1921.959053] env[62824]: value = "task-2145776" [ 1921.959053] env[62824]: _type = "Task" [ 1921.959053] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1921.967418] env[62824]: DEBUG oslo_vmware.api [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] Task: {'id': task-2145776, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1922.004899] env[62824]: DEBUG oslo_vmware.api [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5269052b-87d1-9ce5-d01a-c884c08709ee, 'name': SearchDatastore_Task, 'duration_secs': 0.00977} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1922.005271] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1922.005645] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] [instance: ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1922.005971] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1922.006177] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1922.006396] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1922.006722] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1922.007099] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1922.007628] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-599534c0-0e0a-430a-8b8c-27968ec52f3b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1922.010023] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d8f25776-724e-4592-87c0-6af8fdc2e158 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1922.015063] env[62824]: DEBUG oslo_vmware.api [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the task: (returnval){ [ 1922.015063] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]521b91ba-198b-5f63-85c7-bf569f8156a3" [ 1922.015063] env[62824]: _type = "Task" [ 1922.015063] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1922.027152] env[62824]: DEBUG oslo_vmware.api [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]521b91ba-198b-5f63-85c7-bf569f8156a3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1922.027152] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1922.027152] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1922.027644] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-35ec46a8-f6d6-4680-a346-46f322f03677 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1922.032428] env[62824]: DEBUG oslo_vmware.api [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Waiting for the task: (returnval){ [ 1922.032428] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52ad1afb-e602-e120-2a3b-6d22540365d8" [ 1922.032428] env[62824]: _type = "Task" [ 1922.032428] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1922.041158] env[62824]: DEBUG oslo_vmware.api [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52ad1afb-e602-e120-2a3b-6d22540365d8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1922.274292] env[62824]: DEBUG nova.network.neutron [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] [instance: b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1922.324177] env[62824]: DEBUG oslo_vmware.api [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2145767, 'name': PowerOnVM_Task} progress is 64%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1922.424057] env[62824]: DEBUG nova.compute.utils [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1922.428289] env[62824]: DEBUG nova.compute.manager [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1922.428483] env[62824]: DEBUG nova.network.neutron [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1922.465078] env[62824]: INFO nova.compute.manager [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: f3740d9a-b21a-4a79-9e28-2a89ecd08bb6] Took 33.27 seconds to build instance. [ 1922.475348] env[62824]: DEBUG oslo_vmware.api [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] Task: {'id': task-2145776, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.455327} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1922.475609] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] Copied Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 384a96ea-40ce-43e1-a5f9-82f50b710b0b/384a96ea-40ce-43e1-a5f9-82f50b710b0b.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1922.475813] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1922.476078] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ea2170f2-1c28-4862-92bd-db781a78686b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1922.482554] env[62824]: DEBUG oslo_vmware.api [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] Waiting for the task: (returnval){ [ 1922.482554] env[62824]: value = "task-2145777" [ 1922.482554] env[62824]: _type = "Task" [ 1922.482554] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1922.486114] env[62824]: DEBUG nova.policy [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7b85d4201e7946e6b5b40a6ab8a22d61', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3240be4f32254f54aa87cb045e3d8f31', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 1922.496269] env[62824]: DEBUG oslo_vmware.api [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] Task: {'id': task-2145777, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1922.527680] env[62824]: DEBUG oslo_vmware.api [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]521b91ba-198b-5f63-85c7-bf569f8156a3, 'name': SearchDatastore_Task, 'duration_secs': 0.018231} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1922.530361] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1922.530981] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1922.531220] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1922.546343] env[62824]: DEBUG oslo_vmware.api [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52ad1afb-e602-e120-2a3b-6d22540365d8, 'name': SearchDatastore_Task, 'duration_secs': 0.008035} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1922.550521] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1e0de6f2-650e-4bcc-99da-3916cc23965d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1922.558292] env[62824]: DEBUG oslo_vmware.api [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Waiting for the task: (returnval){ [ 1922.558292] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]521bf644-e217-e0ff-5817-42597ac10164" [ 1922.558292] env[62824]: _type = "Task" [ 1922.558292] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1922.578729] env[62824]: DEBUG oslo_vmware.api [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]521bf644-e217-e0ff-5817-42597ac10164, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1922.590795] env[62824]: DEBUG nova.network.neutron [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] [instance: b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6] Updating instance_info_cache with network_info: [{"id": "c21b7203-99b2-485d-8a8f-db1f6bcdbbe0", "address": "fa:16:3e:62:46:77", "network": {"id": "b6403dc3-30ef-4c76-b521-380c179f345d", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.200", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "22c5c40c2e004babaa15896c89307303", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1002b79b-224e-41e3-a484-4245a767147a", "external-id": "nsx-vlan-transportzone-353", "segmentation_id": 353, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc21b7203-99", "ovs_interfaceid": "c21b7203-99b2-485d-8a8f-db1f6bcdbbe0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1922.644387] env[62824]: DEBUG nova.compute.manager [req-626bcab9-7daf-4541-9b77-4e6260023b5b req-955364b2-d588-4095-b0c1-d1199794f497 service nova] [instance: b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6] Received event network-vif-plugged-c21b7203-99b2-485d-8a8f-db1f6bcdbbe0 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1922.644683] env[62824]: DEBUG oslo_concurrency.lockutils [req-626bcab9-7daf-4541-9b77-4e6260023b5b req-955364b2-d588-4095-b0c1-d1199794f497 service nova] Acquiring lock "b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1922.644819] env[62824]: DEBUG oslo_concurrency.lockutils [req-626bcab9-7daf-4541-9b77-4e6260023b5b req-955364b2-d588-4095-b0c1-d1199794f497 service nova] Lock "b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1922.644985] env[62824]: DEBUG oslo_concurrency.lockutils [req-626bcab9-7daf-4541-9b77-4e6260023b5b req-955364b2-d588-4095-b0c1-d1199794f497 service nova] Lock "b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1922.645169] env[62824]: DEBUG nova.compute.manager [req-626bcab9-7daf-4541-9b77-4e6260023b5b req-955364b2-d588-4095-b0c1-d1199794f497 service nova] [instance: b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6] No waiting events found dispatching network-vif-plugged-c21b7203-99b2-485d-8a8f-db1f6bcdbbe0 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1922.645330] env[62824]: WARNING nova.compute.manager [req-626bcab9-7daf-4541-9b77-4e6260023b5b req-955364b2-d588-4095-b0c1-d1199794f497 service nova] [instance: b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6] Received unexpected event network-vif-plugged-c21b7203-99b2-485d-8a8f-db1f6bcdbbe0 for instance with vm_state building and task_state spawning. [ 1922.645484] env[62824]: DEBUG nova.compute.manager [req-626bcab9-7daf-4541-9b77-4e6260023b5b req-955364b2-d588-4095-b0c1-d1199794f497 service nova] [instance: b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6] Received event network-changed-c21b7203-99b2-485d-8a8f-db1f6bcdbbe0 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1922.645665] env[62824]: DEBUG nova.compute.manager [req-626bcab9-7daf-4541-9b77-4e6260023b5b req-955364b2-d588-4095-b0c1-d1199794f497 service nova] [instance: b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6] Refreshing instance network info cache due to event network-changed-c21b7203-99b2-485d-8a8f-db1f6bcdbbe0. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1922.645842] env[62824]: DEBUG oslo_concurrency.lockutils [req-626bcab9-7daf-4541-9b77-4e6260023b5b req-955364b2-d588-4095-b0c1-d1199794f497 service nova] Acquiring lock "refresh_cache-b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1922.802298] env[62824]: DEBUG nova.network.neutron [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Successfully created port: 60a4fce0-0e63-45af-890c-46ca44ea9a0c {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1922.823034] env[62824]: DEBUG oslo_vmware.api [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2145767, 'name': PowerOnVM_Task} progress is 75%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1922.933259] env[62824]: DEBUG nova.compute.manager [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1922.969670] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a704eff0-a1e9-4504-98a8-5f78ad2fac5a tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Lock "f3740d9a-b21a-4a79-9e28-2a89ecd08bb6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 59.329s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1922.998969] env[62824]: DEBUG oslo_vmware.api [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] Task: {'id': task-2145777, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.081107} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1923.001875] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1923.001875] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd8d1b69-47c8-477d-9769-039108249b87 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1923.004030] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d4a6519-fc26-460e-9348-0471a5a60759 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1923.019445] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86414239-14bc-4d57-b87f-646c63f830d3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1923.033458] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Reconfiguring VM instance instance-00000014 to attach disk [datastore2] 384a96ea-40ce-43e1-a5f9-82f50b710b0b/384a96ea-40ce-43e1-a5f9-82f50b710b0b.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1923.033800] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d2e539f2-03c3-4f63-9ebd-93eb30e5dca6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1923.081710] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba501cff-28ab-4dec-be54-9f2331bcbdee {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1923.085128] env[62824]: DEBUG oslo_vmware.api [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] Waiting for the task: (returnval){ [ 1923.085128] env[62824]: value = "task-2145778" [ 1923.085128] env[62824]: _type = "Task" [ 1923.085128] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1923.095293] env[62824]: DEBUG oslo_concurrency.lockutils [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Releasing lock "refresh_cache-b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1923.095293] env[62824]: DEBUG nova.compute.manager [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] [instance: b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6] Instance network_info: |[{"id": "c21b7203-99b2-485d-8a8f-db1f6bcdbbe0", "address": "fa:16:3e:62:46:77", "network": {"id": "b6403dc3-30ef-4c76-b521-380c179f345d", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.200", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "22c5c40c2e004babaa15896c89307303", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1002b79b-224e-41e3-a484-4245a767147a", "external-id": "nsx-vlan-transportzone-353", "segmentation_id": 353, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc21b7203-99", "ovs_interfaceid": "c21b7203-99b2-485d-8a8f-db1f6bcdbbe0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1923.095413] env[62824]: DEBUG oslo_vmware.api [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]521bf644-e217-e0ff-5817-42597ac10164, 'name': SearchDatastore_Task, 'duration_secs': 0.040084} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1923.097112] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fe4db2e-aab7-40a6-b989-74a2d631c38f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1923.102647] env[62824]: DEBUG oslo_concurrency.lockutils [req-626bcab9-7daf-4541-9b77-4e6260023b5b req-955364b2-d588-4095-b0c1-d1199794f497 service nova] Acquired lock "refresh_cache-b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1923.102845] env[62824]: DEBUG nova.network.neutron [req-626bcab9-7daf-4541-9b77-4e6260023b5b req-955364b2-d588-4095-b0c1-d1199794f497 service nova] [instance: b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6] Refreshing network info cache for port c21b7203-99b2-485d-8a8f-db1f6bcdbbe0 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1923.104241] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] [instance: b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:62:46:77', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1002b79b-224e-41e3-a484-4245a767147a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c21b7203-99b2-485d-8a8f-db1f6bcdbbe0', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1923.112267] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Creating folder: Project (a871dcf63cf7415a8a2b29d6f5d1136c). Parent ref: group-v438503. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1923.112782] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1923.115024] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5/ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1923.117685] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5de9749f-b1ce-42ea-9a85-1d2a45cb6b35 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1923.119552] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1923.119676] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1923.119933] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-037b4ba8-16db-47a5-b3cb-4ad66024817b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1923.122494] env[62824]: DEBUG oslo_vmware.api [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] Task: {'id': task-2145778, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1923.123435] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7109c48e-0e86-44c1-bde7-25db74ff1dc2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1923.136598] env[62824]: DEBUG nova.compute.provider_tree [None req-7f338944-9cb1-4d55-9d22-6a91d729ee9f tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Updating inventory in ProviderTree for provider bf95157b-a274-42de-9ccf-9851128a44a1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1923.141485] env[62824]: DEBUG oslo_vmware.api [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Waiting for the task: (returnval){ [ 1923.141485] env[62824]: value = "task-2145780" [ 1923.141485] env[62824]: _type = "Task" [ 1923.141485] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1923.145706] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Created folder: Project (a871dcf63cf7415a8a2b29d6f5d1136c) in parent group-v438503. [ 1923.145798] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Creating folder: Instances. Parent ref: group-v438556. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1923.146880] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5091d12a-562d-49cf-b784-0a96cb5e20ae {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1923.156245] env[62824]: DEBUG oslo_vmware.api [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Task: {'id': task-2145780, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1923.157642] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1923.157891] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1923.158933] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c9c18f98-9e7d-4704-993b-4bca6ff41b3b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1923.165212] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Created folder: Instances in parent group-v438556. [ 1923.165687] env[62824]: DEBUG oslo.service.loopingcall [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1923.169168] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1923.169168] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f46cbd6d-85f8-47a1-b67f-a76ea272232f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1923.184533] env[62824]: DEBUG oslo_vmware.api [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the task: (returnval){ [ 1923.184533] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]521c80e2-2352-85fe-343e-156ecb9407e1" [ 1923.184533] env[62824]: _type = "Task" [ 1923.184533] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1923.190459] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1923.190459] env[62824]: value = "task-2145782" [ 1923.190459] env[62824]: _type = "Task" [ 1923.190459] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1923.198143] env[62824]: DEBUG oslo_vmware.api [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]521c80e2-2352-85fe-343e-156ecb9407e1, 'name': SearchDatastore_Task, 'duration_secs': 0.013619} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1923.199301] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-737c2af0-ee88-463a-940f-0dc5e4eda81d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1923.205995] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145782, 'name': CreateVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1923.209744] env[62824]: DEBUG oslo_vmware.api [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the task: (returnval){ [ 1923.209744] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]527b03c0-136c-c487-481e-8d796c05f726" [ 1923.209744] env[62824]: _type = "Task" [ 1923.209744] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1923.217111] env[62824]: DEBUG oslo_vmware.api [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]527b03c0-136c-c487-481e-8d796c05f726, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1923.323235] env[62824]: DEBUG oslo_vmware.api [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2145767, 'name': PowerOnVM_Task, 'duration_secs': 4.876671} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1923.323555] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1923.323763] env[62824]: INFO nova.compute.manager [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Took 14.19 seconds to spawn the instance on the hypervisor. [ 1923.323940] env[62824]: DEBUG nova.compute.manager [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1923.324792] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d05b9242-fff6-4a9c-9cb1-d3bc2a9b937d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1923.473610] env[62824]: DEBUG nova.compute.manager [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1923.598066] env[62824]: DEBUG oslo_vmware.api [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] Task: {'id': task-2145778, 'name': ReconfigVM_Task, 'duration_secs': 0.322275} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1923.598396] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Reconfigured VM instance instance-00000014 to attach disk [datastore2] 384a96ea-40ce-43e1-a5f9-82f50b710b0b/384a96ea-40ce-43e1-a5f9-82f50b710b0b.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1923.599078] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-df41c655-47b0-444f-b21c-67d6af635155 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1923.605729] env[62824]: DEBUG oslo_vmware.api [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] Waiting for the task: (returnval){ [ 1923.605729] env[62824]: value = "task-2145783" [ 1923.605729] env[62824]: _type = "Task" [ 1923.605729] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1923.616430] env[62824]: DEBUG oslo_vmware.api [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] Task: {'id': task-2145783, 'name': Rename_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1923.651757] env[62824]: DEBUG oslo_vmware.api [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Task: {'id': task-2145780, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1923.678128] env[62824]: DEBUG nova.scheduler.client.report [None req-7f338944-9cb1-4d55-9d22-6a91d729ee9f tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Updated inventory for provider bf95157b-a274-42de-9ccf-9851128a44a1 with generation 56 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 1923.678422] env[62824]: DEBUG nova.compute.provider_tree [None req-7f338944-9cb1-4d55-9d22-6a91d729ee9f tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Updating resource provider bf95157b-a274-42de-9ccf-9851128a44a1 generation from 56 to 57 during operation: update_inventory {{(pid=62824) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1923.678608] env[62824]: DEBUG nova.compute.provider_tree [None req-7f338944-9cb1-4d55-9d22-6a91d729ee9f tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Updating inventory in ProviderTree for provider bf95157b-a274-42de-9ccf-9851128a44a1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1923.701935] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145782, 'name': CreateVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1923.720268] env[62824]: DEBUG oslo_vmware.api [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]527b03c0-136c-c487-481e-8d796c05f726, 'name': SearchDatastore_Task, 'duration_secs': 0.014722} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1923.720575] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1923.720845] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 16b466c3-c749-4f96-a82c-32dad31138ec/16b466c3-c749-4f96-a82c-32dad31138ec.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1923.721129] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-05dd431b-3e0b-43bf-80f8-8f80f417a6cf {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1923.728503] env[62824]: DEBUG oslo_vmware.api [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the task: (returnval){ [ 1923.728503] env[62824]: value = "task-2145784" [ 1923.728503] env[62824]: _type = "Task" [ 1923.728503] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1923.740892] env[62824]: DEBUG oslo_vmware.api [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2145784, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1923.842968] env[62824]: INFO nova.compute.manager [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Took 37.86 seconds to build instance. [ 1923.944721] env[62824]: DEBUG nova.compute.manager [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1923.970675] env[62824]: DEBUG nova.virt.hardware [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1923.970920] env[62824]: DEBUG nova.virt.hardware [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1923.971098] env[62824]: DEBUG nova.virt.hardware [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1923.971288] env[62824]: DEBUG nova.virt.hardware [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1923.971437] env[62824]: DEBUG nova.virt.hardware [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1923.971593] env[62824]: DEBUG nova.virt.hardware [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1923.971816] env[62824]: DEBUG nova.virt.hardware [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1923.971976] env[62824]: DEBUG nova.virt.hardware [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1923.972170] env[62824]: DEBUG nova.virt.hardware [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1923.972337] env[62824]: DEBUG nova.virt.hardware [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1923.972561] env[62824]: DEBUG nova.virt.hardware [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1923.973415] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54435a68-e2f7-497e-8a51-f5fdc2e727f3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1923.984391] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e35fcc96-de0c-4b61-a55c-9002dbe79c07 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1924.003472] env[62824]: DEBUG oslo_concurrency.lockutils [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1924.116541] env[62824]: DEBUG oslo_vmware.api [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] Task: {'id': task-2145783, 'name': Rename_Task, 'duration_secs': 0.176937} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1924.116830] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1924.117093] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4a2ea3a1-2a8a-4551-9320-d862a08f1f32 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1924.125880] env[62824]: DEBUG oslo_vmware.api [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] Waiting for the task: (returnval){ [ 1924.125880] env[62824]: value = "task-2145785" [ 1924.125880] env[62824]: _type = "Task" [ 1924.125880] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1924.135678] env[62824]: DEBUG oslo_vmware.api [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] Task: {'id': task-2145785, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1924.153416] env[62824]: DEBUG oslo_vmware.api [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Task: {'id': task-2145780, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.622895} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1924.153777] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5/ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1924.154008] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] [instance: ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1924.154460] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-129ce710-9434-4739-887f-0a73fe75078c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1924.161285] env[62824]: DEBUG oslo_vmware.api [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Waiting for the task: (returnval){ [ 1924.161285] env[62824]: value = "task-2145786" [ 1924.161285] env[62824]: _type = "Task" [ 1924.161285] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1924.171226] env[62824]: DEBUG oslo_vmware.api [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Task: {'id': task-2145786, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1924.184514] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7f338944-9cb1-4d55-9d22-6a91d729ee9f tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.264s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1924.188318] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.720s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1924.189192] env[62824]: INFO nova.compute.claims [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] [instance: c36ece43-3d70-4e67-a740-9057f413c722] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1924.202083] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145782, 'name': CreateVM_Task, 'duration_secs': 0.559927} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1924.202083] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1924.203409] env[62824]: DEBUG oslo_concurrency.lockutils [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1924.203409] env[62824]: DEBUG oslo_concurrency.lockutils [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1924.203409] env[62824]: DEBUG oslo_concurrency.lockutils [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1924.203883] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-786133bd-d115-4221-9ac5-874f6294a1e9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1924.209099] env[62824]: DEBUG oslo_vmware.api [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Waiting for the task: (returnval){ [ 1924.209099] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]528e4405-048c-f90f-0612-7d6491f8a8be" [ 1924.209099] env[62824]: _type = "Task" [ 1924.209099] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1924.213969] env[62824]: INFO nova.scheduler.client.report [None req-7f338944-9cb1-4d55-9d22-6a91d729ee9f tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Deleted allocations for instance 517bf2cf-e142-4f36-bf53-79a8d2d8a530 [ 1924.222631] env[62824]: DEBUG oslo_vmware.api [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]528e4405-048c-f90f-0612-7d6491f8a8be, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1924.223777] env[62824]: DEBUG nova.network.neutron [req-626bcab9-7daf-4541-9b77-4e6260023b5b req-955364b2-d588-4095-b0c1-d1199794f497 service nova] [instance: b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6] Updated VIF entry in instance network info cache for port c21b7203-99b2-485d-8a8f-db1f6bcdbbe0. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1924.224135] env[62824]: DEBUG nova.network.neutron [req-626bcab9-7daf-4541-9b77-4e6260023b5b req-955364b2-d588-4095-b0c1-d1199794f497 service nova] [instance: b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6] Updating instance_info_cache with network_info: [{"id": "c21b7203-99b2-485d-8a8f-db1f6bcdbbe0", "address": "fa:16:3e:62:46:77", "network": {"id": "b6403dc3-30ef-4c76-b521-380c179f345d", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.200", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "22c5c40c2e004babaa15896c89307303", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1002b79b-224e-41e3-a484-4245a767147a", "external-id": "nsx-vlan-transportzone-353", "segmentation_id": 353, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc21b7203-99", "ovs_interfaceid": "c21b7203-99b2-485d-8a8f-db1f6bcdbbe0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1924.238687] env[62824]: DEBUG oslo_vmware.api [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2145784, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1924.345619] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7c8943c9-e568-4a3a-8b38-e8347d457c38 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Lock "07ce931f-45ef-409b-b714-9f1cd47a3a88" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 62.452s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1924.446598] env[62824]: DEBUG nova.network.neutron [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Successfully updated port: 60a4fce0-0e63-45af-890c-46ca44ea9a0c {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1924.606338] env[62824]: DEBUG nova.compute.manager [req-d298a5a6-9df6-4e6d-9354-b737db917387 req-a605b34f-c825-45ab-91c1-7705a634bbc9 service nova] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Received event network-vif-plugged-60a4fce0-0e63-45af-890c-46ca44ea9a0c {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1924.606556] env[62824]: DEBUG oslo_concurrency.lockutils [req-d298a5a6-9df6-4e6d-9354-b737db917387 req-a605b34f-c825-45ab-91c1-7705a634bbc9 service nova] Acquiring lock "b8cc8cd3-ea03-40bf-b867-7ad193365552-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1924.606763] env[62824]: DEBUG oslo_concurrency.lockutils [req-d298a5a6-9df6-4e6d-9354-b737db917387 req-a605b34f-c825-45ab-91c1-7705a634bbc9 service nova] Lock "b8cc8cd3-ea03-40bf-b867-7ad193365552-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1924.606943] env[62824]: DEBUG oslo_concurrency.lockutils [req-d298a5a6-9df6-4e6d-9354-b737db917387 req-a605b34f-c825-45ab-91c1-7705a634bbc9 service nova] Lock "b8cc8cd3-ea03-40bf-b867-7ad193365552-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1924.607126] env[62824]: DEBUG nova.compute.manager [req-d298a5a6-9df6-4e6d-9354-b737db917387 req-a605b34f-c825-45ab-91c1-7705a634bbc9 service nova] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] No waiting events found dispatching network-vif-plugged-60a4fce0-0e63-45af-890c-46ca44ea9a0c {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1924.607292] env[62824]: WARNING nova.compute.manager [req-d298a5a6-9df6-4e6d-9354-b737db917387 req-a605b34f-c825-45ab-91c1-7705a634bbc9 service nova] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Received unexpected event network-vif-plugged-60a4fce0-0e63-45af-890c-46ca44ea9a0c for instance with vm_state building and task_state spawning. [ 1924.634428] env[62824]: DEBUG oslo_vmware.api [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] Task: {'id': task-2145785, 'name': PowerOnVM_Task} progress is 64%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1924.669861] env[62824]: DEBUG oslo_vmware.api [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Task: {'id': task-2145786, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.106228} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1924.670136] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] [instance: ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1924.670899] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70b41a09-212f-47a2-8350-d297847d7eb8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1924.693348] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] [instance: ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5] Reconfiguring VM instance instance-0000001a to attach disk [datastore1] ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5/ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1924.693686] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d07797d4-86cd-4f8c-9ca5-42cea90babfd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1924.716239] env[62824]: DEBUG oslo_vmware.api [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Waiting for the task: (returnval){ [ 1924.716239] env[62824]: value = "task-2145787" [ 1924.716239] env[62824]: _type = "Task" [ 1924.716239] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1924.724597] env[62824]: DEBUG oslo_vmware.api [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]528e4405-048c-f90f-0612-7d6491f8a8be, 'name': SearchDatastore_Task, 'duration_secs': 0.026317} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1924.725267] env[62824]: DEBUG oslo_concurrency.lockutils [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1924.725518] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] [instance: b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1924.725776] env[62824]: DEBUG oslo_concurrency.lockutils [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1924.725924] env[62824]: DEBUG oslo_concurrency.lockutils [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1924.726121] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1924.729628] env[62824]: DEBUG oslo_concurrency.lockutils [req-626bcab9-7daf-4541-9b77-4e6260023b5b req-955364b2-d588-4095-b0c1-d1199794f497 service nova] Releasing lock "refresh_cache-b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1924.729958] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6e9f9e6f-506f-473e-956c-5489770c0ca2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1924.735992] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7f338944-9cb1-4d55-9d22-6a91d729ee9f tempest-DeleteServersAdminTestJSON-958320793 tempest-DeleteServersAdminTestJSON-958320793-project-member] Lock "517bf2cf-e142-4f36-bf53-79a8d2d8a530" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.815s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1924.737308] env[62824]: DEBUG oslo_vmware.api [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Task: {'id': task-2145787, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1924.752137] env[62824]: DEBUG oslo_vmware.api [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2145784, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.681046} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1924.752137] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 16b466c3-c749-4f96-a82c-32dad31138ec/16b466c3-c749-4f96-a82c-32dad31138ec.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1924.752137] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1924.752403] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1924.752585] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1924.753285] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-433f6916-92df-46e6-ae15-c7a589d1e362 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1924.755663] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c21338a2-de49-47f9-8483-15e5b3b34831 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1924.765238] env[62824]: DEBUG oslo_vmware.api [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Waiting for the task: (returnval){ [ 1924.765238] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52094ce1-85cc-b8c3-fc99-163dfd9a2153" [ 1924.765238] env[62824]: _type = "Task" [ 1924.765238] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1924.766894] env[62824]: DEBUG oslo_vmware.api [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the task: (returnval){ [ 1924.766894] env[62824]: value = "task-2145788" [ 1924.766894] env[62824]: _type = "Task" [ 1924.766894] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1924.778872] env[62824]: DEBUG oslo_vmware.api [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52094ce1-85cc-b8c3-fc99-163dfd9a2153, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1924.784385] env[62824]: DEBUG oslo_vmware.api [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2145788, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1924.848559] env[62824]: DEBUG nova.compute.manager [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 487c2c9d-2cd2-4912-8613-e1bfac732c40] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1924.952225] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Acquiring lock "refresh_cache-b8cc8cd3-ea03-40bf-b867-7ad193365552" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1924.952419] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Acquired lock "refresh_cache-b8cc8cd3-ea03-40bf-b867-7ad193365552" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1924.952602] env[62824]: DEBUG nova.network.neutron [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1925.135482] env[62824]: DEBUG oslo_vmware.api [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] Task: {'id': task-2145785, 'name': PowerOnVM_Task, 'duration_secs': 0.873198} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1925.135797] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1925.136034] env[62824]: DEBUG nova.compute.manager [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1925.136783] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bc58add-d15a-4355-a358-b416ddcf8d38 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1925.231576] env[62824]: DEBUG oslo_vmware.api [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Task: {'id': task-2145787, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1925.281161] env[62824]: DEBUG oslo_vmware.api [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52094ce1-85cc-b8c3-fc99-163dfd9a2153, 'name': SearchDatastore_Task, 'duration_secs': 0.014325} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1925.282038] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-44b23c27-692c-44dd-b96a-70ac7bc8c1c0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1925.287550] env[62824]: DEBUG oslo_vmware.api [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2145788, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.097917} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1925.287970] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1925.288770] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afee8e42-abd2-4228-8aea-d78630323225 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1925.292492] env[62824]: DEBUG oslo_vmware.api [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Waiting for the task: (returnval){ [ 1925.292492] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52ab5894-4fb3-5e4b-aaa3-174888d18958" [ 1925.292492] env[62824]: _type = "Task" [ 1925.292492] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1925.324703] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] Reconfiguring VM instance instance-0000001b to attach disk [datastore1] 16b466c3-c749-4f96-a82c-32dad31138ec/16b466c3-c749-4f96-a82c-32dad31138ec.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1925.326106] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-32ed671b-66a3-44d3-b9b4-cb36f8a91e93 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1925.347816] env[62824]: DEBUG oslo_vmware.api [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52ab5894-4fb3-5e4b-aaa3-174888d18958, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1925.353320] env[62824]: DEBUG oslo_vmware.api [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the task: (returnval){ [ 1925.353320] env[62824]: value = "task-2145789" [ 1925.353320] env[62824]: _type = "Task" [ 1925.353320] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1925.369049] env[62824]: DEBUG oslo_vmware.api [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2145789, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1925.375115] env[62824]: DEBUG oslo_concurrency.lockutils [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1925.512588] env[62824]: DEBUG nova.network.neutron [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1925.654670] env[62824]: DEBUG oslo_concurrency.lockutils [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1925.740041] env[62824]: DEBUG oslo_vmware.api [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Task: {'id': task-2145787, 'name': ReconfigVM_Task, 'duration_secs': 0.647616} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1925.740980] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] [instance: ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5] Reconfigured VM instance instance-0000001a to attach disk [datastore1] ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5/ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1925.741993] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a696fda6-ea1f-4a2d-a046-8b8e433b4b4e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1925.755072] env[62824]: DEBUG oslo_vmware.api [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Waiting for the task: (returnval){ [ 1925.755072] env[62824]: value = "task-2145790" [ 1925.755072] env[62824]: _type = "Task" [ 1925.755072] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1925.765857] env[62824]: DEBUG oslo_vmware.api [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Task: {'id': task-2145790, 'name': Rename_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1925.784394] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Acquiring lock "02405b17-7022-4e2c-a357-415de88f63d0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1925.784758] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Lock "02405b17-7022-4e2c-a357-415de88f63d0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1925.803892] env[62824]: DEBUG oslo_vmware.api [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52ab5894-4fb3-5e4b-aaa3-174888d18958, 'name': SearchDatastore_Task, 'duration_secs': 0.0386} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1925.803892] env[62824]: DEBUG oslo_concurrency.lockutils [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1925.804162] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6/b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1925.804475] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fd0b2d44-5f6f-4c9d-805a-c51e31a7d62a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1925.813810] env[62824]: DEBUG oslo_vmware.api [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Waiting for the task: (returnval){ [ 1925.813810] env[62824]: value = "task-2145791" [ 1925.813810] env[62824]: _type = "Task" [ 1925.813810] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1925.824073] env[62824]: DEBUG oslo_vmware.api [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Task: {'id': task-2145791, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1925.829390] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89e4c00d-9977-45c7-a287-80b975da09d7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1925.838524] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e268b87c-ab3e-4e9f-b5a3-cd03e9abf7b4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1925.875625] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71dfeff8-73b2-479a-9fc3-30e77fdf8f5b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1925.886400] env[62824]: DEBUG oslo_vmware.api [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2145789, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1925.889802] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e80c70b5-d7ff-46e6-befa-3aea679dc5a5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1925.904662] env[62824]: DEBUG nova.compute.provider_tree [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Updating inventory in ProviderTree for provider bf95157b-a274-42de-9ccf-9851128a44a1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1925.930075] env[62824]: DEBUG nova.network.neutron [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Updating instance_info_cache with network_info: [{"id": "60a4fce0-0e63-45af-890c-46ca44ea9a0c", "address": "fa:16:3e:cf:18:eb", "network": {"id": "db233b0c-27f3-4a76-ae9c-2b5889bf3d3b", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1825671448-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3240be4f32254f54aa87cb045e3d8f31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7041d198-66a3-40de-bf7d-cfc036e6ed69", "external-id": "nsx-vlan-transportzone-278", "segmentation_id": 278, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap60a4fce0-0e", "ovs_interfaceid": "60a4fce0-0e63-45af-890c-46ca44ea9a0c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1926.266293] env[62824]: DEBUG oslo_vmware.api [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Task: {'id': task-2145790, 'name': Rename_Task, 'duration_secs': 0.158534} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1926.266562] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] [instance: ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1926.266826] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7db281d7-1b55-42dc-8443-ff35c219aa89 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1926.273918] env[62824]: DEBUG oslo_vmware.api [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Waiting for the task: (returnval){ [ 1926.273918] env[62824]: value = "task-2145792" [ 1926.273918] env[62824]: _type = "Task" [ 1926.273918] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1926.282323] env[62824]: DEBUG oslo_vmware.api [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Task: {'id': task-2145792, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1926.324671] env[62824]: DEBUG oslo_vmware.api [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Task: {'id': task-2145791, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1926.381828] env[62824]: DEBUG oslo_vmware.api [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2145789, 'name': ReconfigVM_Task, 'duration_secs': 0.578106} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1926.382151] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] Reconfigured VM instance instance-0000001b to attach disk [datastore1] 16b466c3-c749-4f96-a82c-32dad31138ec/16b466c3-c749-4f96-a82c-32dad31138ec.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1926.382811] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-15354194-f1d7-481f-88bb-e56b238e971a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1926.389361] env[62824]: DEBUG oslo_vmware.api [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the task: (returnval){ [ 1926.389361] env[62824]: value = "task-2145793" [ 1926.389361] env[62824]: _type = "Task" [ 1926.389361] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1926.399112] env[62824]: DEBUG oslo_vmware.api [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2145793, 'name': Rename_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1926.416717] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b27ea9b2-e866-4d82-b8e1-ed0ef4938234 tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Acquiring lock "384a96ea-40ce-43e1-a5f9-82f50b710b0b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1926.417222] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b27ea9b2-e866-4d82-b8e1-ed0ef4938234 tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Lock "384a96ea-40ce-43e1-a5f9-82f50b710b0b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1926.417493] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b27ea9b2-e866-4d82-b8e1-ed0ef4938234 tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Acquiring lock "384a96ea-40ce-43e1-a5f9-82f50b710b0b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1926.417762] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b27ea9b2-e866-4d82-b8e1-ed0ef4938234 tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Lock "384a96ea-40ce-43e1-a5f9-82f50b710b0b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1926.417993] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b27ea9b2-e866-4d82-b8e1-ed0ef4938234 tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Lock "384a96ea-40ce-43e1-a5f9-82f50b710b0b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1926.420396] env[62824]: INFO nova.compute.manager [None req-b27ea9b2-e866-4d82-b8e1-ed0ef4938234 tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Terminating instance [ 1926.433015] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Releasing lock "refresh_cache-b8cc8cd3-ea03-40bf-b867-7ad193365552" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1926.433397] env[62824]: DEBUG nova.compute.manager [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Instance network_info: |[{"id": "60a4fce0-0e63-45af-890c-46ca44ea9a0c", "address": "fa:16:3e:cf:18:eb", "network": {"id": "db233b0c-27f3-4a76-ae9c-2b5889bf3d3b", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1825671448-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3240be4f32254f54aa87cb045e3d8f31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7041d198-66a3-40de-bf7d-cfc036e6ed69", "external-id": "nsx-vlan-transportzone-278", "segmentation_id": 278, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap60a4fce0-0e", "ovs_interfaceid": "60a4fce0-0e63-45af-890c-46ca44ea9a0c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1926.434423] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:cf:18:eb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7041d198-66a3-40de-bf7d-cfc036e6ed69', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '60a4fce0-0e63-45af-890c-46ca44ea9a0c', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1926.446219] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Creating folder: Project (3240be4f32254f54aa87cb045e3d8f31). Parent ref: group-v438503. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1926.447083] env[62824]: ERROR nova.scheduler.client.report [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] [req-7fb7552e-bbdc-4d8d-8c56-6df15de1793f] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID bf95157b-a274-42de-9ccf-9851128a44a1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-7fb7552e-bbdc-4d8d-8c56-6df15de1793f"}]} [ 1926.447709] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-667edf00-2ec6-4db0-ab5b-48bb5ba0747a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1926.462895] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Created folder: Project (3240be4f32254f54aa87cb045e3d8f31) in parent group-v438503. [ 1926.463648] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Creating folder: Instances. Parent ref: group-v438559. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1926.468404] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-399488ba-4b61-4c41-b612-53b8784c9630 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1926.471495] env[62824]: DEBUG nova.scheduler.client.report [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Refreshing inventories for resource provider bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1926.481597] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Created folder: Instances in parent group-v438559. [ 1926.481900] env[62824]: DEBUG oslo.service.loopingcall [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1926.482193] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1926.482549] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0f41e4b6-0d86-4525-8cab-cc4a5f52a67d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1926.504018] env[62824]: DEBUG nova.scheduler.client.report [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Updating ProviderTree inventory for provider bf95157b-a274-42de-9ccf-9851128a44a1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1926.504018] env[62824]: DEBUG nova.compute.provider_tree [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Updating inventory in ProviderTree for provider bf95157b-a274-42de-9ccf-9851128a44a1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1926.508121] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1926.508121] env[62824]: value = "task-2145796" [ 1926.508121] env[62824]: _type = "Task" [ 1926.508121] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1926.517642] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145796, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1926.518709] env[62824]: DEBUG nova.scheduler.client.report [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Refreshing aggregate associations for resource provider bf95157b-a274-42de-9ccf-9851128a44a1, aggregates: None {{(pid=62824) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1926.544906] env[62824]: DEBUG nova.scheduler.client.report [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Refreshing trait associations for resource provider bf95157b-a274-42de-9ccf-9851128a44a1, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_ISO {{(pid=62824) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1926.691284] env[62824]: DEBUG nova.compute.manager [req-0f232ae8-bf52-4c2a-88ac-33f25c627cc6 req-f5c3214f-e19d-4a91-9d90-e085f31c0d3f service nova] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Received event network-changed-60a4fce0-0e63-45af-890c-46ca44ea9a0c {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1926.691284] env[62824]: DEBUG nova.compute.manager [req-0f232ae8-bf52-4c2a-88ac-33f25c627cc6 req-f5c3214f-e19d-4a91-9d90-e085f31c0d3f service nova] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Refreshing instance network info cache due to event network-changed-60a4fce0-0e63-45af-890c-46ca44ea9a0c. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1926.691284] env[62824]: DEBUG oslo_concurrency.lockutils [req-0f232ae8-bf52-4c2a-88ac-33f25c627cc6 req-f5c3214f-e19d-4a91-9d90-e085f31c0d3f service nova] Acquiring lock "refresh_cache-b8cc8cd3-ea03-40bf-b867-7ad193365552" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1926.691284] env[62824]: DEBUG oslo_concurrency.lockutils [req-0f232ae8-bf52-4c2a-88ac-33f25c627cc6 req-f5c3214f-e19d-4a91-9d90-e085f31c0d3f service nova] Acquired lock "refresh_cache-b8cc8cd3-ea03-40bf-b867-7ad193365552" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1926.691284] env[62824]: DEBUG nova.network.neutron [req-0f232ae8-bf52-4c2a-88ac-33f25c627cc6 req-f5c3214f-e19d-4a91-9d90-e085f31c0d3f service nova] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Refreshing network info cache for port 60a4fce0-0e63-45af-890c-46ca44ea9a0c {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1926.786095] env[62824]: DEBUG oslo_vmware.api [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Task: {'id': task-2145792, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1926.824416] env[62824]: DEBUG oslo_vmware.api [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Task: {'id': task-2145791, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1926.898484] env[62824]: DEBUG oslo_vmware.api [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2145793, 'name': Rename_Task, 'duration_secs': 0.451193} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1926.900879] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1926.901300] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c4381e84-14de-4fe1-9178-2e20628b9cf0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1926.907419] env[62824]: DEBUG oslo_vmware.api [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the task: (returnval){ [ 1926.907419] env[62824]: value = "task-2145797" [ 1926.907419] env[62824]: _type = "Task" [ 1926.907419] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1926.917403] env[62824]: DEBUG oslo_vmware.api [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2145797, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1926.925120] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b27ea9b2-e866-4d82-b8e1-ed0ef4938234 tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Acquiring lock "refresh_cache-384a96ea-40ce-43e1-a5f9-82f50b710b0b" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1926.925285] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b27ea9b2-e866-4d82-b8e1-ed0ef4938234 tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Acquired lock "refresh_cache-384a96ea-40ce-43e1-a5f9-82f50b710b0b" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1926.925546] env[62824]: DEBUG nova.network.neutron [None req-b27ea9b2-e866-4d82-b8e1-ed0ef4938234 tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1927.020651] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145796, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1927.033532] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c0144eb-afae-43d2-93ed-6b6ac32af241 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1927.040795] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e5c24c4-17f5-413d-9cb1-eec82add32ce {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1927.072945] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da0ac56a-09c0-44ea-a1f7-0ddf2f7d6aff {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1927.080976] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8713994-d676-4f5d-8747-6525f1da25ee {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1927.094998] env[62824]: DEBUG nova.compute.provider_tree [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Updating inventory in ProviderTree for provider bf95157b-a274-42de-9ccf-9851128a44a1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1927.285720] env[62824]: DEBUG oslo_vmware.api [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Task: {'id': task-2145792, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1927.324839] env[62824]: DEBUG oslo_vmware.api [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Task: {'id': task-2145791, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1927.409417] env[62824]: DEBUG nova.network.neutron [req-0f232ae8-bf52-4c2a-88ac-33f25c627cc6 req-f5c3214f-e19d-4a91-9d90-e085f31c0d3f service nova] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Updated VIF entry in instance network info cache for port 60a4fce0-0e63-45af-890c-46ca44ea9a0c. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1927.409591] env[62824]: DEBUG nova.network.neutron [req-0f232ae8-bf52-4c2a-88ac-33f25c627cc6 req-f5c3214f-e19d-4a91-9d90-e085f31c0d3f service nova] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Updating instance_info_cache with network_info: [{"id": "60a4fce0-0e63-45af-890c-46ca44ea9a0c", "address": "fa:16:3e:cf:18:eb", "network": {"id": "db233b0c-27f3-4a76-ae9c-2b5889bf3d3b", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1825671448-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3240be4f32254f54aa87cb045e3d8f31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7041d198-66a3-40de-bf7d-cfc036e6ed69", "external-id": "nsx-vlan-transportzone-278", "segmentation_id": 278, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap60a4fce0-0e", "ovs_interfaceid": "60a4fce0-0e63-45af-890c-46ca44ea9a0c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1927.420428] env[62824]: DEBUG oslo_vmware.api [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2145797, 'name': PowerOnVM_Task} progress is 1%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1927.443474] env[62824]: DEBUG nova.network.neutron [None req-b27ea9b2-e866-4d82-b8e1-ed0ef4938234 tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1927.501083] env[62824]: DEBUG nova.network.neutron [None req-b27ea9b2-e866-4d82-b8e1-ed0ef4938234 tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1927.525351] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145796, 'name': CreateVM_Task, 'duration_secs': 0.767449} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1927.525582] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1927.526351] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1927.526540] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1927.526867] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1927.527152] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-43b4c0e0-518c-414c-97c5-5b9c466bcb47 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1927.532351] env[62824]: DEBUG oslo_vmware.api [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Waiting for the task: (returnval){ [ 1927.532351] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52adfd0e-a2e8-b578-fc0b-052eb8f79e25" [ 1927.532351] env[62824]: _type = "Task" [ 1927.532351] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1927.541330] env[62824]: DEBUG oslo_vmware.api [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52adfd0e-a2e8-b578-fc0b-052eb8f79e25, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1927.637241] env[62824]: DEBUG nova.scheduler.client.report [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Updated inventory for provider bf95157b-a274-42de-9ccf-9851128a44a1 with generation 58 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 1927.637241] env[62824]: DEBUG nova.compute.provider_tree [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Updating resource provider bf95157b-a274-42de-9ccf-9851128a44a1 generation from 58 to 59 during operation: update_inventory {{(pid=62824) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1927.637241] env[62824]: DEBUG nova.compute.provider_tree [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Updating inventory in ProviderTree for provider bf95157b-a274-42de-9ccf-9851128a44a1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1927.786305] env[62824]: DEBUG oslo_vmware.api [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Task: {'id': task-2145792, 'name': PowerOnVM_Task, 'duration_secs': 1.128881} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1927.787276] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] [instance: ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1927.787792] env[62824]: INFO nova.compute.manager [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] [instance: ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5] Took 13.22 seconds to spawn the instance on the hypervisor. [ 1927.788321] env[62824]: DEBUG nova.compute.manager [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] [instance: ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1927.790429] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70ac8bbd-a40e-4ecf-943d-a3190293d0a6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1927.825652] env[62824]: DEBUG oslo_vmware.api [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Task: {'id': task-2145791, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.55108} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1927.825991] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6/b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1927.826289] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] [instance: b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1927.826610] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cb6be76f-605d-4743-8aa8-351f83a090bb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1927.833114] env[62824]: DEBUG oslo_vmware.api [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Waiting for the task: (returnval){ [ 1927.833114] env[62824]: value = "task-2145798" [ 1927.833114] env[62824]: _type = "Task" [ 1927.833114] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1927.841342] env[62824]: DEBUG oslo_vmware.api [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Task: {'id': task-2145798, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1927.916811] env[62824]: DEBUG oslo_concurrency.lockutils [req-0f232ae8-bf52-4c2a-88ac-33f25c627cc6 req-f5c3214f-e19d-4a91-9d90-e085f31c0d3f service nova] Releasing lock "refresh_cache-b8cc8cd3-ea03-40bf-b867-7ad193365552" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1927.920668] env[62824]: DEBUG oslo_vmware.api [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2145797, 'name': PowerOnVM_Task} progress is 64%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1928.002994] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b27ea9b2-e866-4d82-b8e1-ed0ef4938234 tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Releasing lock "refresh_cache-384a96ea-40ce-43e1-a5f9-82f50b710b0b" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1928.003483] env[62824]: DEBUG nova.compute.manager [None req-b27ea9b2-e866-4d82-b8e1-ed0ef4938234 tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1928.003681] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-b27ea9b2-e866-4d82-b8e1-ed0ef4938234 tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1928.004598] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5dee0b9-ed4b-4306-81a5-69f5cc6b2a9d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1928.013107] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-b27ea9b2-e866-4d82-b8e1-ed0ef4938234 tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1928.013779] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9d6765fc-6722-478b-926b-ae3f6bc5abf3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1928.020671] env[62824]: DEBUG oslo_vmware.api [None req-b27ea9b2-e866-4d82-b8e1-ed0ef4938234 tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Waiting for the task: (returnval){ [ 1928.020671] env[62824]: value = "task-2145799" [ 1928.020671] env[62824]: _type = "Task" [ 1928.020671] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1928.030397] env[62824]: DEBUG oslo_vmware.api [None req-b27ea9b2-e866-4d82-b8e1-ed0ef4938234 tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Task: {'id': task-2145799, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1928.042425] env[62824]: DEBUG oslo_vmware.api [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52adfd0e-a2e8-b578-fc0b-052eb8f79e25, 'name': SearchDatastore_Task, 'duration_secs': 0.0105} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1928.042838] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1928.043375] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1928.043454] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1928.043614] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1928.043821] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1928.044100] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dfb00a38-5238-4ea6-aa7e-951993f15d65 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1928.053142] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1928.054136] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1928.054249] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cdfcd667-4efd-462a-ae9a-4653801900ff {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1928.059507] env[62824]: DEBUG oslo_vmware.api [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Waiting for the task: (returnval){ [ 1928.059507] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]523420d2-92bf-8f9e-0af8-47cdcec4c185" [ 1928.059507] env[62824]: _type = "Task" [ 1928.059507] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1928.068024] env[62824]: DEBUG oslo_vmware.api [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]523420d2-92bf-8f9e-0af8-47cdcec4c185, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1928.144549] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.957s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1928.146024] env[62824]: DEBUG nova.compute.manager [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] [instance: c36ece43-3d70-4e67-a740-9057f413c722] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1928.148012] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 18.232s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1928.312564] env[62824]: INFO nova.compute.manager [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] [instance: ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5] Took 37.11 seconds to build instance. [ 1928.343400] env[62824]: DEBUG oslo_vmware.api [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Task: {'id': task-2145798, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.10045} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1928.343680] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] [instance: b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1928.345467] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f148b07-0b5f-4e47-ade5-5bcc998fd13b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1928.373780] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] [instance: b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6] Reconfiguring VM instance instance-0000001c to attach disk [datastore2] b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6/b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1928.373780] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0e5bc96a-bc58-45a6-997e-c8c17dad8ea5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1928.393345] env[62824]: DEBUG oslo_vmware.api [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Waiting for the task: (returnval){ [ 1928.393345] env[62824]: value = "task-2145800" [ 1928.393345] env[62824]: _type = "Task" [ 1928.393345] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1928.401781] env[62824]: DEBUG oslo_vmware.api [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Task: {'id': task-2145800, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1928.422491] env[62824]: DEBUG oslo_vmware.api [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2145797, 'name': PowerOnVM_Task} progress is 91%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1928.530296] env[62824]: DEBUG oslo_vmware.api [None req-b27ea9b2-e866-4d82-b8e1-ed0ef4938234 tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Task: {'id': task-2145799, 'name': PowerOffVM_Task, 'duration_secs': 0.2208} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1928.530587] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-b27ea9b2-e866-4d82-b8e1-ed0ef4938234 tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1928.530757] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-b27ea9b2-e866-4d82-b8e1-ed0ef4938234 tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1928.531106] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-621898ae-fea8-46fe-9aeb-d63c25d88a2d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1928.557855] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-b27ea9b2-e866-4d82-b8e1-ed0ef4938234 tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1928.558103] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-b27ea9b2-e866-4d82-b8e1-ed0ef4938234 tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Deleting contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1928.558290] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-b27ea9b2-e866-4d82-b8e1-ed0ef4938234 tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Deleting the datastore file [datastore2] 384a96ea-40ce-43e1-a5f9-82f50b710b0b {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1928.558549] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f8f3bf89-b1ed-4aa4-bd53-614bfebf10f7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1928.569383] env[62824]: DEBUG oslo_vmware.api [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]523420d2-92bf-8f9e-0af8-47cdcec4c185, 'name': SearchDatastore_Task, 'duration_secs': 0.010176} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1928.571307] env[62824]: DEBUG oslo_vmware.api [None req-b27ea9b2-e866-4d82-b8e1-ed0ef4938234 tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Waiting for the task: (returnval){ [ 1928.571307] env[62824]: value = "task-2145802" [ 1928.571307] env[62824]: _type = "Task" [ 1928.571307] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1928.571975] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ffd61b43-3ae0-4aed-8256-2c518b62b804 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1928.579409] env[62824]: DEBUG oslo_vmware.api [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Waiting for the task: (returnval){ [ 1928.579409] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52f24a5c-1a88-eebd-ccb3-3415b15fa766" [ 1928.579409] env[62824]: _type = "Task" [ 1928.579409] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1928.582798] env[62824]: DEBUG oslo_vmware.api [None req-b27ea9b2-e866-4d82-b8e1-ed0ef4938234 tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Task: {'id': task-2145802, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1928.590156] env[62824]: DEBUG oslo_vmware.api [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52f24a5c-1a88-eebd-ccb3-3415b15fa766, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1928.653177] env[62824]: DEBUG nova.compute.utils [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1928.662711] env[62824]: DEBUG nova.compute.manager [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] [instance: c36ece43-3d70-4e67-a740-9057f413c722] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1928.662890] env[62824]: DEBUG nova.network.neutron [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] [instance: c36ece43-3d70-4e67-a740-9057f413c722] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1928.721542] env[62824]: DEBUG nova.policy [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0059ec60e4884d22addac429b94794e9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '78831267dd354243b0f284437b835ef5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 1928.814944] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c9420c52-2e7b-4179-a724-dc8220206e37 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Lock "ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 65.105s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1928.903453] env[62824]: DEBUG oslo_vmware.api [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Task: {'id': task-2145800, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1928.923830] env[62824]: DEBUG oslo_vmware.api [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2145797, 'name': PowerOnVM_Task, 'duration_secs': 1.542152} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1928.925016] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1928.925016] env[62824]: INFO nova.compute.manager [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] Took 11.59 seconds to spawn the instance on the hypervisor. [ 1928.925016] env[62824]: DEBUG nova.compute.manager [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1928.932105] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e0ee250-4f6f-47a2-a187-1bd79280c132 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1929.085527] env[62824]: DEBUG oslo_vmware.api [None req-b27ea9b2-e866-4d82-b8e1-ed0ef4938234 tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Task: {'id': task-2145802, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.223915} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1929.087547] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-b27ea9b2-e866-4d82-b8e1-ed0ef4938234 tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1929.087805] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-b27ea9b2-e866-4d82-b8e1-ed0ef4938234 tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Deleted contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1929.087988] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-b27ea9b2-e866-4d82-b8e1-ed0ef4938234 tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1929.089553] env[62824]: INFO nova.compute.manager [None req-b27ea9b2-e866-4d82-b8e1-ed0ef4938234 tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Took 1.08 seconds to destroy the instance on the hypervisor. [ 1929.089817] env[62824]: DEBUG oslo.service.loopingcall [None req-b27ea9b2-e866-4d82-b8e1-ed0ef4938234 tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1929.090028] env[62824]: DEBUG nova.compute.manager [-] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1929.090125] env[62824]: DEBUG nova.network.neutron [-] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1929.097853] env[62824]: DEBUG oslo_vmware.api [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52f24a5c-1a88-eebd-ccb3-3415b15fa766, 'name': SearchDatastore_Task, 'duration_secs': 0.010961} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1929.098147] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1929.098401] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] b8cc8cd3-ea03-40bf-b867-7ad193365552/b8cc8cd3-ea03-40bf-b867-7ad193365552.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1929.098660] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2e12a186-a1ad-44a5-8d57-d29971ea5c32 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1929.107039] env[62824]: DEBUG oslo_vmware.api [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Waiting for the task: (returnval){ [ 1929.107039] env[62824]: value = "task-2145803" [ 1929.107039] env[62824]: _type = "Task" [ 1929.107039] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1929.113322] env[62824]: DEBUG oslo_vmware.api [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2145803, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1929.118501] env[62824]: DEBUG nova.network.neutron [-] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1929.133106] env[62824]: DEBUG nova.network.neutron [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] [instance: c36ece43-3d70-4e67-a740-9057f413c722] Successfully created port: eb387e38-7606-4160-a960-43d51598b964 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1929.167418] env[62824]: INFO nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Updating resource usage from migration bc55ced6-f1ca-4506-869b-2d7d5683b841 [ 1929.174142] env[62824]: DEBUG nova.compute.manager [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] [instance: c36ece43-3d70-4e67-a740-9057f413c722] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1929.198280] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 29b8dd5f-1855-490c-a01b-54840073a753 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1929.198280] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 82382932-7302-4441-a6f8-9aa2300ec0f6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1929.198280] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance f4d63a93-23af-470c-b36a-662af81dc386 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1929.198280] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1929.198428] env[62824]: WARNING nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 9062e606-511f-4fe7-9621-90b3c0e51eec is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1929.198428] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 384a96ea-40ce-43e1-a5f9-82f50b710b0b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1929.198428] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1929.198428] env[62824]: WARNING nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 4dfe9f82-7f89-486b-b159-948656d2896c is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1929.198549] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 07ce931f-45ef-409b-b714-9f1cd47a3a88 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1929.198549] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance f3740d9a-b21a-4a79-9e28-2a89ecd08bb6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1929.198549] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1929.198549] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 16b466c3-c749-4f96-a82c-32dad31138ec actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1929.198668] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1929.198668] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance b8cc8cd3-ea03-40bf-b867-7ad193365552 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1929.198668] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance c36ece43-3d70-4e67-a740-9057f413c722 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1929.321539] env[62824]: DEBUG nova.compute.manager [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] [instance: 69bcac33-d6ac-4ee7-b674-58ab25389149] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1929.404817] env[62824]: DEBUG oslo_vmware.api [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Task: {'id': task-2145800, 'name': ReconfigVM_Task, 'duration_secs': 0.524749} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1929.405128] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] [instance: b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6] Reconfigured VM instance instance-0000001c to attach disk [datastore2] b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6/b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1929.405822] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6c2c36ea-2529-424c-8e5b-15c38b308c16 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1929.413408] env[62824]: DEBUG oslo_vmware.api [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Waiting for the task: (returnval){ [ 1929.413408] env[62824]: value = "task-2145804" [ 1929.413408] env[62824]: _type = "Task" [ 1929.413408] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1929.421818] env[62824]: DEBUG oslo_vmware.api [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Task: {'id': task-2145804, 'name': Rename_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1929.448299] env[62824]: INFO nova.compute.manager [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] Took 32.98 seconds to build instance. [ 1929.616051] env[62824]: DEBUG oslo_vmware.api [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2145803, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1929.624640] env[62824]: DEBUG nova.network.neutron [-] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1929.626199] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d7484a56-617e-4824-9c3a-1e532dad4eb7 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Acquiring lock "interface-ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1929.626327] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d7484a56-617e-4824-9c3a-1e532dad4eb7 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Lock "interface-ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1929.626631] env[62824]: DEBUG nova.objects.instance [None req-d7484a56-617e-4824-9c3a-1e532dad4eb7 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Lazy-loading 'flavor' on Instance uuid ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1929.702908] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance b96d1351-f3a7-4bac-998c-a34ab2606041 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1929.843142] env[62824]: DEBUG oslo_concurrency.lockutils [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1929.923422] env[62824]: DEBUG oslo_vmware.api [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Task: {'id': task-2145804, 'name': Rename_Task, 'duration_secs': 0.188211} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1929.923695] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] [instance: b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1929.923940] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ab009b14-fc92-4ed2-93c8-da7763f9ecaa {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1929.929971] env[62824]: DEBUG oslo_vmware.api [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Waiting for the task: (returnval){ [ 1929.929971] env[62824]: value = "task-2145805" [ 1929.929971] env[62824]: _type = "Task" [ 1929.929971] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1929.938127] env[62824]: DEBUG oslo_vmware.api [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Task: {'id': task-2145805, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1929.950078] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ca8423a1-497e-4c59-a756-4cd1d8035445 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Lock "16b466c3-c749-4f96-a82c-32dad31138ec" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 65.354s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1930.118813] env[62824]: DEBUG oslo_vmware.api [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2145803, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.762284} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1930.119212] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] b8cc8cd3-ea03-40bf-b867-7ad193365552/b8cc8cd3-ea03-40bf-b867-7ad193365552.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1930.119518] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1930.119884] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ac11421e-4348-4a34-a9ee-858cc0054aba {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1930.127430] env[62824]: DEBUG oslo_vmware.api [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Waiting for the task: (returnval){ [ 1930.127430] env[62824]: value = "task-2145806" [ 1930.127430] env[62824]: _type = "Task" [ 1930.127430] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1930.133361] env[62824]: DEBUG nova.objects.instance [None req-d7484a56-617e-4824-9c3a-1e532dad4eb7 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Lazy-loading 'pci_requests' on Instance uuid ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1930.135117] env[62824]: INFO nova.compute.manager [-] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Took 1.04 seconds to deallocate network for instance. [ 1930.144877] env[62824]: DEBUG oslo_vmware.api [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2145806, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1930.180712] env[62824]: DEBUG nova.compute.manager [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] [instance: c36ece43-3d70-4e67-a740-9057f413c722] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1930.206227] env[62824]: DEBUG nova.virt.hardware [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1930.206492] env[62824]: DEBUG nova.virt.hardware [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1930.206650] env[62824]: DEBUG nova.virt.hardware [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1930.206844] env[62824]: DEBUG nova.virt.hardware [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1930.207065] env[62824]: DEBUG nova.virt.hardware [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1930.207173] env[62824]: DEBUG nova.virt.hardware [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1930.207480] env[62824]: DEBUG nova.virt.hardware [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1930.207581] env[62824]: DEBUG nova.virt.hardware [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1930.207737] env[62824]: DEBUG nova.virt.hardware [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1930.207848] env[62824]: DEBUG nova.virt.hardware [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1930.208051] env[62824]: DEBUG nova.virt.hardware [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1930.208754] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance b3003c4b-ae5a-48df-8c12-a915a76253f4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1930.211415] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-388e6e63-71a3-4537-a376-8d8c09072de7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1930.219364] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d8b5c1d-3dcc-49b3-adb1-e9da70a08050 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1930.439987] env[62824]: DEBUG oslo_vmware.api [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Task: {'id': task-2145805, 'name': PowerOnVM_Task} progress is 19%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1930.452202] env[62824]: DEBUG nova.compute.manager [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] [instance: 5f1237ff-5841-41f1-8d77-df9fabaa62b6] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1930.637745] env[62824]: DEBUG nova.objects.base [None req-d7484a56-617e-4824-9c3a-1e532dad4eb7 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=62824) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1930.637993] env[62824]: DEBUG nova.network.neutron [None req-d7484a56-617e-4824-9c3a-1e532dad4eb7 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] [instance: ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1930.639786] env[62824]: DEBUG oslo_vmware.api [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2145806, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075849} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1930.640044] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1930.641176] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fac4bfbf-6a2d-4e11-93be-4dbe28e98e67 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1930.656195] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b27ea9b2-e866-4d82-b8e1-ed0ef4938234 tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1930.666516] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Reconfiguring VM instance instance-0000001d to attach disk [datastore1] b8cc8cd3-ea03-40bf-b867-7ad193365552/b8cc8cd3-ea03-40bf-b867-7ad193365552.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1930.666839] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e0332776-46d3-4e31-bec5-7016335f0379 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1930.690062] env[62824]: DEBUG oslo_vmware.api [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Waiting for the task: (returnval){ [ 1930.690062] env[62824]: value = "task-2145807" [ 1930.690062] env[62824]: _type = "Task" [ 1930.690062] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1930.699223] env[62824]: DEBUG oslo_vmware.api [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2145807, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1930.715826] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 6d80ec06-8559-4964-8577-a2512aa366ed has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1930.739277] env[62824]: DEBUG nova.network.neutron [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] [instance: c36ece43-3d70-4e67-a740-9057f413c722] Successfully updated port: eb387e38-7606-4160-a960-43d51598b964 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1930.774288] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d7484a56-617e-4824-9c3a-1e532dad4eb7 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Lock "interface-ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.148s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1930.817533] env[62824]: DEBUG oslo_concurrency.lockutils [None req-6c7bf437-76f6-4a22-a22b-d3d2820ad6e7 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Acquiring lock "16b466c3-c749-4f96-a82c-32dad31138ec" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1930.817866] env[62824]: DEBUG oslo_concurrency.lockutils [None req-6c7bf437-76f6-4a22-a22b-d3d2820ad6e7 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Lock "16b466c3-c749-4f96-a82c-32dad31138ec" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1930.818223] env[62824]: DEBUG nova.compute.manager [None req-6c7bf437-76f6-4a22-a22b-d3d2820ad6e7 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1930.819030] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91405206-36e2-4239-8720-d1ffb1e1438f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1930.826667] env[62824]: DEBUG nova.compute.manager [None req-6c7bf437-76f6-4a22-a22b-d3d2820ad6e7 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62824) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3404}} [ 1930.827506] env[62824]: DEBUG nova.objects.instance [None req-6c7bf437-76f6-4a22-a22b-d3d2820ad6e7 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Lazy-loading 'flavor' on Instance uuid 16b466c3-c749-4f96-a82c-32dad31138ec {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1930.945459] env[62824]: DEBUG oslo_vmware.api [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Task: {'id': task-2145805, 'name': PowerOnVM_Task} progress is 91%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1930.983527] env[62824]: DEBUG oslo_concurrency.lockutils [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1931.204596] env[62824]: DEBUG oslo_vmware.api [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2145807, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1931.219616] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 59b5b883-4188-471c-8862-444f3ce08cb0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1931.241645] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Acquiring lock "refresh_cache-c36ece43-3d70-4e67-a740-9057f413c722" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1931.241791] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Acquired lock "refresh_cache-c36ece43-3d70-4e67-a740-9057f413c722" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1931.241939] env[62824]: DEBUG nova.network.neutron [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] [instance: c36ece43-3d70-4e67-a740-9057f413c722] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1931.310541] env[62824]: DEBUG nova.compute.manager [req-40a39184-1cf5-4d17-ae06-0cab872b21c3 req-f26334b4-6488-4231-823d-4a8da8541b1e service nova] [instance: c36ece43-3d70-4e67-a740-9057f413c722] Received event network-vif-plugged-eb387e38-7606-4160-a960-43d51598b964 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1931.310765] env[62824]: DEBUG oslo_concurrency.lockutils [req-40a39184-1cf5-4d17-ae06-0cab872b21c3 req-f26334b4-6488-4231-823d-4a8da8541b1e service nova] Acquiring lock "c36ece43-3d70-4e67-a740-9057f413c722-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1931.310972] env[62824]: DEBUG oslo_concurrency.lockutils [req-40a39184-1cf5-4d17-ae06-0cab872b21c3 req-f26334b4-6488-4231-823d-4a8da8541b1e service nova] Lock "c36ece43-3d70-4e67-a740-9057f413c722-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1931.311152] env[62824]: DEBUG oslo_concurrency.lockutils [req-40a39184-1cf5-4d17-ae06-0cab872b21c3 req-f26334b4-6488-4231-823d-4a8da8541b1e service nova] Lock "c36ece43-3d70-4e67-a740-9057f413c722-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1931.311317] env[62824]: DEBUG nova.compute.manager [req-40a39184-1cf5-4d17-ae06-0cab872b21c3 req-f26334b4-6488-4231-823d-4a8da8541b1e service nova] [instance: c36ece43-3d70-4e67-a740-9057f413c722] No waiting events found dispatching network-vif-plugged-eb387e38-7606-4160-a960-43d51598b964 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1931.311482] env[62824]: WARNING nova.compute.manager [req-40a39184-1cf5-4d17-ae06-0cab872b21c3 req-f26334b4-6488-4231-823d-4a8da8541b1e service nova] [instance: c36ece43-3d70-4e67-a740-9057f413c722] Received unexpected event network-vif-plugged-eb387e38-7606-4160-a960-43d51598b964 for instance with vm_state building and task_state spawning. [ 1931.445737] env[62824]: DEBUG oslo_vmware.api [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Task: {'id': task-2145805, 'name': PowerOnVM_Task, 'duration_secs': 1.043115} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1931.446066] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] [instance: b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1931.446353] env[62824]: INFO nova.compute.manager [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] [instance: b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6] Took 11.41 seconds to spawn the instance on the hypervisor. [ 1931.446604] env[62824]: DEBUG nova.compute.manager [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] [instance: b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1931.447708] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a308786c-4ca5-411e-b000-3aef7d4e41f0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1931.700651] env[62824]: DEBUG oslo_vmware.api [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2145807, 'name': ReconfigVM_Task, 'duration_secs': 0.811152} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1931.700954] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Reconfigured VM instance instance-0000001d to attach disk [datastore1] b8cc8cd3-ea03-40bf-b867-7ad193365552/b8cc8cd3-ea03-40bf-b867-7ad193365552.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1931.701570] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4457ad88-9eb4-43b9-a52a-fb492e84533c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1931.707555] env[62824]: DEBUG oslo_vmware.api [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Waiting for the task: (returnval){ [ 1931.707555] env[62824]: value = "task-2145808" [ 1931.707555] env[62824]: _type = "Task" [ 1931.707555] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1931.715122] env[62824]: DEBUG oslo_vmware.api [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2145808, 'name': Rename_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1931.722959] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 487c2c9d-2cd2-4912-8613-e1bfac732c40 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1931.819299] env[62824]: DEBUG nova.network.neutron [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] [instance: c36ece43-3d70-4e67-a740-9057f413c722] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1931.834712] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c7bf437-76f6-4a22-a22b-d3d2820ad6e7 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1931.835045] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-10e5d854-7829-4f72-a6c5-a64a2ac0f2c4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1931.842294] env[62824]: DEBUG oslo_vmware.api [None req-6c7bf437-76f6-4a22-a22b-d3d2820ad6e7 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the task: (returnval){ [ 1931.842294] env[62824]: value = "task-2145809" [ 1931.842294] env[62824]: _type = "Task" [ 1931.842294] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1931.855194] env[62824]: DEBUG oslo_vmware.api [None req-6c7bf437-76f6-4a22-a22b-d3d2820ad6e7 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2145809, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1931.970606] env[62824]: INFO nova.compute.manager [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] [instance: b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6] Took 34.42 seconds to build instance. [ 1932.116109] env[62824]: DEBUG nova.network.neutron [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] [instance: c36ece43-3d70-4e67-a740-9057f413c722] Updating instance_info_cache with network_info: [{"id": "eb387e38-7606-4160-a960-43d51598b964", "address": "fa:16:3e:47:61:95", "network": {"id": "b6403dc3-30ef-4c76-b521-380c179f345d", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.139", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "22c5c40c2e004babaa15896c89307303", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1002b79b-224e-41e3-a484-4245a767147a", "external-id": "nsx-vlan-transportzone-353", "segmentation_id": 353, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeb387e38-76", "ovs_interfaceid": "eb387e38-7606-4160-a960-43d51598b964", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1932.218322] env[62824]: DEBUG oslo_vmware.api [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2145808, 'name': Rename_Task, 'duration_secs': 0.24176} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1932.218479] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1932.218706] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7ed4bea4-3600-49c3-8b86-c82c7ed8d07b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1932.225312] env[62824]: DEBUG oslo_vmware.api [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Waiting for the task: (returnval){ [ 1932.225312] env[62824]: value = "task-2145810" [ 1932.225312] env[62824]: _type = "Task" [ 1932.225312] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1932.228781] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 69bcac33-d6ac-4ee7-b674-58ab25389149 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1932.234968] env[62824]: DEBUG oslo_vmware.api [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2145810, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1932.354277] env[62824]: DEBUG oslo_vmware.api [None req-6c7bf437-76f6-4a22-a22b-d3d2820ad6e7 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2145809, 'name': PowerOffVM_Task, 'duration_secs': 0.2404} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1932.354570] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c7bf437-76f6-4a22-a22b-d3d2820ad6e7 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1932.354765] env[62824]: DEBUG nova.compute.manager [None req-6c7bf437-76f6-4a22-a22b-d3d2820ad6e7 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1932.358019] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34d827e9-b2ed-4810-987c-30fe3fb1652e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1932.474182] env[62824]: DEBUG oslo_concurrency.lockutils [None req-370e56f6-0096-46b3-ad39-684a37405f9a tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Lock "b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 67.428s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1932.621015] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Releasing lock "refresh_cache-c36ece43-3d70-4e67-a740-9057f413c722" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1932.621015] env[62824]: DEBUG nova.compute.manager [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] [instance: c36ece43-3d70-4e67-a740-9057f413c722] Instance network_info: |[{"id": "eb387e38-7606-4160-a960-43d51598b964", "address": "fa:16:3e:47:61:95", "network": {"id": "b6403dc3-30ef-4c76-b521-380c179f345d", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.139", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "22c5c40c2e004babaa15896c89307303", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1002b79b-224e-41e3-a484-4245a767147a", "external-id": "nsx-vlan-transportzone-353", "segmentation_id": 353, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeb387e38-76", "ovs_interfaceid": "eb387e38-7606-4160-a960-43d51598b964", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1932.621373] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] [instance: c36ece43-3d70-4e67-a740-9057f413c722] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:47:61:95', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1002b79b-224e-41e3-a484-4245a767147a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'eb387e38-7606-4160-a960-43d51598b964', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1932.628265] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Creating folder: Project (78831267dd354243b0f284437b835ef5). Parent ref: group-v438503. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1932.629345] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-dd5da0fe-ce26-4ff3-9da1-5b3e26442496 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1932.641652] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Created folder: Project (78831267dd354243b0f284437b835ef5) in parent group-v438503. [ 1932.641652] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Creating folder: Instances. Parent ref: group-v438562. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1932.641652] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e7cc4cb6-5c63-4b50-8c89-58e55d640bcf {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1932.651590] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Created folder: Instances in parent group-v438562. [ 1932.651865] env[62824]: DEBUG oslo.service.loopingcall [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1932.653364] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c36ece43-3d70-4e67-a740-9057f413c722] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1932.653364] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bdff7361-d2f1-4592-aef3-5d0317b2e29c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1932.670996] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1932.670996] env[62824]: value = "task-2145813" [ 1932.670996] env[62824]: _type = "Task" [ 1932.670996] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1932.678570] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145813, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1932.731583] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 5f1237ff-5841-41f1-8d77-df9fabaa62b6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1932.737768] env[62824]: DEBUG oslo_vmware.api [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2145810, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1932.869261] env[62824]: DEBUG oslo_concurrency.lockutils [None req-6c7bf437-76f6-4a22-a22b-d3d2820ad6e7 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Lock "16b466c3-c749-4f96-a82c-32dad31138ec" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.051s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1932.976123] env[62824]: DEBUG nova.compute.manager [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: dc3317e4-6958-4df6-8e39-c89b4a5d861c] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1933.079563] env[62824]: DEBUG nova.compute.manager [None req-be0fa035-e0fc-46c3-b3d9-c94dc60ee21d tempest-ServerExternalEventsTest-1049562868 tempest-ServerExternalEventsTest-1049562868-project] [instance: b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6] Received event network-changed {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1933.079802] env[62824]: DEBUG nova.compute.manager [None req-be0fa035-e0fc-46c3-b3d9-c94dc60ee21d tempest-ServerExternalEventsTest-1049562868 tempest-ServerExternalEventsTest-1049562868-project] [instance: b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6] Refreshing instance network info cache due to event network-changed. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1933.080197] env[62824]: DEBUG oslo_concurrency.lockutils [None req-be0fa035-e0fc-46c3-b3d9-c94dc60ee21d tempest-ServerExternalEventsTest-1049562868 tempest-ServerExternalEventsTest-1049562868-project] Acquiring lock "refresh_cache-b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1933.080432] env[62824]: DEBUG oslo_concurrency.lockutils [None req-be0fa035-e0fc-46c3-b3d9-c94dc60ee21d tempest-ServerExternalEventsTest-1049562868 tempest-ServerExternalEventsTest-1049562868-project] Acquired lock "refresh_cache-b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1933.080626] env[62824]: DEBUG nova.network.neutron [None req-be0fa035-e0fc-46c3-b3d9-c94dc60ee21d tempest-ServerExternalEventsTest-1049562868 tempest-ServerExternalEventsTest-1049562868-project] [instance: b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1933.089300] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8453c213-135b-48f6-a29c-da7b36504504 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Acquiring lock "ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1933.089874] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8453c213-135b-48f6-a29c-da7b36504504 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Lock "ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1933.090155] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8453c213-135b-48f6-a29c-da7b36504504 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Acquiring lock "ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1933.090379] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8453c213-135b-48f6-a29c-da7b36504504 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Lock "ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1933.090803] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8453c213-135b-48f6-a29c-da7b36504504 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Lock "ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1933.093014] env[62824]: INFO nova.compute.manager [None req-8453c213-135b-48f6-a29c-da7b36504504 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] [instance: ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5] Terminating instance [ 1933.184348] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145813, 'name': CreateVM_Task, 'duration_secs': 0.491444} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1933.184791] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c36ece43-3d70-4e67-a740-9057f413c722] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1933.187073] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1933.187073] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1933.187073] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1933.187073] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2878c33d-1d56-44aa-9fe4-64b425139d0c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1933.193025] env[62824]: DEBUG oslo_vmware.api [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Waiting for the task: (returnval){ [ 1933.193025] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5286c835-3380-22d5-1646-b74e75d847aa" [ 1933.193025] env[62824]: _type = "Task" [ 1933.193025] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1933.203728] env[62824]: DEBUG oslo_vmware.api [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5286c835-3380-22d5-1646-b74e75d847aa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1933.234855] env[62824]: DEBUG oslo_vmware.api [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2145810, 'name': PowerOnVM_Task, 'duration_secs': 0.784327} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1933.235341] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1933.235576] env[62824]: INFO nova.compute.manager [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Took 9.29 seconds to spawn the instance on the hypervisor. [ 1933.235758] env[62824]: DEBUG nova.compute.manager [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1933.236533] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e762f86-6e14-4883-ae8c-c29bdeae9c71 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1933.240114] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance dc3317e4-6958-4df6-8e39-c89b4a5d861c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1933.513411] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1933.599399] env[62824]: DEBUG nova.compute.manager [None req-8453c213-135b-48f6-a29c-da7b36504504 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] [instance: ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1933.599689] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-8453c213-135b-48f6-a29c-da7b36504504 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] [instance: ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1933.600686] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7174d83f-5e65-45e9-9b75-d0a8b8967a16 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1933.609587] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-8453c213-135b-48f6-a29c-da7b36504504 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] [instance: ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1933.609820] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d0a96d8c-08f6-4299-b5b5-8f9cd273dfec {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1933.617496] env[62824]: DEBUG oslo_vmware.api [None req-8453c213-135b-48f6-a29c-da7b36504504 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Waiting for the task: (returnval){ [ 1933.617496] env[62824]: value = "task-2145814" [ 1933.617496] env[62824]: _type = "Task" [ 1933.617496] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1933.626919] env[62824]: DEBUG oslo_vmware.api [None req-8453c213-135b-48f6-a29c-da7b36504504 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Task: {'id': task-2145814, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1933.703410] env[62824]: DEBUG oslo_vmware.api [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5286c835-3380-22d5-1646-b74e75d847aa, 'name': SearchDatastore_Task, 'duration_secs': 0.012975} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1933.704920] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1933.705092] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] [instance: c36ece43-3d70-4e67-a740-9057f413c722] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1933.705334] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1933.705476] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1933.705867] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1933.711878] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d78fe48e-c87a-4ffb-83c4-d450d8c52847 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1933.718187] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1933.718354] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1933.719821] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6fabdc4a-848b-4e72-9524-e337a15f167b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1933.727109] env[62824]: DEBUG oslo_vmware.api [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Waiting for the task: (returnval){ [ 1933.727109] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52f4f3e3-4b32-1ddf-8a91-f7d6c0a49336" [ 1933.727109] env[62824]: _type = "Task" [ 1933.727109] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1933.740814] env[62824]: DEBUG oslo_vmware.api [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52f4f3e3-4b32-1ddf-8a91-f7d6c0a49336, 'name': SearchDatastore_Task, 'duration_secs': 0.00929} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1933.741711] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e1aa051e-ac02-4dde-bd70-ed58e9de77b3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1933.745275] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance bbf1440b-8681-48c8-a178-9a83b925c695 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1933.762014] env[62824]: DEBUG oslo_vmware.api [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Waiting for the task: (returnval){ [ 1933.762014] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]525e311a-9b93-b9c6-462b-1b2e9bdda251" [ 1933.762014] env[62824]: _type = "Task" [ 1933.762014] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1933.762014] env[62824]: INFO nova.compute.manager [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Took 31.28 seconds to build instance. [ 1933.773401] env[62824]: DEBUG oslo_vmware.api [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]525e311a-9b93-b9c6-462b-1b2e9bdda251, 'name': SearchDatastore_Task, 'duration_secs': 0.009788} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1933.773749] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1933.774044] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] c36ece43-3d70-4e67-a740-9057f413c722/c36ece43-3d70-4e67-a740-9057f413c722.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1933.778211] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7fd06cd6-cc76-45da-a99c-499623a403b5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1933.787149] env[62824]: DEBUG oslo_vmware.api [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Waiting for the task: (returnval){ [ 1933.787149] env[62824]: value = "task-2145815" [ 1933.787149] env[62824]: _type = "Task" [ 1933.787149] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1933.802700] env[62824]: DEBUG oslo_vmware.api [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Task: {'id': task-2145815, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1933.912666] env[62824]: DEBUG nova.compute.manager [req-3e0674b8-989b-4ef9-a8f5-c50fbefb3d66 req-81c3da41-d6a6-45e2-8fa0-8ddf26d64ba3 service nova] [instance: c36ece43-3d70-4e67-a740-9057f413c722] Received event network-changed-eb387e38-7606-4160-a960-43d51598b964 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1933.912794] env[62824]: DEBUG nova.compute.manager [req-3e0674b8-989b-4ef9-a8f5-c50fbefb3d66 req-81c3da41-d6a6-45e2-8fa0-8ddf26d64ba3 service nova] [instance: c36ece43-3d70-4e67-a740-9057f413c722] Refreshing instance network info cache due to event network-changed-eb387e38-7606-4160-a960-43d51598b964. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1933.912989] env[62824]: DEBUG oslo_concurrency.lockutils [req-3e0674b8-989b-4ef9-a8f5-c50fbefb3d66 req-81c3da41-d6a6-45e2-8fa0-8ddf26d64ba3 service nova] Acquiring lock "refresh_cache-c36ece43-3d70-4e67-a740-9057f413c722" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1933.913153] env[62824]: DEBUG oslo_concurrency.lockutils [req-3e0674b8-989b-4ef9-a8f5-c50fbefb3d66 req-81c3da41-d6a6-45e2-8fa0-8ddf26d64ba3 service nova] Acquired lock "refresh_cache-c36ece43-3d70-4e67-a740-9057f413c722" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1933.913343] env[62824]: DEBUG nova.network.neutron [req-3e0674b8-989b-4ef9-a8f5-c50fbefb3d66 req-81c3da41-d6a6-45e2-8fa0-8ddf26d64ba3 service nova] [instance: c36ece43-3d70-4e67-a740-9057f413c722] Refreshing network info cache for port eb387e38-7606-4160-a960-43d51598b964 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1934.073414] env[62824]: DEBUG nova.network.neutron [None req-be0fa035-e0fc-46c3-b3d9-c94dc60ee21d tempest-ServerExternalEventsTest-1049562868 tempest-ServerExternalEventsTest-1049562868-project] [instance: b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6] Updating instance_info_cache with network_info: [{"id": "c21b7203-99b2-485d-8a8f-db1f6bcdbbe0", "address": "fa:16:3e:62:46:77", "network": {"id": "b6403dc3-30ef-4c76-b521-380c179f345d", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.200", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "22c5c40c2e004babaa15896c89307303", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1002b79b-224e-41e3-a484-4245a767147a", "external-id": "nsx-vlan-transportzone-353", "segmentation_id": 353, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc21b7203-99", "ovs_interfaceid": "c21b7203-99b2-485d-8a8f-db1f6bcdbbe0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1934.132615] env[62824]: DEBUG oslo_vmware.api [None req-8453c213-135b-48f6-a29c-da7b36504504 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Task: {'id': task-2145814, 'name': PowerOffVM_Task, 'duration_secs': 0.268583} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1934.132837] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-8453c213-135b-48f6-a29c-da7b36504504 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] [instance: ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1934.133019] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-8453c213-135b-48f6-a29c-da7b36504504 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] [instance: ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1934.133310] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bd269fdf-10fa-43f0-9d20-caed0c5a185d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1934.212029] env[62824]: DEBUG oslo_concurrency.lockutils [None req-995ba297-d41c-4960-b797-2a0283f98d18 tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Acquiring lock "b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1934.212643] env[62824]: DEBUG oslo_concurrency.lockutils [None req-995ba297-d41c-4960-b797-2a0283f98d18 tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Lock "b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1934.212959] env[62824]: DEBUG oslo_concurrency.lockutils [None req-995ba297-d41c-4960-b797-2a0283f98d18 tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Acquiring lock "b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1934.213219] env[62824]: DEBUG oslo_concurrency.lockutils [None req-995ba297-d41c-4960-b797-2a0283f98d18 tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Lock "b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1934.213437] env[62824]: DEBUG oslo_concurrency.lockutils [None req-995ba297-d41c-4960-b797-2a0283f98d18 tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Lock "b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1934.217310] env[62824]: INFO nova.compute.manager [None req-995ba297-d41c-4960-b797-2a0283f98d18 tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] [instance: b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6] Terminating instance [ 1934.236038] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-8453c213-135b-48f6-a29c-da7b36504504 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] [instance: ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1934.236657] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-8453c213-135b-48f6-a29c-da7b36504504 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] [instance: ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5] Deleting contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1934.237201] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-8453c213-135b-48f6-a29c-da7b36504504 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Deleting the datastore file [datastore1] ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1934.238915] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8f05eb76-b5c4-4975-a73a-d135d9bf95f0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1934.247081] env[62824]: DEBUG oslo_vmware.api [None req-8453c213-135b-48f6-a29c-da7b36504504 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Waiting for the task: (returnval){ [ 1934.247081] env[62824]: value = "task-2145817" [ 1934.247081] env[62824]: _type = "Task" [ 1934.247081] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1934.259672] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 7febb8b4-f984-4d79-a888-b2829f2a9df6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1934.261297] env[62824]: DEBUG oslo_vmware.api [None req-8453c213-135b-48f6-a29c-da7b36504504 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Task: {'id': task-2145817, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1934.264333] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8b9834a0-6525-490b-8619-56a995b26b7c tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Lock "b8cc8cd3-ea03-40bf-b867-7ad193365552" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 69.143s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1934.298551] env[62824]: DEBUG oslo_vmware.api [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Task: {'id': task-2145815, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1934.497017] env[62824]: DEBUG nova.compute.manager [None req-e84c4fa9-f5ca-4343-835a-d82a25102db8 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1934.500017] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a0e068c-ca17-4b7e-bb3b-3a67ed7e600c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1934.577199] env[62824]: DEBUG oslo_concurrency.lockutils [None req-be0fa035-e0fc-46c3-b3d9-c94dc60ee21d tempest-ServerExternalEventsTest-1049562868 tempest-ServerExternalEventsTest-1049562868-project] Releasing lock "refresh_cache-b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1934.729622] env[62824]: DEBUG nova.compute.manager [None req-995ba297-d41c-4960-b797-2a0283f98d18 tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] [instance: b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1934.732789] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-995ba297-d41c-4960-b797-2a0283f98d18 tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] [instance: b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1934.733988] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb82ef04-89a1-4010-817b-2e9d88133d16 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1934.742586] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-995ba297-d41c-4960-b797-2a0283f98d18 tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] [instance: b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1934.742586] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a3026b1b-9733-4737-af84-b6593ca85ea5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1934.752110] env[62824]: DEBUG oslo_vmware.api [None req-995ba297-d41c-4960-b797-2a0283f98d18 tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Waiting for the task: (returnval){ [ 1934.752110] env[62824]: value = "task-2145818" [ 1934.752110] env[62824]: _type = "Task" [ 1934.752110] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1934.758792] env[62824]: DEBUG oslo_vmware.api [None req-8453c213-135b-48f6-a29c-da7b36504504 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Task: {'id': task-2145817, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.184647} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1934.759063] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-8453c213-135b-48f6-a29c-da7b36504504 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1934.759278] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-8453c213-135b-48f6-a29c-da7b36504504 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] [instance: ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5] Deleted contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1934.759465] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-8453c213-135b-48f6-a29c-da7b36504504 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] [instance: ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1934.759634] env[62824]: INFO nova.compute.manager [None req-8453c213-135b-48f6-a29c-da7b36504504 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] [instance: ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5] Took 1.16 seconds to destroy the instance on the hypervisor. [ 1934.759869] env[62824]: DEBUG oslo.service.loopingcall [None req-8453c213-135b-48f6-a29c-da7b36504504 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1934.760104] env[62824]: DEBUG nova.compute.manager [-] [instance: ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1934.760213] env[62824]: DEBUG nova.network.neutron [-] [instance: ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1934.765346] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1934.767161] env[62824]: DEBUG oslo_vmware.api [None req-995ba297-d41c-4960-b797-2a0283f98d18 tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Task: {'id': task-2145818, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1934.767583] env[62824]: DEBUG nova.compute.manager [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1934.802099] env[62824]: DEBUG oslo_vmware.api [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Task: {'id': task-2145815, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.540985} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1934.802650] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] c36ece43-3d70-4e67-a740-9057f413c722/c36ece43-3d70-4e67-a740-9057f413c722.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1934.803851] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] [instance: c36ece43-3d70-4e67-a740-9057f413c722] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1934.804373] env[62824]: DEBUG nova.network.neutron [req-3e0674b8-989b-4ef9-a8f5-c50fbefb3d66 req-81c3da41-d6a6-45e2-8fa0-8ddf26d64ba3 service nova] [instance: c36ece43-3d70-4e67-a740-9057f413c722] Updated VIF entry in instance network info cache for port eb387e38-7606-4160-a960-43d51598b964. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1934.804925] env[62824]: DEBUG nova.network.neutron [req-3e0674b8-989b-4ef9-a8f5-c50fbefb3d66 req-81c3da41-d6a6-45e2-8fa0-8ddf26d64ba3 service nova] [instance: c36ece43-3d70-4e67-a740-9057f413c722] Updating instance_info_cache with network_info: [{"id": "eb387e38-7606-4160-a960-43d51598b964", "address": "fa:16:3e:47:61:95", "network": {"id": "b6403dc3-30ef-4c76-b521-380c179f345d", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.139", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "22c5c40c2e004babaa15896c89307303", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1002b79b-224e-41e3-a484-4245a767147a", "external-id": "nsx-vlan-transportzone-353", "segmentation_id": 353, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeb387e38-76", "ovs_interfaceid": "eb387e38-7606-4160-a960-43d51598b964", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1934.806187] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7cde423e-98d8-410a-990b-6b1f5b6f1a88 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1934.816024] env[62824]: DEBUG oslo_vmware.api [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Waiting for the task: (returnval){ [ 1934.816024] env[62824]: value = "task-2145819" [ 1934.816024] env[62824]: _type = "Task" [ 1934.816024] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1934.826302] env[62824]: DEBUG oslo_vmware.api [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Task: {'id': task-2145819, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1935.012735] env[62824]: INFO nova.compute.manager [None req-e84c4fa9-f5ca-4343-835a-d82a25102db8 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] instance snapshotting [ 1935.013159] env[62824]: WARNING nova.compute.manager [None req-e84c4fa9-f5ca-4343-835a-d82a25102db8 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] trying to snapshot a non-running instance: (state: 4 expected: 1) [ 1935.016305] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24495555-39a0-4978-b403-c7fc121f67da {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1935.054979] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43b72afb-a716-439f-b0dc-50957fadd9e5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1935.262695] env[62824]: DEBUG oslo_vmware.api [None req-995ba297-d41c-4960-b797-2a0283f98d18 tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Task: {'id': task-2145818, 'name': PowerOffVM_Task, 'duration_secs': 0.17424} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1935.263039] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-995ba297-d41c-4960-b797-2a0283f98d18 tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] [instance: b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1935.263473] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-995ba297-d41c-4960-b797-2a0283f98d18 tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] [instance: b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1935.263473] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9146ade1-24af-4ad0-8012-04fd1b4769cd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1935.269074] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 8bb90a4d-93c9-4f54-b15e-48fe966e6c97 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1935.290568] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1935.310885] env[62824]: DEBUG oslo_concurrency.lockutils [req-3e0674b8-989b-4ef9-a8f5-c50fbefb3d66 req-81c3da41-d6a6-45e2-8fa0-8ddf26d64ba3 service nova] Releasing lock "refresh_cache-c36ece43-3d70-4e67-a740-9057f413c722" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1935.332360] env[62824]: DEBUG oslo_vmware.api [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Task: {'id': task-2145819, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072028} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1935.332360] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] [instance: c36ece43-3d70-4e67-a740-9057f413c722] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1935.332360] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4653b589-fa63-489f-87ae-1dab8dd5095d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1935.357663] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] [instance: c36ece43-3d70-4e67-a740-9057f413c722] Reconfiguring VM instance instance-0000001e to attach disk [datastore2] c36ece43-3d70-4e67-a740-9057f413c722/c36ece43-3d70-4e67-a740-9057f413c722.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1935.359081] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2beef62c-3436-4373-9b3e-6bdc22aee011 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1935.378891] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-995ba297-d41c-4960-b797-2a0283f98d18 tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] [instance: b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1935.379194] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-995ba297-d41c-4960-b797-2a0283f98d18 tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] [instance: b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6] Deleting contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1935.379288] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-995ba297-d41c-4960-b797-2a0283f98d18 tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Deleting the datastore file [datastore2] b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1935.379915] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-99503bfa-ba4d-4ee0-98dd-d820f7436730 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1935.385474] env[62824]: DEBUG oslo_vmware.api [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Waiting for the task: (returnval){ [ 1935.385474] env[62824]: value = "task-2145821" [ 1935.385474] env[62824]: _type = "Task" [ 1935.385474] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1935.389864] env[62824]: DEBUG oslo_vmware.api [None req-995ba297-d41c-4960-b797-2a0283f98d18 tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Waiting for the task: (returnval){ [ 1935.389864] env[62824]: value = "task-2145822" [ 1935.389864] env[62824]: _type = "Task" [ 1935.389864] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1935.396367] env[62824]: DEBUG oslo_vmware.api [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Task: {'id': task-2145821, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1935.400957] env[62824]: DEBUG oslo_vmware.api [None req-995ba297-d41c-4960-b797-2a0283f98d18 tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Task: {'id': task-2145822, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1935.566506] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e84c4fa9-f5ca-4343-835a-d82a25102db8 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] Creating Snapshot of the VM instance {{(pid=62824) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1935.567513] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-113582ff-4079-4c1d-a896-e5ad989acb57 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1935.575419] env[62824]: DEBUG oslo_vmware.api [None req-e84c4fa9-f5ca-4343-835a-d82a25102db8 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the task: (returnval){ [ 1935.575419] env[62824]: value = "task-2145823" [ 1935.575419] env[62824]: _type = "Task" [ 1935.575419] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1935.587115] env[62824]: DEBUG oslo_vmware.api [None req-e84c4fa9-f5ca-4343-835a-d82a25102db8 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2145823, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1935.627876] env[62824]: DEBUG nova.network.neutron [-] [instance: ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1935.690895] env[62824]: DEBUG oslo_concurrency.lockutils [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Acquiring lock "c65ccd20-b682-420a-9c1a-47a45959197e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1935.690895] env[62824]: DEBUG oslo_concurrency.lockutils [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Lock "c65ccd20-b682-420a-9c1a-47a45959197e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1935.771671] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 2afecec5-763d-4616-b690-41d3101cfc47 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1935.898470] env[62824]: DEBUG oslo_vmware.api [None req-995ba297-d41c-4960-b797-2a0283f98d18 tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Task: {'id': task-2145822, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.208107} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1935.901495] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-995ba297-d41c-4960-b797-2a0283f98d18 tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1935.901706] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-995ba297-d41c-4960-b797-2a0283f98d18 tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] [instance: b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6] Deleted contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1935.901899] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-995ba297-d41c-4960-b797-2a0283f98d18 tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] [instance: b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1935.902098] env[62824]: INFO nova.compute.manager [None req-995ba297-d41c-4960-b797-2a0283f98d18 tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] [instance: b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6] Took 1.17 seconds to destroy the instance on the hypervisor. [ 1935.902356] env[62824]: DEBUG oslo.service.loopingcall [None req-995ba297-d41c-4960-b797-2a0283f98d18 tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1935.902580] env[62824]: DEBUG oslo_vmware.api [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Task: {'id': task-2145821, 'name': ReconfigVM_Task, 'duration_secs': 0.290563} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1935.902792] env[62824]: DEBUG nova.compute.manager [-] [instance: b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1935.902895] env[62824]: DEBUG nova.network.neutron [-] [instance: b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1935.904598] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] [instance: c36ece43-3d70-4e67-a740-9057f413c722] Reconfigured VM instance instance-0000001e to attach disk [datastore2] c36ece43-3d70-4e67-a740-9057f413c722/c36ece43-3d70-4e67-a740-9057f413c722.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1935.905245] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-db05c553-a6fc-458c-9f84-88f3e489a285 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1935.912975] env[62824]: DEBUG oslo_vmware.api [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Waiting for the task: (returnval){ [ 1935.912975] env[62824]: value = "task-2145824" [ 1935.912975] env[62824]: _type = "Task" [ 1935.912975] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1935.922388] env[62824]: DEBUG oslo_vmware.api [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Task: {'id': task-2145824, 'name': Rename_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1935.944468] env[62824]: DEBUG nova.compute.manager [req-c1a131ec-b241-449c-86db-e88e14166b0e req-e3153c3f-9a16-45ec-9aea-ec90913a8aee service nova] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Received event network-changed-60a4fce0-0e63-45af-890c-46ca44ea9a0c {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1935.944778] env[62824]: DEBUG nova.compute.manager [req-c1a131ec-b241-449c-86db-e88e14166b0e req-e3153c3f-9a16-45ec-9aea-ec90913a8aee service nova] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Refreshing instance network info cache due to event network-changed-60a4fce0-0e63-45af-890c-46ca44ea9a0c. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1935.945066] env[62824]: DEBUG oslo_concurrency.lockutils [req-c1a131ec-b241-449c-86db-e88e14166b0e req-e3153c3f-9a16-45ec-9aea-ec90913a8aee service nova] Acquiring lock "refresh_cache-b8cc8cd3-ea03-40bf-b867-7ad193365552" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1935.945239] env[62824]: DEBUG oslo_concurrency.lockutils [req-c1a131ec-b241-449c-86db-e88e14166b0e req-e3153c3f-9a16-45ec-9aea-ec90913a8aee service nova] Acquired lock "refresh_cache-b8cc8cd3-ea03-40bf-b867-7ad193365552" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1935.945402] env[62824]: DEBUG nova.network.neutron [req-c1a131ec-b241-449c-86db-e88e14166b0e req-e3153c3f-9a16-45ec-9aea-ec90913a8aee service nova] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Refreshing network info cache for port 60a4fce0-0e63-45af-890c-46ca44ea9a0c {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1936.086959] env[62824]: DEBUG oslo_vmware.api [None req-e84c4fa9-f5ca-4343-835a-d82a25102db8 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2145823, 'name': CreateSnapshot_Task, 'duration_secs': 0.416772} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1936.087112] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e84c4fa9-f5ca-4343-835a-d82a25102db8 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] Created Snapshot of the VM instance {{(pid=62824) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1936.087989] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b96b7d8e-4fef-4088-b904-0ec55383367c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1936.131766] env[62824]: INFO nova.compute.manager [-] [instance: ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5] Took 1.37 seconds to deallocate network for instance. [ 1936.275119] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 41f068d3-2c8d-46f0-8d84-78531d28b0dd has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1936.423287] env[62824]: DEBUG oslo_vmware.api [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Task: {'id': task-2145824, 'name': Rename_Task, 'duration_secs': 0.146803} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1936.423561] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] [instance: c36ece43-3d70-4e67-a740-9057f413c722] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1936.423798] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ec5eb962-f1ec-4e2f-b081-2419f9fe648d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1936.429850] env[62824]: DEBUG oslo_vmware.api [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Waiting for the task: (returnval){ [ 1936.429850] env[62824]: value = "task-2145825" [ 1936.429850] env[62824]: _type = "Task" [ 1936.429850] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1936.437287] env[62824]: DEBUG oslo_vmware.api [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Task: {'id': task-2145825, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1936.608858] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e84c4fa9-f5ca-4343-835a-d82a25102db8 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] Creating linked-clone VM from snapshot {{(pid=62824) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1936.611761] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-01b49a11-20b2-4e38-abe9-449a1cd6402e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1936.621175] env[62824]: DEBUG oslo_vmware.api [None req-e84c4fa9-f5ca-4343-835a-d82a25102db8 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the task: (returnval){ [ 1936.621175] env[62824]: value = "task-2145826" [ 1936.621175] env[62824]: _type = "Task" [ 1936.621175] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1936.629770] env[62824]: DEBUG oslo_vmware.api [None req-e84c4fa9-f5ca-4343-835a-d82a25102db8 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2145826, 'name': CloneVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1936.638881] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8453c213-135b-48f6-a29c-da7b36504504 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1936.683424] env[62824]: DEBUG nova.network.neutron [req-c1a131ec-b241-449c-86db-e88e14166b0e req-e3153c3f-9a16-45ec-9aea-ec90913a8aee service nova] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Updated VIF entry in instance network info cache for port 60a4fce0-0e63-45af-890c-46ca44ea9a0c. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1936.683950] env[62824]: DEBUG nova.network.neutron [req-c1a131ec-b241-449c-86db-e88e14166b0e req-e3153c3f-9a16-45ec-9aea-ec90913a8aee service nova] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Updating instance_info_cache with network_info: [{"id": "60a4fce0-0e63-45af-890c-46ca44ea9a0c", "address": "fa:16:3e:cf:18:eb", "network": {"id": "db233b0c-27f3-4a76-ae9c-2b5889bf3d3b", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1825671448-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.231", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3240be4f32254f54aa87cb045e3d8f31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7041d198-66a3-40de-bf7d-cfc036e6ed69", "external-id": "nsx-vlan-transportzone-278", "segmentation_id": 278, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap60a4fce0-0e", "ovs_interfaceid": "60a4fce0-0e63-45af-890c-46ca44ea9a0c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1936.711493] env[62824]: DEBUG nova.network.neutron [-] [instance: b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1936.778617] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 35f0435c-fe46-46ab-bc71-c21bb0ad3d55 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1936.941548] env[62824]: DEBUG oslo_vmware.api [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Task: {'id': task-2145825, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1937.132133] env[62824]: DEBUG oslo_vmware.api [None req-e84c4fa9-f5ca-4343-835a-d82a25102db8 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2145826, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1937.187091] env[62824]: DEBUG oslo_concurrency.lockutils [req-c1a131ec-b241-449c-86db-e88e14166b0e req-e3153c3f-9a16-45ec-9aea-ec90913a8aee service nova] Releasing lock "refresh_cache-b8cc8cd3-ea03-40bf-b867-7ad193365552" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1937.187421] env[62824]: DEBUG nova.compute.manager [req-c1a131ec-b241-449c-86db-e88e14166b0e req-e3153c3f-9a16-45ec-9aea-ec90913a8aee service nova] [instance: ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5] Received event network-vif-deleted-e8535330-ccec-4e36-8f78-cf45232f2a44 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1937.214456] env[62824]: INFO nova.compute.manager [-] [instance: b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6] Took 1.31 seconds to deallocate network for instance. [ 1937.282611] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance db018af0-7bdf-40c5-b129-fde15dfb8ab4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1937.282611] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Migration bc55ced6-f1ca-4506-869b-2d7d5683b841 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1937.282611] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance e158b5d4-c120-4e6c-89c4-7668e097926d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1937.441548] env[62824]: DEBUG oslo_vmware.api [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Task: {'id': task-2145825, 'name': PowerOnVM_Task, 'duration_secs': 0.529284} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1937.441824] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] [instance: c36ece43-3d70-4e67-a740-9057f413c722] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1937.442058] env[62824]: INFO nova.compute.manager [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] [instance: c36ece43-3d70-4e67-a740-9057f413c722] Took 7.26 seconds to spawn the instance on the hypervisor. [ 1937.442357] env[62824]: DEBUG nova.compute.manager [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] [instance: c36ece43-3d70-4e67-a740-9057f413c722] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1937.443341] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc0cf088-811d-40ee-86a4-f03deb260cf8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1937.631762] env[62824]: DEBUG oslo_vmware.api [None req-e84c4fa9-f5ca-4343-835a-d82a25102db8 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2145826, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1937.722081] env[62824]: DEBUG oslo_concurrency.lockutils [None req-995ba297-d41c-4960-b797-2a0283f98d18 tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1937.787739] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance b9f30f87-3594-4468-9d29-70890d8761e3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1937.960751] env[62824]: INFO nova.compute.manager [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] [instance: c36ece43-3d70-4e67-a740-9057f413c722] Took 33.51 seconds to build instance. [ 1937.987913] env[62824]: DEBUG nova.compute.manager [req-53f0f7d0-4b3a-4234-95d3-fcfecefbce52 req-a8607120-f242-4339-8b0d-1b7f0596f008 service nova] [instance: b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6] Received event network-vif-deleted-c21b7203-99b2-485d-8a8f-db1f6bcdbbe0 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1938.133067] env[62824]: DEBUG oslo_vmware.api [None req-e84c4fa9-f5ca-4343-835a-d82a25102db8 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2145826, 'name': CloneVM_Task, 'duration_secs': 1.404815} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1938.133367] env[62824]: INFO nova.virt.vmwareapi.vmops [None req-e84c4fa9-f5ca-4343-835a-d82a25102db8 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] Created linked-clone VM from snapshot [ 1938.134119] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a8eb12a-f4c7-45e5-b26a-a57ab6d6be5c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1938.141519] env[62824]: DEBUG nova.virt.vmwareapi.images [None req-e84c4fa9-f5ca-4343-835a-d82a25102db8 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] Uploading image b918e7a3-3ced-4879-ba36-91fdfd2da288 {{(pid=62824) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1938.161470] env[62824]: DEBUG oslo_vmware.rw_handles [None req-e84c4fa9-f5ca-4343-835a-d82a25102db8 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1938.161470] env[62824]: value = "vm-438566" [ 1938.161470] env[62824]: _type = "VirtualMachine" [ 1938.161470] env[62824]: }. {{(pid=62824) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1938.161728] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-494109aa-1daa-4e58-9d67-8f27490f5128 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1938.168690] env[62824]: DEBUG oslo_vmware.rw_handles [None req-e84c4fa9-f5ca-4343-835a-d82a25102db8 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Lease: (returnval){ [ 1938.168690] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52d67d14-682e-92ff-e7ae-f08101517d4d" [ 1938.168690] env[62824]: _type = "HttpNfcLease" [ 1938.168690] env[62824]: } obtained for exporting VM: (result){ [ 1938.168690] env[62824]: value = "vm-438566" [ 1938.168690] env[62824]: _type = "VirtualMachine" [ 1938.168690] env[62824]: }. {{(pid=62824) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1938.168963] env[62824]: DEBUG oslo_vmware.api [None req-e84c4fa9-f5ca-4343-835a-d82a25102db8 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the lease: (returnval){ [ 1938.168963] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52d67d14-682e-92ff-e7ae-f08101517d4d" [ 1938.168963] env[62824]: _type = "HttpNfcLease" [ 1938.168963] env[62824]: } to be ready. {{(pid=62824) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1938.175316] env[62824]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1938.175316] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52d67d14-682e-92ff-e7ae-f08101517d4d" [ 1938.175316] env[62824]: _type = "HttpNfcLease" [ 1938.175316] env[62824]: } is initializing. {{(pid=62824) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1938.290398] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 02405b17-7022-4e2c-a357-415de88f63d0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1938.290676] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Total usable vcpus: 48, total allocated vcpus: 14 {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1938.290825] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3200MB phys_disk=200GB used_disk=14GB total_vcpus=48 used_vcpus=14 pci_stats=[] {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1938.464047] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2094d162-d263-46e8-9604-cab54e7ec5fd tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Lock "c36ece43-3d70-4e67-a740-9057f413c722" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 72.936s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1938.680026] env[62824]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1938.680026] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52d67d14-682e-92ff-e7ae-f08101517d4d" [ 1938.680026] env[62824]: _type = "HttpNfcLease" [ 1938.680026] env[62824]: } is ready. {{(pid=62824) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1938.680026] env[62824]: DEBUG oslo_vmware.rw_handles [None req-e84c4fa9-f5ca-4343-835a-d82a25102db8 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1938.680026] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52d67d14-682e-92ff-e7ae-f08101517d4d" [ 1938.680026] env[62824]: _type = "HttpNfcLease" [ 1938.680026] env[62824]: }. {{(pid=62824) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1938.680326] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcba064c-4be7-4001-9e05-45d0f966bcaf {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1938.690412] env[62824]: DEBUG oslo_vmware.rw_handles [None req-e84c4fa9-f5ca-4343-835a-d82a25102db8 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52941f11-5ad4-3873-2134-894d413e38a2/disk-0.vmdk from lease info. {{(pid=62824) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1938.690593] env[62824]: DEBUG oslo_vmware.rw_handles [None req-e84c4fa9-f5ca-4343-835a-d82a25102db8 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52941f11-5ad4-3873-2134-894d413e38a2/disk-0.vmdk for reading. {{(pid=62824) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1938.778060] env[62824]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-6459d56b-5c27-4af4-962c-7687c9fa80a2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1938.887760] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d79eedc8-715c-41f4-b8ab-79cd85c975ac {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1938.895258] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5296439c-1905-45fe-8ae9-e3f1e30d594a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1938.928259] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44ffc233-8d16-4748-9229-041fc4b28d8e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1938.938874] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e314c70a-7ad1-449b-b00b-d580a82f007b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1938.953216] env[62824]: DEBUG nova.compute.provider_tree [None req-41517485-76de-4e8f-a181-a012faec848d None None] Updating inventory in ProviderTree for provider bf95157b-a274-42de-9ccf-9851128a44a1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1938.958430] env[62824]: DEBUG oslo_concurrency.lockutils [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Acquiring lock "4423d172-acdd-4c69-b3b8-ff166e1b8548" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1938.958648] env[62824]: DEBUG oslo_concurrency.lockutils [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Lock "4423d172-acdd-4c69-b3b8-ff166e1b8548" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1938.965724] env[62824]: DEBUG nova.compute.manager [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1939.483991] env[62824]: ERROR nova.scheduler.client.report [None req-41517485-76de-4e8f-a181-a012faec848d None None] [req-2eea9661-c1de-41a3-bafc-4a785a81df48] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID bf95157b-a274-42de-9ccf-9851128a44a1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-2eea9661-c1de-41a3-bafc-4a785a81df48"}]} [ 1939.494199] env[62824]: DEBUG oslo_concurrency.lockutils [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1939.508612] env[62824]: DEBUG nova.scheduler.client.report [None req-41517485-76de-4e8f-a181-a012faec848d None None] Refreshing inventories for resource provider bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1939.523643] env[62824]: DEBUG nova.scheduler.client.report [None req-41517485-76de-4e8f-a181-a012faec848d None None] Updating ProviderTree inventory for provider bf95157b-a274-42de-9ccf-9851128a44a1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1939.523874] env[62824]: DEBUG nova.compute.provider_tree [None req-41517485-76de-4e8f-a181-a012faec848d None None] Updating inventory in ProviderTree for provider bf95157b-a274-42de-9ccf-9851128a44a1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1939.538393] env[62824]: DEBUG nova.scheduler.client.report [None req-41517485-76de-4e8f-a181-a012faec848d None None] Refreshing aggregate associations for resource provider bf95157b-a274-42de-9ccf-9851128a44a1, aggregates: None {{(pid=62824) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1939.565271] env[62824]: DEBUG nova.scheduler.client.report [None req-41517485-76de-4e8f-a181-a012faec848d None None] Refreshing trait associations for resource provider bf95157b-a274-42de-9ccf-9851128a44a1, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_ISO {{(pid=62824) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1940.123079] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1499350b-b42f-4096-b144-2a2439a04bde {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1940.131033] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-549d9872-9ac0-4a26-9e5c-aa14f13792d0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1940.162587] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bf56ce0-aed3-4a7e-9163-1a562242aeae {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1940.170891] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9914411d-f7b9-48a7-a724-0efa2dfcfbca {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1940.185627] env[62824]: DEBUG nova.compute.provider_tree [None req-41517485-76de-4e8f-a181-a012faec848d None None] Updating inventory in ProviderTree for provider bf95157b-a274-42de-9ccf-9851128a44a1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1940.720913] env[62824]: DEBUG nova.scheduler.client.report [None req-41517485-76de-4e8f-a181-a012faec848d None None] Updated inventory for provider bf95157b-a274-42de-9ccf-9851128a44a1 with generation 61 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 1940.721381] env[62824]: DEBUG nova.compute.provider_tree [None req-41517485-76de-4e8f-a181-a012faec848d None None] Updating resource provider bf95157b-a274-42de-9ccf-9851128a44a1 generation from 61 to 62 during operation: update_inventory {{(pid=62824) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1940.721615] env[62824]: DEBUG nova.compute.provider_tree [None req-41517485-76de-4e8f-a181-a012faec848d None None] Updating inventory in ProviderTree for provider bf95157b-a274-42de-9ccf-9851128a44a1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1941.229553] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62824) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1941.229789] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 13.082s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1941.230136] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.288s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1941.231819] env[62824]: INFO nova.compute.claims [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] [instance: b96d1351-f3a7-4bac-998c-a34ab2606041] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1942.776634] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ca8b9e2-a264-47c9-a7f8-6d849d0b7d8a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1942.784949] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b9f43c7-f693-4364-ad65-7f2a25be7448 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1942.818372] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9349bc4-5fc6-4808-8a41-730f1ee70130 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1942.827924] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da1a066c-30b6-49e2-8ad0-b90d11df6aa8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1942.844664] env[62824]: DEBUG nova.compute.provider_tree [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1943.229567] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1943.229811] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1943.229934] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Starting heal instance info cache {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10445}} [ 1943.230067] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Rebuilding the list of instances to heal {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10449}} [ 1943.349050] env[62824]: DEBUG nova.scheduler.client.report [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1943.738364] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: b96d1351-f3a7-4bac-998c-a34ab2606041] Skipping network cache update for instance because it is Building. {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10458}} [ 1943.768514] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "refresh_cache-29b8dd5f-1855-490c-a01b-54840073a753" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1943.768724] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquired lock "refresh_cache-29b8dd5f-1855-490c-a01b-54840073a753" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1943.768949] env[62824]: DEBUG nova.network.neutron [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Forcefully refreshing network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1943.769140] env[62824]: DEBUG nova.objects.instance [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lazy-loading 'info_cache' on Instance uuid 29b8dd5f-1855-490c-a01b-54840073a753 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1943.854734] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.624s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1943.854734] env[62824]: DEBUG nova.compute.manager [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] [instance: b96d1351-f3a7-4bac-998c-a34ab2606041] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1943.857796] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1a251766-5a96-4ca3-afcd-ba6cbf0f0b78 tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 32.714s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1943.858068] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1a251766-5a96-4ca3-afcd-ba6cbf0f0b78 tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1943.860320] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.652s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1943.861810] env[62824]: INFO nova.compute.claims [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1943.890289] env[62824]: INFO nova.scheduler.client.report [None req-1a251766-5a96-4ca3-afcd-ba6cbf0f0b78 tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Deleted allocations for instance 9062e606-511f-4fe7-9621-90b3c0e51eec [ 1944.367646] env[62824]: DEBUG nova.compute.utils [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1944.371192] env[62824]: DEBUG nova.compute.manager [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] [instance: b96d1351-f3a7-4bac-998c-a34ab2606041] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1944.371192] env[62824]: DEBUG nova.network.neutron [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] [instance: b96d1351-f3a7-4bac-998c-a34ab2606041] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1944.395433] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1a251766-5a96-4ca3-afcd-ba6cbf0f0b78 tempest-InstanceActionsTestJSON-57745707 tempest-InstanceActionsTestJSON-57745707-project-member] Lock "9062e606-511f-4fe7-9621-90b3c0e51eec" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 36.859s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1944.419845] env[62824]: DEBUG nova.policy [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '44d9c422d3984a32854e3843fd3330e0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4279c92ea7004c74bdbf768b38d9b372', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 1944.755666] env[62824]: DEBUG nova.network.neutron [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] [instance: b96d1351-f3a7-4bac-998c-a34ab2606041] Successfully created port: c7dfce05-2238-418a-b414-395b1f1fb4f6 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1944.877304] env[62824]: DEBUG nova.compute.manager [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] [instance: b96d1351-f3a7-4bac-998c-a34ab2606041] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1945.276018] env[62824]: DEBUG oslo_vmware.rw_handles [None req-e84c4fa9-f5ca-4343-835a-d82a25102db8 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52941f11-5ad4-3873-2134-894d413e38a2/disk-0.vmdk. {{(pid=62824) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1945.278861] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d634d758-8580-413a-ac0d-c4fe8ea87066 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1945.292348] env[62824]: DEBUG oslo_vmware.rw_handles [None req-e84c4fa9-f5ca-4343-835a-d82a25102db8 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52941f11-5ad4-3873-2134-894d413e38a2/disk-0.vmdk is in state: ready. {{(pid=62824) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1945.292555] env[62824]: ERROR oslo_vmware.rw_handles [None req-e84c4fa9-f5ca-4343-835a-d82a25102db8 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52941f11-5ad4-3873-2134-894d413e38a2/disk-0.vmdk due to incomplete transfer. [ 1945.292798] env[62824]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-cf1531a9-cc6f-4283-bf8e-6ac8c829ef0c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1945.301430] env[62824]: DEBUG oslo_vmware.rw_handles [None req-e84c4fa9-f5ca-4343-835a-d82a25102db8 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52941f11-5ad4-3873-2134-894d413e38a2/disk-0.vmdk. {{(pid=62824) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1945.301628] env[62824]: DEBUG nova.virt.vmwareapi.images [None req-e84c4fa9-f5ca-4343-835a-d82a25102db8 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] Uploaded image b918e7a3-3ced-4879-ba36-91fdfd2da288 to the Glance image server {{(pid=62824) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1945.303661] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-e84c4fa9-f5ca-4343-835a-d82a25102db8 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] Destroying the VM {{(pid=62824) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1945.303911] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-6db8d908-3d2a-4917-be0f-36977b34a2a8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1945.309610] env[62824]: DEBUG oslo_vmware.api [None req-e84c4fa9-f5ca-4343-835a-d82a25102db8 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the task: (returnval){ [ 1945.309610] env[62824]: value = "task-2145828" [ 1945.309610] env[62824]: _type = "Task" [ 1945.309610] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1945.318738] env[62824]: DEBUG oslo_vmware.api [None req-e84c4fa9-f5ca-4343-835a-d82a25102db8 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2145828, 'name': Destroy_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1945.389328] env[62824]: INFO nova.virt.block_device [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] [instance: b96d1351-f3a7-4bac-998c-a34ab2606041] Booting with volume 64941ab3-102e-48a5-a3ad-433fb818850b at /dev/sda [ 1945.433231] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe475ff6-62e4-43c1-bba2-3428ec41c93f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1945.437733] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f75be9b1-db8f-4490-9d95-f77926c4fa89 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1945.445220] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3426c391-6e07-4710-8ac7-c803bfc498bb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1945.450925] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-513caf49-e519-41ec-8b20-108e15354adf {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1945.501626] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a271f2b0-abc8-47f3-9f49-b79a2970dadb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1945.504267] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1634a63c-dc0f-469d-be97-65662268ffbd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1945.512114] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea98b256-4d6e-4ece-80d1-c3cc12cbdd3f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1945.519380] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a17ae624-125a-4be5-a8c6-75cb6559f5ff {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1945.538954] env[62824]: DEBUG nova.compute.provider_tree [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1945.551429] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77025147-a759-4456-b3b7-e192cfabb66d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1945.558524] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d760a828-9813-4948-927e-2faa88b1928e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1945.573208] env[62824]: DEBUG nova.virt.block_device [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] [instance: b96d1351-f3a7-4bac-998c-a34ab2606041] Updating existing volume attachment record: b7919e55-53e5-4aad-8562-b40291e3954b {{(pid=62824) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1945.726966] env[62824]: DEBUG nova.network.neutron [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Updating instance_info_cache with network_info: [{"id": "2d861cfc-3ad4-4d40-ad77-e4530d363421", "address": "fa:16:3e:59:84:dd", "network": {"id": "f18b4230-e950-4957-a02b-107d27729346", "bridge": "br-int", "label": "tempest-VolumesAssistedSnapshotsTest-633702241-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.144", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e2e61ea652f94257b884d2e1b495e446", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4b5c60ce-845e-4506-bc10-348461fece6d", "external-id": "nsx-vlan-transportzone-831", "segmentation_id": 831, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2d861cfc-3a", "ovs_interfaceid": "2d861cfc-3ad4-4d40-ad77-e4530d363421", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1945.824361] env[62824]: DEBUG oslo_vmware.api [None req-e84c4fa9-f5ca-4343-835a-d82a25102db8 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2145828, 'name': Destroy_Task, 'duration_secs': 0.453437} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1945.824361] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-e84c4fa9-f5ca-4343-835a-d82a25102db8 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] Destroyed the VM [ 1945.824361] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e84c4fa9-f5ca-4343-835a-d82a25102db8 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] Deleting Snapshot of the VM instance {{(pid=62824) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1945.824361] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-82c9a868-87b1-4a09-9127-c83bb50522e4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1945.829023] env[62824]: DEBUG oslo_vmware.api [None req-e84c4fa9-f5ca-4343-835a-d82a25102db8 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the task: (returnval){ [ 1945.829023] env[62824]: value = "task-2145829" [ 1945.829023] env[62824]: _type = "Task" [ 1945.829023] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1945.837526] env[62824]: DEBUG oslo_vmware.api [None req-e84c4fa9-f5ca-4343-835a-d82a25102db8 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2145829, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1946.052449] env[62824]: DEBUG nova.scheduler.client.report [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1946.232899] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Releasing lock "refresh_cache-29b8dd5f-1855-490c-a01b-54840073a753" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1946.233149] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Updated the network info_cache for instance {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10516}} [ 1946.233352] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1946.233499] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1946.233637] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1946.233775] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1946.233951] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1946.234089] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62824) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11064}} [ 1946.339440] env[62824]: DEBUG oslo_vmware.api [None req-e84c4fa9-f5ca-4343-835a-d82a25102db8 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2145829, 'name': RemoveSnapshot_Task, 'duration_secs': 0.375495} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1946.339718] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e84c4fa9-f5ca-4343-835a-d82a25102db8 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] Deleted Snapshot of the VM instance {{(pid=62824) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1946.339987] env[62824]: INFO nova.compute.manager [None req-e84c4fa9-f5ca-4343-835a-d82a25102db8 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] Took 11.32 seconds to snapshot the instance on the hypervisor. [ 1946.536485] env[62824]: DEBUG nova.network.neutron [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] [instance: b96d1351-f3a7-4bac-998c-a34ab2606041] Successfully updated port: c7dfce05-2238-418a-b414-395b1f1fb4f6 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1946.557887] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.697s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1946.558446] env[62824]: DEBUG nova.compute.manager [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1946.560842] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.884s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1946.562556] env[62824]: INFO nova.compute.claims [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: 6d80ec06-8559-4964-8577-a2512aa366ed] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1946.614650] env[62824]: DEBUG nova.compute.manager [req-6f65e491-8afd-45f3-97bb-3b5efd6aa2ed req-e28cb905-62be-47a0-a94f-d6375530d49c service nova] [instance: b96d1351-f3a7-4bac-998c-a34ab2606041] Received event network-vif-plugged-c7dfce05-2238-418a-b414-395b1f1fb4f6 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1946.614882] env[62824]: DEBUG oslo_concurrency.lockutils [req-6f65e491-8afd-45f3-97bb-3b5efd6aa2ed req-e28cb905-62be-47a0-a94f-d6375530d49c service nova] Acquiring lock "b96d1351-f3a7-4bac-998c-a34ab2606041-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1946.615117] env[62824]: DEBUG oslo_concurrency.lockutils [req-6f65e491-8afd-45f3-97bb-3b5efd6aa2ed req-e28cb905-62be-47a0-a94f-d6375530d49c service nova] Lock "b96d1351-f3a7-4bac-998c-a34ab2606041-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1946.615301] env[62824]: DEBUG oslo_concurrency.lockutils [req-6f65e491-8afd-45f3-97bb-3b5efd6aa2ed req-e28cb905-62be-47a0-a94f-d6375530d49c service nova] Lock "b96d1351-f3a7-4bac-998c-a34ab2606041-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1946.615468] env[62824]: DEBUG nova.compute.manager [req-6f65e491-8afd-45f3-97bb-3b5efd6aa2ed req-e28cb905-62be-47a0-a94f-d6375530d49c service nova] [instance: b96d1351-f3a7-4bac-998c-a34ab2606041] No waiting events found dispatching network-vif-plugged-c7dfce05-2238-418a-b414-395b1f1fb4f6 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1946.615638] env[62824]: WARNING nova.compute.manager [req-6f65e491-8afd-45f3-97bb-3b5efd6aa2ed req-e28cb905-62be-47a0-a94f-d6375530d49c service nova] [instance: b96d1351-f3a7-4bac-998c-a34ab2606041] Received unexpected event network-vif-plugged-c7dfce05-2238-418a-b414-395b1f1fb4f6 for instance with vm_state building and task_state block_device_mapping. [ 1946.688475] env[62824]: DEBUG oslo_concurrency.lockutils [None req-89ab4335-c79f-4e08-ad2d-7263741cbc31 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Acquiring lock "16b466c3-c749-4f96-a82c-32dad31138ec" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1946.688475] env[62824]: DEBUG oslo_concurrency.lockutils [None req-89ab4335-c79f-4e08-ad2d-7263741cbc31 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Lock "16b466c3-c749-4f96-a82c-32dad31138ec" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1946.688803] env[62824]: DEBUG oslo_concurrency.lockutils [None req-89ab4335-c79f-4e08-ad2d-7263741cbc31 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Acquiring lock "16b466c3-c749-4f96-a82c-32dad31138ec-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1946.688865] env[62824]: DEBUG oslo_concurrency.lockutils [None req-89ab4335-c79f-4e08-ad2d-7263741cbc31 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Lock "16b466c3-c749-4f96-a82c-32dad31138ec-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1946.688972] env[62824]: DEBUG oslo_concurrency.lockutils [None req-89ab4335-c79f-4e08-ad2d-7263741cbc31 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Lock "16b466c3-c749-4f96-a82c-32dad31138ec-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1946.692651] env[62824]: INFO nova.compute.manager [None req-89ab4335-c79f-4e08-ad2d-7263741cbc31 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] Terminating instance [ 1946.845660] env[62824]: DEBUG nova.compute.manager [None req-e84c4fa9-f5ca-4343-835a-d82a25102db8 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] Instance disappeared during snapshot {{(pid=62824) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4595}} [ 1946.860508] env[62824]: DEBUG nova.compute.manager [None req-e84c4fa9-f5ca-4343-835a-d82a25102db8 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Image not found during clean up b918e7a3-3ced-4879-ba36-91fdfd2da288 {{(pid=62824) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4601}} [ 1947.040275] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Acquiring lock "refresh_cache-b96d1351-f3a7-4bac-998c-a34ab2606041" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1947.040438] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Acquired lock "refresh_cache-b96d1351-f3a7-4bac-998c-a34ab2606041" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1947.040595] env[62824]: DEBUG nova.network.neutron [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] [instance: b96d1351-f3a7-4bac-998c-a34ab2606041] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1947.067357] env[62824]: DEBUG nova.compute.utils [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1947.072188] env[62824]: DEBUG nova.compute.manager [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1947.072188] env[62824]: DEBUG nova.network.neutron [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1947.123595] env[62824]: DEBUG nova.policy [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6ac24e1886434785b90dfba135fb7c8f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e10208040df34eb5adfe8dcbc76043d5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 1947.196087] env[62824]: DEBUG nova.compute.manager [None req-89ab4335-c79f-4e08-ad2d-7263741cbc31 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1947.196357] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-89ab4335-c79f-4e08-ad2d-7263741cbc31 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1947.197213] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb8db438-2e3b-4a75-81f8-5af93c4c01a6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1947.205297] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-89ab4335-c79f-4e08-ad2d-7263741cbc31 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1947.205541] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ef27d3f9-698f-4476-a1ad-a30ab9b07292 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1947.281485] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-89ab4335-c79f-4e08-ad2d-7263741cbc31 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1947.281926] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-89ab4335-c79f-4e08-ad2d-7263741cbc31 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] Deleting contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1947.282099] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-89ab4335-c79f-4e08-ad2d-7263741cbc31 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Deleting the datastore file [datastore1] 16b466c3-c749-4f96-a82c-32dad31138ec {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1947.283429] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c91b545c-4a1a-4237-b18b-b180280d5be5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1947.290103] env[62824]: DEBUG oslo_vmware.api [None req-89ab4335-c79f-4e08-ad2d-7263741cbc31 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the task: (returnval){ [ 1947.290103] env[62824]: value = "task-2145831" [ 1947.290103] env[62824]: _type = "Task" [ 1947.290103] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1947.298548] env[62824]: DEBUG oslo_vmware.api [None req-89ab4335-c79f-4e08-ad2d-7263741cbc31 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2145831, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1947.320639] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Acquiring lock "88bad2dd-dce7-41df-b56c-93a5d054c11e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1947.320937] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Lock "88bad2dd-dce7-41df-b56c-93a5d054c11e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1947.517750] env[62824]: DEBUG nova.network.neutron [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Successfully created port: 4e0d51df-4419-4059-9dab-be44957e4f77 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1947.572113] env[62824]: DEBUG nova.compute.manager [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1947.600370] env[62824]: DEBUG nova.network.neutron [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] [instance: b96d1351-f3a7-4bac-998c-a34ab2606041] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1947.698334] env[62824]: DEBUG nova.compute.manager [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] [instance: b96d1351-f3a7-4bac-998c-a34ab2606041] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1947.698892] env[62824]: DEBUG nova.virt.hardware [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1947.699146] env[62824]: DEBUG nova.virt.hardware [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1947.699323] env[62824]: DEBUG nova.virt.hardware [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1947.699518] env[62824]: DEBUG nova.virt.hardware [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1947.699671] env[62824]: DEBUG nova.virt.hardware [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1947.699874] env[62824]: DEBUG nova.virt.hardware [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1947.700117] env[62824]: DEBUG nova.virt.hardware [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1947.700282] env[62824]: DEBUG nova.virt.hardware [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1947.700461] env[62824]: DEBUG nova.virt.hardware [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1947.702191] env[62824]: DEBUG nova.virt.hardware [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1947.702191] env[62824]: DEBUG nova.virt.hardware [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1947.702191] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d626a7bc-9944-46d4-88bc-e3fd9976724d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1947.712948] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aca81c74-6e0b-4883-926c-9d74a8ea5df1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1947.799412] env[62824]: DEBUG oslo_vmware.api [None req-89ab4335-c79f-4e08-ad2d-7263741cbc31 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2145831, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.141071} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1947.799670] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-89ab4335-c79f-4e08-ad2d-7263741cbc31 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1947.799856] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-89ab4335-c79f-4e08-ad2d-7263741cbc31 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] Deleted contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1947.800100] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-89ab4335-c79f-4e08-ad2d-7263741cbc31 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1947.800277] env[62824]: INFO nova.compute.manager [None req-89ab4335-c79f-4e08-ad2d-7263741cbc31 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] Took 0.60 seconds to destroy the instance on the hypervisor. [ 1947.800560] env[62824]: DEBUG oslo.service.loopingcall [None req-89ab4335-c79f-4e08-ad2d-7263741cbc31 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1947.803380] env[62824]: DEBUG nova.compute.manager [-] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1947.803380] env[62824]: DEBUG nova.network.neutron [-] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1947.811061] env[62824]: DEBUG nova.network.neutron [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] [instance: b96d1351-f3a7-4bac-998c-a34ab2606041] Updating instance_info_cache with network_info: [{"id": "c7dfce05-2238-418a-b414-395b1f1fb4f6", "address": "fa:16:3e:34:b9:b4", "network": {"id": "0a7184c9-f122-4c98-a7d1-3637da9e3669", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-1696959333-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4279c92ea7004c74bdbf768b38d9b372", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a2e2e51-010f-4535-ba88-433663275996", "external-id": "nsx-vlan-transportzone-915", "segmentation_id": 915, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc7dfce05-22", "ovs_interfaceid": "c7dfce05-2238-418a-b414-395b1f1fb4f6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1948.130738] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ee07717-ef8d-4249-bd53-e5c463fda3eb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1948.140151] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57d9ba66-08d2-4b54-bdcb-0ed50bcee95a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1948.171150] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6107054-a664-41b4-a609-841777cb15e1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1948.179890] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6a81f61-7191-4a99-8d65-cf1587cc4d7a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1948.193814] env[62824]: DEBUG nova.compute.provider_tree [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1948.315482] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Releasing lock "refresh_cache-b96d1351-f3a7-4bac-998c-a34ab2606041" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1948.315862] env[62824]: DEBUG nova.compute.manager [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] [instance: b96d1351-f3a7-4bac-998c-a34ab2606041] Instance network_info: |[{"id": "c7dfce05-2238-418a-b414-395b1f1fb4f6", "address": "fa:16:3e:34:b9:b4", "network": {"id": "0a7184c9-f122-4c98-a7d1-3637da9e3669", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-1696959333-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4279c92ea7004c74bdbf768b38d9b372", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a2e2e51-010f-4535-ba88-433663275996", "external-id": "nsx-vlan-transportzone-915", "segmentation_id": 915, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc7dfce05-22", "ovs_interfaceid": "c7dfce05-2238-418a-b414-395b1f1fb4f6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1948.316308] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] [instance: b96d1351-f3a7-4bac-998c-a34ab2606041] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:34:b9:b4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6a2e2e51-010f-4535-ba88-433663275996', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c7dfce05-2238-418a-b414-395b1f1fb4f6', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1948.324924] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Creating folder: Project (4279c92ea7004c74bdbf768b38d9b372). Parent ref: group-v438503. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1948.325237] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-491d90cd-95f5-4f3c-9b2d-16e95be7abb8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1948.338723] env[62824]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 1948.339067] env[62824]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=62824) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 1948.339474] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Folder already exists: Project (4279c92ea7004c74bdbf768b38d9b372). Parent ref: group-v438503. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1609}} [ 1948.339710] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Creating folder: Instances. Parent ref: group-v438507. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1948.339985] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f43373a2-b4f9-4cd2-a274-eeb86fe4aa2e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1948.349673] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Created folder: Instances in parent group-v438507. [ 1948.349996] env[62824]: DEBUG oslo.service.loopingcall [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1948.350234] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b96d1351-f3a7-4bac-998c-a34ab2606041] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1948.350846] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2873b290-65cb-4ec6-93c1-8082e58497e4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1948.371105] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1948.371105] env[62824]: value = "task-2145834" [ 1948.371105] env[62824]: _type = "Task" [ 1948.371105] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1948.380870] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145834, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1948.583419] env[62824]: DEBUG nova.compute.manager [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1948.614344] env[62824]: DEBUG nova.virt.hardware [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1948.614718] env[62824]: DEBUG nova.virt.hardware [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1948.614904] env[62824]: DEBUG nova.virt.hardware [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1948.615168] env[62824]: DEBUG nova.virt.hardware [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1948.615372] env[62824]: DEBUG nova.virt.hardware [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1948.615563] env[62824]: DEBUG nova.virt.hardware [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1948.615842] env[62824]: DEBUG nova.virt.hardware [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1948.615923] env[62824]: DEBUG nova.virt.hardware [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1948.616091] env[62824]: DEBUG nova.virt.hardware [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1948.616295] env[62824]: DEBUG nova.virt.hardware [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1948.616420] env[62824]: DEBUG nova.virt.hardware [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1948.617289] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29964bd2-d137-495f-8258-4a51e94f2369 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1948.625497] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ca9a872-7410-48dc-8c03-d82624cf14b2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1948.659102] env[62824]: DEBUG nova.compute.manager [req-b4722135-7e21-4106-a65f-ab3a5af26f86 req-5cf3a1b2-f3da-4baf-9b76-9041827caa75 service nova] [instance: b96d1351-f3a7-4bac-998c-a34ab2606041] Received event network-changed-c7dfce05-2238-418a-b414-395b1f1fb4f6 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1948.659377] env[62824]: DEBUG nova.compute.manager [req-b4722135-7e21-4106-a65f-ab3a5af26f86 req-5cf3a1b2-f3da-4baf-9b76-9041827caa75 service nova] [instance: b96d1351-f3a7-4bac-998c-a34ab2606041] Refreshing instance network info cache due to event network-changed-c7dfce05-2238-418a-b414-395b1f1fb4f6. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1948.659407] env[62824]: DEBUG oslo_concurrency.lockutils [req-b4722135-7e21-4106-a65f-ab3a5af26f86 req-5cf3a1b2-f3da-4baf-9b76-9041827caa75 service nova] Acquiring lock "refresh_cache-b96d1351-f3a7-4bac-998c-a34ab2606041" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1948.659693] env[62824]: DEBUG oslo_concurrency.lockutils [req-b4722135-7e21-4106-a65f-ab3a5af26f86 req-5cf3a1b2-f3da-4baf-9b76-9041827caa75 service nova] Acquired lock "refresh_cache-b96d1351-f3a7-4bac-998c-a34ab2606041" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1948.659790] env[62824]: DEBUG nova.network.neutron [req-b4722135-7e21-4106-a65f-ab3a5af26f86 req-5cf3a1b2-f3da-4baf-9b76-9041827caa75 service nova] [instance: b96d1351-f3a7-4bac-998c-a34ab2606041] Refreshing network info cache for port c7dfce05-2238-418a-b414-395b1f1fb4f6 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1948.696955] env[62824]: DEBUG nova.scheduler.client.report [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1948.882909] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145834, 'name': CreateVM_Task, 'duration_secs': 0.388238} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1948.883351] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b96d1351-f3a7-4bac-998c-a34ab2606041] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1948.884249] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] [instance: b96d1351-f3a7-4bac-998c-a34ab2606041] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'device_type': None, 'boot_index': 0, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-438510', 'volume_id': '64941ab3-102e-48a5-a3ad-433fb818850b', 'name': 'volume-64941ab3-102e-48a5-a3ad-433fb818850b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'b96d1351-f3a7-4bac-998c-a34ab2606041', 'attached_at': '', 'detached_at': '', 'volume_id': '64941ab3-102e-48a5-a3ad-433fb818850b', 'serial': '64941ab3-102e-48a5-a3ad-433fb818850b'}, 'disk_bus': None, 'mount_device': '/dev/sda', 'delete_on_termination': True, 'attachment_id': 'b7919e55-53e5-4aad-8562-b40291e3954b', 'guest_format': None, 'volume_type': None}], 'swap': None} {{(pid=62824) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1948.885449] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] [instance: b96d1351-f3a7-4bac-998c-a34ab2606041] Root volume attach. Driver type: vmdk {{(pid=62824) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 1948.891187] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35387fb3-963d-4ce0-8998-29524be62f9c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1948.899686] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c09b9ea-9bf6-471d-a817-eb1bd499a0be {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1948.902290] env[62824]: DEBUG nova.network.neutron [-] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1948.909315] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d3b9c36-e9f8-42dd-b7a8-a2e9825ed53d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1948.916534] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-5c023d73-0320-4677-bba4-a6ce5dff3857 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1948.923738] env[62824]: DEBUG oslo_vmware.api [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Waiting for the task: (returnval){ [ 1948.923738] env[62824]: value = "task-2145835" [ 1948.923738] env[62824]: _type = "Task" [ 1948.923738] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1948.932068] env[62824]: DEBUG oslo_vmware.api [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Task: {'id': task-2145835, 'name': RelocateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1949.201796] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.641s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1949.202412] env[62824]: DEBUG nova.compute.manager [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: 6d80ec06-8559-4964-8577-a2512aa366ed] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1949.205675] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 33.930s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1949.206056] env[62824]: DEBUG nova.objects.instance [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62824) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1949.303748] env[62824]: DEBUG nova.network.neutron [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Successfully updated port: 4e0d51df-4419-4059-9dab-be44957e4f77 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1949.408926] env[62824]: INFO nova.compute.manager [-] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] Took 1.60 seconds to deallocate network for instance. [ 1949.444799] env[62824]: DEBUG oslo_vmware.api [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Task: {'id': task-2145835, 'name': RelocateVM_Task} progress is 43%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1949.592349] env[62824]: DEBUG nova.network.neutron [req-b4722135-7e21-4106-a65f-ab3a5af26f86 req-5cf3a1b2-f3da-4baf-9b76-9041827caa75 service nova] [instance: b96d1351-f3a7-4bac-998c-a34ab2606041] Updated VIF entry in instance network info cache for port c7dfce05-2238-418a-b414-395b1f1fb4f6. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1949.594578] env[62824]: DEBUG nova.network.neutron [req-b4722135-7e21-4106-a65f-ab3a5af26f86 req-5cf3a1b2-f3da-4baf-9b76-9041827caa75 service nova] [instance: b96d1351-f3a7-4bac-998c-a34ab2606041] Updating instance_info_cache with network_info: [{"id": "c7dfce05-2238-418a-b414-395b1f1fb4f6", "address": "fa:16:3e:34:b9:b4", "network": {"id": "0a7184c9-f122-4c98-a7d1-3637da9e3669", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-1696959333-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4279c92ea7004c74bdbf768b38d9b372", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a2e2e51-010f-4535-ba88-433663275996", "external-id": "nsx-vlan-transportzone-915", "segmentation_id": 915, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc7dfce05-22", "ovs_interfaceid": "c7dfce05-2238-418a-b414-395b1f1fb4f6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1949.714290] env[62824]: DEBUG nova.compute.utils [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1949.717360] env[62824]: DEBUG nova.compute.manager [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: 6d80ec06-8559-4964-8577-a2512aa366ed] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1949.717858] env[62824]: DEBUG nova.network.neutron [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: 6d80ec06-8559-4964-8577-a2512aa366ed] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1949.795166] env[62824]: DEBUG nova.policy [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6ac24e1886434785b90dfba135fb7c8f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e10208040df34eb5adfe8dcbc76043d5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 1949.810987] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Acquiring lock "refresh_cache-b3003c4b-ae5a-48df-8c12-a915a76253f4" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1949.810987] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Acquired lock "refresh_cache-b3003c4b-ae5a-48df-8c12-a915a76253f4" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1949.810987] env[62824]: DEBUG nova.network.neutron [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1949.916664] env[62824]: DEBUG oslo_concurrency.lockutils [None req-89ab4335-c79f-4e08-ad2d-7263741cbc31 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1949.938920] env[62824]: DEBUG oslo_vmware.api [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Task: {'id': task-2145835, 'name': RelocateVM_Task} progress is 56%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1950.098096] env[62824]: DEBUG oslo_concurrency.lockutils [req-b4722135-7e21-4106-a65f-ab3a5af26f86 req-5cf3a1b2-f3da-4baf-9b76-9041827caa75 service nova] Releasing lock "refresh_cache-b96d1351-f3a7-4bac-998c-a34ab2606041" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1950.098096] env[62824]: DEBUG nova.compute.manager [req-b4722135-7e21-4106-a65f-ab3a5af26f86 req-5cf3a1b2-f3da-4baf-9b76-9041827caa75 service nova] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] Received event network-vif-deleted-e0e9725d-7e5a-452f-bd1e-960934705a7f {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1950.098096] env[62824]: INFO nova.compute.manager [req-b4722135-7e21-4106-a65f-ab3a5af26f86 req-5cf3a1b2-f3da-4baf-9b76-9041827caa75 service nova] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] Neutron deleted interface e0e9725d-7e5a-452f-bd1e-960934705a7f; detaching it from the instance and deleting it from the info cache [ 1950.098096] env[62824]: DEBUG nova.network.neutron [req-b4722135-7e21-4106-a65f-ab3a5af26f86 req-5cf3a1b2-f3da-4baf-9b76-9041827caa75 service nova] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1950.221243] env[62824]: DEBUG nova.compute.manager [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: 6d80ec06-8559-4964-8577-a2512aa366ed] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1950.226213] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1ce86287-50d5-43c6-bfdd-d665182fa98d tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.019s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1950.227627] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 33.443s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1950.244177] env[62824]: DEBUG nova.network.neutron [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: 6d80ec06-8559-4964-8577-a2512aa366ed] Successfully created port: 24e47f2d-ba7e-48db-9b98-c4f3f6a78930 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1950.362051] env[62824]: DEBUG nova.network.neutron [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1950.444209] env[62824]: DEBUG oslo_vmware.api [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Task: {'id': task-2145835, 'name': RelocateVM_Task} progress is 71%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1950.600727] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1b746ceb-d891-4ede-bbc7-83b7c7e89845 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1950.615532] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dec282fc-7f7b-439c-a0ed-f4d5a9fe2d5a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1950.631236] env[62824]: DEBUG nova.network.neutron [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Updating instance_info_cache with network_info: [{"id": "4e0d51df-4419-4059-9dab-be44957e4f77", "address": "fa:16:3e:15:1a:78", "network": {"id": "a410f4f3-d072-4034-b797-8210e1a79093", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1482574391-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e10208040df34eb5adfe8dcbc76043d5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f4399275-8e92-4448-be9e-d4984e93e89c", "external-id": "nsx-vlan-transportzone-192", "segmentation_id": 192, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e0d51df-44", "ovs_interfaceid": "4e0d51df-4419-4059-9dab-be44957e4f77", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1950.652420] env[62824]: DEBUG nova.compute.manager [req-b4722135-7e21-4106-a65f-ab3a5af26f86 req-5cf3a1b2-f3da-4baf-9b76-9041827caa75 service nova] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] Detach interface failed, port_id=e0e9725d-7e5a-452f-bd1e-960934705a7f, reason: Instance 16b466c3-c749-4f96-a82c-32dad31138ec could not be found. {{(pid=62824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11482}} [ 1950.732930] env[62824]: INFO nova.compute.claims [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1950.773294] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1950.938019] env[62824]: DEBUG oslo_vmware.api [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Task: {'id': task-2145835, 'name': RelocateVM_Task} progress is 84%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1950.991302] env[62824]: DEBUG nova.compute.manager [req-8d7f5f67-00dd-42c1-99ec-124ee744a630 req-6255bcbd-a7f7-4ac7-a456-2f4f14b531fc service nova] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Received event network-vif-plugged-4e0d51df-4419-4059-9dab-be44957e4f77 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1950.991673] env[62824]: DEBUG oslo_concurrency.lockutils [req-8d7f5f67-00dd-42c1-99ec-124ee744a630 req-6255bcbd-a7f7-4ac7-a456-2f4f14b531fc service nova] Acquiring lock "b3003c4b-ae5a-48df-8c12-a915a76253f4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1950.991919] env[62824]: DEBUG oslo_concurrency.lockutils [req-8d7f5f67-00dd-42c1-99ec-124ee744a630 req-6255bcbd-a7f7-4ac7-a456-2f4f14b531fc service nova] Lock "b3003c4b-ae5a-48df-8c12-a915a76253f4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1950.993594] env[62824]: DEBUG oslo_concurrency.lockutils [req-8d7f5f67-00dd-42c1-99ec-124ee744a630 req-6255bcbd-a7f7-4ac7-a456-2f4f14b531fc service nova] Lock "b3003c4b-ae5a-48df-8c12-a915a76253f4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1950.993821] env[62824]: DEBUG nova.compute.manager [req-8d7f5f67-00dd-42c1-99ec-124ee744a630 req-6255bcbd-a7f7-4ac7-a456-2f4f14b531fc service nova] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] No waiting events found dispatching network-vif-plugged-4e0d51df-4419-4059-9dab-be44957e4f77 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1950.993999] env[62824]: WARNING nova.compute.manager [req-8d7f5f67-00dd-42c1-99ec-124ee744a630 req-6255bcbd-a7f7-4ac7-a456-2f4f14b531fc service nova] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Received unexpected event network-vif-plugged-4e0d51df-4419-4059-9dab-be44957e4f77 for instance with vm_state building and task_state spawning. [ 1950.994185] env[62824]: DEBUG nova.compute.manager [req-8d7f5f67-00dd-42c1-99ec-124ee744a630 req-6255bcbd-a7f7-4ac7-a456-2f4f14b531fc service nova] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Received event network-changed-4e0d51df-4419-4059-9dab-be44957e4f77 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1950.994504] env[62824]: DEBUG nova.compute.manager [req-8d7f5f67-00dd-42c1-99ec-124ee744a630 req-6255bcbd-a7f7-4ac7-a456-2f4f14b531fc service nova] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Refreshing instance network info cache due to event network-changed-4e0d51df-4419-4059-9dab-be44957e4f77. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1950.994756] env[62824]: DEBUG oslo_concurrency.lockutils [req-8d7f5f67-00dd-42c1-99ec-124ee744a630 req-6255bcbd-a7f7-4ac7-a456-2f4f14b531fc service nova] Acquiring lock "refresh_cache-b3003c4b-ae5a-48df-8c12-a915a76253f4" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1951.134570] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Releasing lock "refresh_cache-b3003c4b-ae5a-48df-8c12-a915a76253f4" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1951.134570] env[62824]: DEBUG nova.compute.manager [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Instance network_info: |[{"id": "4e0d51df-4419-4059-9dab-be44957e4f77", "address": "fa:16:3e:15:1a:78", "network": {"id": "a410f4f3-d072-4034-b797-8210e1a79093", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1482574391-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e10208040df34eb5adfe8dcbc76043d5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f4399275-8e92-4448-be9e-d4984e93e89c", "external-id": "nsx-vlan-transportzone-192", "segmentation_id": 192, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e0d51df-44", "ovs_interfaceid": "4e0d51df-4419-4059-9dab-be44957e4f77", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1951.134732] env[62824]: DEBUG oslo_concurrency.lockutils [req-8d7f5f67-00dd-42c1-99ec-124ee744a630 req-6255bcbd-a7f7-4ac7-a456-2f4f14b531fc service nova] Acquired lock "refresh_cache-b3003c4b-ae5a-48df-8c12-a915a76253f4" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1951.134732] env[62824]: DEBUG nova.network.neutron [req-8d7f5f67-00dd-42c1-99ec-124ee744a630 req-6255bcbd-a7f7-4ac7-a456-2f4f14b531fc service nova] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Refreshing network info cache for port 4e0d51df-4419-4059-9dab-be44957e4f77 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1951.136136] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:15:1a:78', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f4399275-8e92-4448-be9e-d4984e93e89c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4e0d51df-4419-4059-9dab-be44957e4f77', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1951.145948] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Creating folder: Project (e10208040df34eb5adfe8dcbc76043d5). Parent ref: group-v438503. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1951.149351] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ca548a88-572d-424f-aa34-aaeac16dce01 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1951.163247] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Created folder: Project (e10208040df34eb5adfe8dcbc76043d5) in parent group-v438503. [ 1951.163445] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Creating folder: Instances. Parent ref: group-v438569. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1951.163699] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-87b8bc05-ea6a-4e1e-ae6e-8db78dc6e8cd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1951.175306] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Created folder: Instances in parent group-v438569. [ 1951.175626] env[62824]: DEBUG oslo.service.loopingcall [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1951.175807] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1951.176064] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2ca55873-fafe-4de6-9d8d-64f49fded1d0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1951.201375] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1951.201375] env[62824]: value = "task-2145838" [ 1951.201375] env[62824]: _type = "Task" [ 1951.201375] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1951.212995] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145838, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1951.241881] env[62824]: DEBUG nova.compute.manager [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: 6d80ec06-8559-4964-8577-a2512aa366ed] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1951.243102] env[62824]: INFO nova.compute.resource_tracker [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Updating resource usage from migration bc55ced6-f1ca-4506-869b-2d7d5683b841 [ 1951.276714] env[62824]: DEBUG nova.virt.hardware [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1951.276933] env[62824]: DEBUG nova.virt.hardware [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1951.277105] env[62824]: DEBUG nova.virt.hardware [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1951.277290] env[62824]: DEBUG nova.virt.hardware [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1951.277430] env[62824]: DEBUG nova.virt.hardware [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1951.277573] env[62824]: DEBUG nova.virt.hardware [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1951.277795] env[62824]: DEBUG nova.virt.hardware [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1951.277946] env[62824]: DEBUG nova.virt.hardware [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1951.278124] env[62824]: DEBUG nova.virt.hardware [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1951.278283] env[62824]: DEBUG nova.virt.hardware [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1951.278463] env[62824]: DEBUG nova.virt.hardware [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1951.279406] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64ef53f1-e331-492e-a71f-d9e0de15de33 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1951.303142] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c5f2a3a-e560-47d4-8daa-16aed521c67a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1951.421692] env[62824]: DEBUG nova.network.neutron [req-8d7f5f67-00dd-42c1-99ec-124ee744a630 req-6255bcbd-a7f7-4ac7-a456-2f4f14b531fc service nova] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Updated VIF entry in instance network info cache for port 4e0d51df-4419-4059-9dab-be44957e4f77. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1951.422141] env[62824]: DEBUG nova.network.neutron [req-8d7f5f67-00dd-42c1-99ec-124ee744a630 req-6255bcbd-a7f7-4ac7-a456-2f4f14b531fc service nova] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Updating instance_info_cache with network_info: [{"id": "4e0d51df-4419-4059-9dab-be44957e4f77", "address": "fa:16:3e:15:1a:78", "network": {"id": "a410f4f3-d072-4034-b797-8210e1a79093", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1482574391-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e10208040df34eb5adfe8dcbc76043d5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f4399275-8e92-4448-be9e-d4984e93e89c", "external-id": "nsx-vlan-transportzone-192", "segmentation_id": 192, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e0d51df-44", "ovs_interfaceid": "4e0d51df-4419-4059-9dab-be44957e4f77", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1951.434296] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Acquiring lock "c63b05b7-1dbf-4637-8728-33673be0d3ea" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1951.434522] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Lock "c63b05b7-1dbf-4637-8728-33673be0d3ea" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1951.439580] env[62824]: DEBUG oslo_vmware.api [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Task: {'id': task-2145835, 'name': RelocateVM_Task} progress is 97%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1951.716028] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145838, 'name': CreateVM_Task, 'duration_secs': 0.40475} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1951.716028] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1951.716028] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1951.716028] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1951.716028] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1951.716332] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-68acc582-b33f-4458-be69-4a1587bdcc9d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1951.723445] env[62824]: DEBUG oslo_vmware.api [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for the task: (returnval){ [ 1951.723445] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52052667-7636-35d4-bad7-fa238ea23cab" [ 1951.723445] env[62824]: _type = "Task" [ 1951.723445] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1951.736479] env[62824]: DEBUG oslo_vmware.api [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52052667-7636-35d4-bad7-fa238ea23cab, 'name': SearchDatastore_Task, 'duration_secs': 0.008532} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1951.739208] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1951.739478] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1951.739671] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1951.739813] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1951.739988] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1951.740451] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-44ef219b-d0b8-4e22-a3ea-2b6a91ad0aba {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1951.747956] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1951.748300] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1951.749054] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b72604cd-a813-4607-b781-d5e6a9f14505 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1951.754189] env[62824]: DEBUG oslo_vmware.api [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for the task: (returnval){ [ 1951.754189] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5232d22f-5991-9ce9-53a7-bc7bb0b1a523" [ 1951.754189] env[62824]: _type = "Task" [ 1951.754189] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1951.766514] env[62824]: DEBUG oslo_vmware.api [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5232d22f-5991-9ce9-53a7-bc7bb0b1a523, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1951.773053] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager.update_available_resource {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1951.840831] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60990cac-752b-4cee-bef2-c893d19f73c0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1951.849419] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f73a2103-961c-4e8d-bb19-9acef85874ed {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1951.882016] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d34e8518-21a4-4c6d-90d3-e6d9f032f12e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1951.890225] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-834ccad8-83c2-4aa9-81f2-9631929fe8a1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1951.904270] env[62824]: DEBUG nova.compute.provider_tree [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1951.932197] env[62824]: DEBUG oslo_concurrency.lockutils [req-8d7f5f67-00dd-42c1-99ec-124ee744a630 req-6255bcbd-a7f7-4ac7-a456-2f4f14b531fc service nova] Releasing lock "refresh_cache-b3003c4b-ae5a-48df-8c12-a915a76253f4" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1951.936174] env[62824]: DEBUG oslo_vmware.api [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Task: {'id': task-2145835, 'name': RelocateVM_Task} progress is 98%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1951.962690] env[62824]: DEBUG nova.network.neutron [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: 6d80ec06-8559-4964-8577-a2512aa366ed] Successfully updated port: 24e47f2d-ba7e-48db-9b98-c4f3f6a78930 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1952.267235] env[62824]: DEBUG oslo_vmware.api [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5232d22f-5991-9ce9-53a7-bc7bb0b1a523, 'name': SearchDatastore_Task, 'duration_secs': 0.020991} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1952.267955] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5cdb6b9c-2cde-4fd7-ae8c-501fb9925188 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1952.273667] env[62824]: DEBUG oslo_vmware.api [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for the task: (returnval){ [ 1952.273667] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52e87c0b-f69b-3693-e400-51dd51a4c1c8" [ 1952.273667] env[62824]: _type = "Task" [ 1952.273667] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1952.279083] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1952.283559] env[62824]: DEBUG oslo_vmware.api [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52e87c0b-f69b-3693-e400-51dd51a4c1c8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1952.408068] env[62824]: DEBUG nova.scheduler.client.report [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1952.436649] env[62824]: DEBUG oslo_vmware.api [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Task: {'id': task-2145835, 'name': RelocateVM_Task, 'duration_secs': 3.045853} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1952.436989] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] [instance: b96d1351-f3a7-4bac-998c-a34ab2606041] Volume attach. Driver type: vmdk {{(pid=62824) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1952.437172] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] [instance: b96d1351-f3a7-4bac-998c-a34ab2606041] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-438510', 'volume_id': '64941ab3-102e-48a5-a3ad-433fb818850b', 'name': 'volume-64941ab3-102e-48a5-a3ad-433fb818850b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'b96d1351-f3a7-4bac-998c-a34ab2606041', 'attached_at': '', 'detached_at': '', 'volume_id': '64941ab3-102e-48a5-a3ad-433fb818850b', 'serial': '64941ab3-102e-48a5-a3ad-433fb818850b'} {{(pid=62824) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1952.437918] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbf71d4e-33d9-4504-8c2f-7fb58c7ef4bb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1952.453324] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-089aa85c-2f56-4cfc-a015-6841142b4567 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1952.466559] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Acquiring lock "refresh_cache-6d80ec06-8559-4964-8577-a2512aa366ed" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1952.466701] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Acquired lock "refresh_cache-6d80ec06-8559-4964-8577-a2512aa366ed" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1952.466846] env[62824]: DEBUG nova.network.neutron [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: 6d80ec06-8559-4964-8577-a2512aa366ed] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1952.476170] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] [instance: b96d1351-f3a7-4bac-998c-a34ab2606041] Reconfiguring VM instance instance-0000001f to attach disk [datastore2] volume-64941ab3-102e-48a5-a3ad-433fb818850b/volume-64941ab3-102e-48a5-a3ad-433fb818850b.vmdk or device None with type thin {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1952.479026] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4dfb3f6a-7b0f-4ffc-961e-c22d61d6cdd0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1952.498746] env[62824]: DEBUG oslo_vmware.api [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Waiting for the task: (returnval){ [ 1952.498746] env[62824]: value = "task-2145839" [ 1952.498746] env[62824]: _type = "Task" [ 1952.498746] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1952.507180] env[62824]: DEBUG oslo_vmware.api [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Task: {'id': task-2145839, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1952.523646] env[62824]: DEBUG nova.network.neutron [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: 6d80ec06-8559-4964-8577-a2512aa366ed] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1952.668052] env[62824]: DEBUG nova.network.neutron [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: 6d80ec06-8559-4964-8577-a2512aa366ed] Updating instance_info_cache with network_info: [{"id": "24e47f2d-ba7e-48db-9b98-c4f3f6a78930", "address": "fa:16:3e:62:14:0e", "network": {"id": "a410f4f3-d072-4034-b797-8210e1a79093", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1482574391-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e10208040df34eb5adfe8dcbc76043d5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f4399275-8e92-4448-be9e-d4984e93e89c", "external-id": "nsx-vlan-transportzone-192", "segmentation_id": 192, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap24e47f2d-ba", "ovs_interfaceid": "24e47f2d-ba7e-48db-9b98-c4f3f6a78930", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1952.784256] env[62824]: DEBUG oslo_vmware.api [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52e87c0b-f69b-3693-e400-51dd51a4c1c8, 'name': SearchDatastore_Task, 'duration_secs': 0.028201} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1952.784497] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1952.784777] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] b3003c4b-ae5a-48df-8c12-a915a76253f4/b3003c4b-ae5a-48df-8c12-a915a76253f4.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1952.785048] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8b13384b-3ece-48c1-bb6d-a1bfc25b7e5c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1952.791035] env[62824]: DEBUG oslo_vmware.api [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for the task: (returnval){ [ 1952.791035] env[62824]: value = "task-2145840" [ 1952.791035] env[62824]: _type = "Task" [ 1952.791035] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1952.799069] env[62824]: DEBUG oslo_vmware.api [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2145840, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1952.915301] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.689s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1952.915419] env[62824]: INFO nova.compute.manager [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Migrating [ 1952.915758] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Acquiring lock "compute-rpcapi-router" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1952.915846] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Acquired lock "compute-rpcapi-router" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1952.918063] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0b8e688c-e818-48ea-b84f-ab6b2e8d5cfe tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 35.513s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1952.918292] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0b8e688c-e818-48ea-b84f-ab6b2e8d5cfe tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1952.921047] env[62824]: DEBUG oslo_concurrency.lockutils [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.917s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1952.923078] env[62824]: INFO nova.compute.claims [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1952.950496] env[62824]: INFO nova.scheduler.client.report [None req-0b8e688c-e818-48ea-b84f-ab6b2e8d5cfe tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Deleted allocations for instance 4dfe9f82-7f89-486b-b159-948656d2896c [ 1953.008829] env[62824]: DEBUG oslo_vmware.api [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Task: {'id': task-2145839, 'name': ReconfigVM_Task, 'duration_secs': 0.257429} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1953.009216] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] [instance: b96d1351-f3a7-4bac-998c-a34ab2606041] Reconfigured VM instance instance-0000001f to attach disk [datastore2] volume-64941ab3-102e-48a5-a3ad-433fb818850b/volume-64941ab3-102e-48a5-a3ad-433fb818850b.vmdk or device None with type thin {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1953.014164] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ee227233-8d33-42eb-bee4-929d8baaee3c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1953.026344] env[62824]: DEBUG nova.compute.manager [req-a01741f0-08fe-4200-9ccc-15f040000193 req-e1b9b2e3-887f-48d9-a2a3-d9156b96fc56 service nova] [instance: 6d80ec06-8559-4964-8577-a2512aa366ed] Received event network-vif-plugged-24e47f2d-ba7e-48db-9b98-c4f3f6a78930 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1953.026669] env[62824]: DEBUG oslo_concurrency.lockutils [req-a01741f0-08fe-4200-9ccc-15f040000193 req-e1b9b2e3-887f-48d9-a2a3-d9156b96fc56 service nova] Acquiring lock "6d80ec06-8559-4964-8577-a2512aa366ed-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1953.026771] env[62824]: DEBUG oslo_concurrency.lockutils [req-a01741f0-08fe-4200-9ccc-15f040000193 req-e1b9b2e3-887f-48d9-a2a3-d9156b96fc56 service nova] Lock "6d80ec06-8559-4964-8577-a2512aa366ed-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1953.026939] env[62824]: DEBUG oslo_concurrency.lockutils [req-a01741f0-08fe-4200-9ccc-15f040000193 req-e1b9b2e3-887f-48d9-a2a3-d9156b96fc56 service nova] Lock "6d80ec06-8559-4964-8577-a2512aa366ed-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1953.027167] env[62824]: DEBUG nova.compute.manager [req-a01741f0-08fe-4200-9ccc-15f040000193 req-e1b9b2e3-887f-48d9-a2a3-d9156b96fc56 service nova] [instance: 6d80ec06-8559-4964-8577-a2512aa366ed] No waiting events found dispatching network-vif-plugged-24e47f2d-ba7e-48db-9b98-c4f3f6a78930 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1953.027340] env[62824]: WARNING nova.compute.manager [req-a01741f0-08fe-4200-9ccc-15f040000193 req-e1b9b2e3-887f-48d9-a2a3-d9156b96fc56 service nova] [instance: 6d80ec06-8559-4964-8577-a2512aa366ed] Received unexpected event network-vif-plugged-24e47f2d-ba7e-48db-9b98-c4f3f6a78930 for instance with vm_state building and task_state spawning. [ 1953.027500] env[62824]: DEBUG nova.compute.manager [req-a01741f0-08fe-4200-9ccc-15f040000193 req-e1b9b2e3-887f-48d9-a2a3-d9156b96fc56 service nova] [instance: 6d80ec06-8559-4964-8577-a2512aa366ed] Received event network-changed-24e47f2d-ba7e-48db-9b98-c4f3f6a78930 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1953.027653] env[62824]: DEBUG nova.compute.manager [req-a01741f0-08fe-4200-9ccc-15f040000193 req-e1b9b2e3-887f-48d9-a2a3-d9156b96fc56 service nova] [instance: 6d80ec06-8559-4964-8577-a2512aa366ed] Refreshing instance network info cache due to event network-changed-24e47f2d-ba7e-48db-9b98-c4f3f6a78930. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1953.027826] env[62824]: DEBUG oslo_concurrency.lockutils [req-a01741f0-08fe-4200-9ccc-15f040000193 req-e1b9b2e3-887f-48d9-a2a3-d9156b96fc56 service nova] Acquiring lock "refresh_cache-6d80ec06-8559-4964-8577-a2512aa366ed" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1953.035191] env[62824]: DEBUG oslo_vmware.api [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Waiting for the task: (returnval){ [ 1953.035191] env[62824]: value = "task-2145841" [ 1953.035191] env[62824]: _type = "Task" [ 1953.035191] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1953.044164] env[62824]: DEBUG oslo_vmware.api [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Task: {'id': task-2145841, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1953.171976] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Releasing lock "refresh_cache-6d80ec06-8559-4964-8577-a2512aa366ed" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1953.171976] env[62824]: DEBUG nova.compute.manager [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: 6d80ec06-8559-4964-8577-a2512aa366ed] Instance network_info: |[{"id": "24e47f2d-ba7e-48db-9b98-c4f3f6a78930", "address": "fa:16:3e:62:14:0e", "network": {"id": "a410f4f3-d072-4034-b797-8210e1a79093", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1482574391-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e10208040df34eb5adfe8dcbc76043d5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f4399275-8e92-4448-be9e-d4984e93e89c", "external-id": "nsx-vlan-transportzone-192", "segmentation_id": 192, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap24e47f2d-ba", "ovs_interfaceid": "24e47f2d-ba7e-48db-9b98-c4f3f6a78930", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1953.172234] env[62824]: DEBUG oslo_concurrency.lockutils [req-a01741f0-08fe-4200-9ccc-15f040000193 req-e1b9b2e3-887f-48d9-a2a3-d9156b96fc56 service nova] Acquired lock "refresh_cache-6d80ec06-8559-4964-8577-a2512aa366ed" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1953.172302] env[62824]: DEBUG nova.network.neutron [req-a01741f0-08fe-4200-9ccc-15f040000193 req-e1b9b2e3-887f-48d9-a2a3-d9156b96fc56 service nova] [instance: 6d80ec06-8559-4964-8577-a2512aa366ed] Refreshing network info cache for port 24e47f2d-ba7e-48db-9b98-c4f3f6a78930 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1953.173815] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: 6d80ec06-8559-4964-8577-a2512aa366ed] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:62:14:0e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f4399275-8e92-4448-be9e-d4984e93e89c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '24e47f2d-ba7e-48db-9b98-c4f3f6a78930', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1953.182818] env[62824]: DEBUG oslo.service.loopingcall [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1953.183457] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6d80ec06-8559-4964-8577-a2512aa366ed] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1953.183744] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-45e09fb3-5246-46ad-b063-fcce050dcc6e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1953.207070] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1953.207070] env[62824]: value = "task-2145842" [ 1953.207070] env[62824]: _type = "Task" [ 1953.207070] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1953.217922] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145842, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1953.301149] env[62824]: DEBUG oslo_vmware.api [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2145840, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.419011} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1953.301511] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] b3003c4b-ae5a-48df-8c12-a915a76253f4/b3003c4b-ae5a-48df-8c12-a915a76253f4.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1953.301589] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1953.301787] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-11a1c410-e48b-4ff6-adc0-b8ca571a1a16 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1953.308165] env[62824]: DEBUG oslo_vmware.api [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for the task: (returnval){ [ 1953.308165] env[62824]: value = "task-2145843" [ 1953.308165] env[62824]: _type = "Task" [ 1953.308165] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1953.316048] env[62824]: DEBUG oslo_vmware.api [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2145843, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1953.427589] env[62824]: INFO nova.compute.rpcapi [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Automatically selected compute RPC version 6.4 from minimum service version 68 [ 1953.428073] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Releasing lock "compute-rpcapi-router" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1953.459183] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0b8e688c-e818-48ea-b84f-ab6b2e8d5cfe tempest-ServerMetadataNegativeTestJSON-671663195 tempest-ServerMetadataNegativeTestJSON-671663195-project-member] Lock "4dfe9f82-7f89-486b-b159-948656d2896c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 39.584s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1953.545035] env[62824]: DEBUG oslo_vmware.api [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Task: {'id': task-2145841, 'name': ReconfigVM_Task, 'duration_secs': 0.224697} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1953.545426] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] [instance: b96d1351-f3a7-4bac-998c-a34ab2606041] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-438510', 'volume_id': '64941ab3-102e-48a5-a3ad-433fb818850b', 'name': 'volume-64941ab3-102e-48a5-a3ad-433fb818850b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'b96d1351-f3a7-4bac-998c-a34ab2606041', 'attached_at': '', 'detached_at': '', 'volume_id': '64941ab3-102e-48a5-a3ad-433fb818850b', 'serial': '64941ab3-102e-48a5-a3ad-433fb818850b'} {{(pid=62824) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1953.546084] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a8b5542d-e4a7-4da4-9264-440023b8ca23 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1953.552591] env[62824]: DEBUG oslo_vmware.api [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Waiting for the task: (returnval){ [ 1953.552591] env[62824]: value = "task-2145844" [ 1953.552591] env[62824]: _type = "Task" [ 1953.552591] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1953.560844] env[62824]: DEBUG oslo_vmware.api [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Task: {'id': task-2145844, 'name': Rename_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1953.718727] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145842, 'name': CreateVM_Task, 'duration_secs': 0.31182} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1953.718905] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6d80ec06-8559-4964-8577-a2512aa366ed] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1953.719683] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1953.719857] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1953.720414] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1953.720664] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cc39d076-d91f-4d00-b3c7-a2273f9e9761 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1953.724964] env[62824]: DEBUG oslo_vmware.api [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for the task: (returnval){ [ 1953.724964] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52d68858-0400-f624-2eef-4121e8765e3e" [ 1953.724964] env[62824]: _type = "Task" [ 1953.724964] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1953.732659] env[62824]: DEBUG oslo_vmware.api [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52d68858-0400-f624-2eef-4121e8765e3e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1953.821946] env[62824]: DEBUG oslo_vmware.api [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2145843, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061691} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1953.822232] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1953.823084] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ca63feb-d31c-4e58-85dc-2496d26d0f3b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1953.845857] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Reconfiguring VM instance instance-00000020 to attach disk [datastore2] b3003c4b-ae5a-48df-8c12-a915a76253f4/b3003c4b-ae5a-48df-8c12-a915a76253f4.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1953.846155] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f485d026-be39-4872-87b8-9baca5f9e34e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1953.868237] env[62824]: DEBUG oslo_vmware.api [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for the task: (returnval){ [ 1953.868237] env[62824]: value = "task-2145845" [ 1953.868237] env[62824]: _type = "Task" [ 1953.868237] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1953.878667] env[62824]: DEBUG oslo_vmware.api [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2145845, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1953.896545] env[62824]: DEBUG nova.network.neutron [req-a01741f0-08fe-4200-9ccc-15f040000193 req-e1b9b2e3-887f-48d9-a2a3-d9156b96fc56 service nova] [instance: 6d80ec06-8559-4964-8577-a2512aa366ed] Updated VIF entry in instance network info cache for port 24e47f2d-ba7e-48db-9b98-c4f3f6a78930. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1953.896910] env[62824]: DEBUG nova.network.neutron [req-a01741f0-08fe-4200-9ccc-15f040000193 req-e1b9b2e3-887f-48d9-a2a3-d9156b96fc56 service nova] [instance: 6d80ec06-8559-4964-8577-a2512aa366ed] Updating instance_info_cache with network_info: [{"id": "24e47f2d-ba7e-48db-9b98-c4f3f6a78930", "address": "fa:16:3e:62:14:0e", "network": {"id": "a410f4f3-d072-4034-b797-8210e1a79093", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1482574391-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e10208040df34eb5adfe8dcbc76043d5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f4399275-8e92-4448-be9e-d4984e93e89c", "external-id": "nsx-vlan-transportzone-192", "segmentation_id": 192, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap24e47f2d-ba", "ovs_interfaceid": "24e47f2d-ba7e-48db-9b98-c4f3f6a78930", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1953.944327] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Acquiring lock "refresh_cache-e158b5d4-c120-4e6c-89c4-7668e097926d" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1953.944515] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Acquired lock "refresh_cache-e158b5d4-c120-4e6c-89c4-7668e097926d" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1953.944724] env[62824]: DEBUG nova.network.neutron [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1954.062617] env[62824]: DEBUG oslo_vmware.api [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Task: {'id': task-2145844, 'name': Rename_Task, 'duration_secs': 0.17181} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1954.065252] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] [instance: b96d1351-f3a7-4bac-998c-a34ab2606041] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1954.065780] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c614e6b5-87fb-476b-bec1-7f971992a607 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1954.072602] env[62824]: DEBUG oslo_vmware.api [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Waiting for the task: (returnval){ [ 1954.072602] env[62824]: value = "task-2145846" [ 1954.072602] env[62824]: _type = "Task" [ 1954.072602] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1954.084424] env[62824]: DEBUG oslo_vmware.api [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Task: {'id': task-2145846, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1954.236078] env[62824]: DEBUG oslo_vmware.api [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52d68858-0400-f624-2eef-4121e8765e3e, 'name': SearchDatastore_Task, 'duration_secs': 0.010157} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1954.236939] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1954.236939] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: 6d80ec06-8559-4964-8577-a2512aa366ed] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1954.237340] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1954.237517] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1954.237701] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1954.238926] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7526d7c7-fb87-4e9f-bc0d-643cb4f21749 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1954.253908] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1954.254128] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1954.257769] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-732fae67-e825-48e1-8001-90988e4e4cb5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1954.262795] env[62824]: DEBUG oslo_vmware.api [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for the task: (returnval){ [ 1954.262795] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52f39582-3fb6-c7ec-da04-f9b73c52b494" [ 1954.262795] env[62824]: _type = "Task" [ 1954.262795] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1954.271671] env[62824]: DEBUG oslo_vmware.api [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52f39582-3fb6-c7ec-da04-f9b73c52b494, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1954.381128] env[62824]: DEBUG oslo_vmware.api [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2145845, 'name': ReconfigVM_Task, 'duration_secs': 0.279366} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1954.381432] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Reconfigured VM instance instance-00000020 to attach disk [datastore2] b3003c4b-ae5a-48df-8c12-a915a76253f4/b3003c4b-ae5a-48df-8c12-a915a76253f4.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1954.382152] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4fbec7c3-c1f8-44c9-9fb0-5010a95df7fb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1954.390639] env[62824]: DEBUG oslo_vmware.api [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for the task: (returnval){ [ 1954.390639] env[62824]: value = "task-2145847" [ 1954.390639] env[62824]: _type = "Task" [ 1954.390639] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1954.402032] env[62824]: DEBUG oslo_vmware.api [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2145847, 'name': Rename_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1954.402032] env[62824]: DEBUG oslo_concurrency.lockutils [req-a01741f0-08fe-4200-9ccc-15f040000193 req-e1b9b2e3-887f-48d9-a2a3-d9156b96fc56 service nova] Releasing lock "refresh_cache-6d80ec06-8559-4964-8577-a2512aa366ed" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1954.454128] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e70e0b8-c487-42e4-8597-911ae3ad442f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1954.464178] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-100441e1-85c7-405a-9f0d-08fca81a708f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1954.500331] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdc96004-4a27-4c3e-b314-a79106f6087a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1954.508494] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9225ac64-4027-4f0b-af95-a46a482d2b30 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1954.524551] env[62824]: DEBUG nova.compute.provider_tree [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Updating inventory in ProviderTree for provider bf95157b-a274-42de-9ccf-9851128a44a1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1954.585526] env[62824]: DEBUG oslo_vmware.api [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Task: {'id': task-2145846, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1954.678216] env[62824]: DEBUG nova.network.neutron [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Updating instance_info_cache with network_info: [{"id": "02e90c28-7f36-4d83-84bb-fec194a04dab", "address": "fa:16:3e:53:1a:4c", "network": {"id": "b6403dc3-30ef-4c76-b521-380c179f345d", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.19", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "22c5c40c2e004babaa15896c89307303", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1002b79b-224e-41e3-a484-4245a767147a", "external-id": "nsx-vlan-transportzone-353", "segmentation_id": 353, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap02e90c28-7f", "ovs_interfaceid": "02e90c28-7f36-4d83-84bb-fec194a04dab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1954.773888] env[62824]: DEBUG oslo_vmware.api [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52f39582-3fb6-c7ec-da04-f9b73c52b494, 'name': SearchDatastore_Task, 'duration_secs': 0.034377} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1954.774746] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d9ee1873-28bc-407d-9c02-17318fb31a5c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1954.780170] env[62824]: DEBUG oslo_vmware.api [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for the task: (returnval){ [ 1954.780170] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52f1b474-0f8d-c654-27ee-1ac722a8f763" [ 1954.780170] env[62824]: _type = "Task" [ 1954.780170] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1954.788075] env[62824]: DEBUG oslo_vmware.api [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52f1b474-0f8d-c654-27ee-1ac722a8f763, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1954.900255] env[62824]: DEBUG oslo_vmware.api [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2145847, 'name': Rename_Task, 'duration_secs': 0.179712} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1954.900620] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1954.900918] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-211c47ab-b6fc-45c9-a1f9-6c84ac8d0b1e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1954.907763] env[62824]: DEBUG oslo_vmware.api [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for the task: (returnval){ [ 1954.907763] env[62824]: value = "task-2145848" [ 1954.907763] env[62824]: _type = "Task" [ 1954.907763] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1954.917876] env[62824]: DEBUG oslo_vmware.api [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2145848, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1955.049625] env[62824]: ERROR nova.scheduler.client.report [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [req-209e7ff7-b582-4147-b89c-4a68f0e40b5a] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID bf95157b-a274-42de-9ccf-9851128a44a1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-209e7ff7-b582-4147-b89c-4a68f0e40b5a"}]} [ 1955.085157] env[62824]: DEBUG oslo_vmware.api [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Task: {'id': task-2145846, 'name': PowerOnVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1955.086538] env[62824]: DEBUG nova.scheduler.client.report [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Refreshing inventories for resource provider bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1955.103738] env[62824]: DEBUG nova.scheduler.client.report [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Updating ProviderTree inventory for provider bf95157b-a274-42de-9ccf-9851128a44a1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1955.103981] env[62824]: DEBUG nova.compute.provider_tree [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Updating inventory in ProviderTree for provider bf95157b-a274-42de-9ccf-9851128a44a1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1955.117044] env[62824]: DEBUG nova.scheduler.client.report [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Refreshing aggregate associations for resource provider bf95157b-a274-42de-9ccf-9851128a44a1, aggregates: None {{(pid=62824) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1955.138288] env[62824]: DEBUG nova.scheduler.client.report [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Refreshing trait associations for resource provider bf95157b-a274-42de-9ccf-9851128a44a1, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_ISO {{(pid=62824) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1955.183953] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Releasing lock "refresh_cache-e158b5d4-c120-4e6c-89c4-7668e097926d" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1955.290102] env[62824]: DEBUG oslo_vmware.api [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52f1b474-0f8d-c654-27ee-1ac722a8f763, 'name': SearchDatastore_Task, 'duration_secs': 0.009048} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1955.292692] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1955.293075] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 6d80ec06-8559-4964-8577-a2512aa366ed/6d80ec06-8559-4964-8577-a2512aa366ed.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1955.294609] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1528b03f-8621-4c6a-bded-e23d62a8301f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1955.302746] env[62824]: DEBUG oslo_vmware.api [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for the task: (returnval){ [ 1955.302746] env[62824]: value = "task-2145849" [ 1955.302746] env[62824]: _type = "Task" [ 1955.302746] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1955.312689] env[62824]: DEBUG oslo_vmware.api [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2145849, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1955.419405] env[62824]: DEBUG oslo_vmware.api [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2145848, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1955.584514] env[62824]: DEBUG oslo_vmware.api [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Task: {'id': task-2145846, 'name': PowerOnVM_Task, 'duration_secs': 1.080866} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1955.587186] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] [instance: b96d1351-f3a7-4bac-998c-a34ab2606041] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1955.587458] env[62824]: INFO nova.compute.manager [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] [instance: b96d1351-f3a7-4bac-998c-a34ab2606041] Took 7.89 seconds to spawn the instance on the hypervisor. [ 1955.587749] env[62824]: DEBUG nova.compute.manager [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] [instance: b96d1351-f3a7-4bac-998c-a34ab2606041] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1955.588912] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1012d8d-24af-4553-a814-7fabf2c0a9e3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1955.726886] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ed1d5c3-6ae2-46ec-9298-43bfa473e407 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1955.738111] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b68cbc0a-5184-4732-92a4-430e41e2c43d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1955.781938] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78d9b502-50eb-451c-9dec-953ef986bb2c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1955.791933] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-048b7a25-6e39-4432-a6c0-c51c4e473f45 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1955.808468] env[62824]: DEBUG nova.compute.provider_tree [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Updating inventory in ProviderTree for provider bf95157b-a274-42de-9ccf-9851128a44a1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1955.819257] env[62824]: DEBUG oslo_vmware.api [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2145849, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.496498} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1955.820131] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 6d80ec06-8559-4964-8577-a2512aa366ed/6d80ec06-8559-4964-8577-a2512aa366ed.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1955.820360] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: 6d80ec06-8559-4964-8577-a2512aa366ed] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1955.820609] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3e21b4c4-f378-4887-ad9e-c48b65846891 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1955.829441] env[62824]: DEBUG oslo_vmware.api [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for the task: (returnval){ [ 1955.829441] env[62824]: value = "task-2145850" [ 1955.829441] env[62824]: _type = "Task" [ 1955.829441] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1955.840655] env[62824]: DEBUG oslo_vmware.api [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2145850, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1955.918095] env[62824]: DEBUG oslo_vmware.api [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2145848, 'name': PowerOnVM_Task, 'duration_secs': 0.884564} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1955.918374] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1955.918586] env[62824]: INFO nova.compute.manager [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Took 7.33 seconds to spawn the instance on the hypervisor. [ 1955.918762] env[62824]: DEBUG nova.compute.manager [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1955.919549] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-716e8b33-9ff7-45ac-95ac-36e28d3930a6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1956.110125] env[62824]: INFO nova.compute.manager [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] [instance: b96d1351-f3a7-4bac-998c-a34ab2606041] Took 45.18 seconds to build instance. [ 1956.340487] env[62824]: DEBUG oslo_vmware.api [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2145850, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060976} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1956.340808] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: 6d80ec06-8559-4964-8577-a2512aa366ed] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1956.341606] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-300d07d8-4626-4b85-94a6-28420bc6c965 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1956.346396] env[62824]: DEBUG nova.scheduler.client.report [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Updated inventory for provider bf95157b-a274-42de-9ccf-9851128a44a1 with generation 64 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 1956.346656] env[62824]: DEBUG nova.compute.provider_tree [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Updating resource provider bf95157b-a274-42de-9ccf-9851128a44a1 generation from 64 to 65 during operation: update_inventory {{(pid=62824) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1956.346845] env[62824]: DEBUG nova.compute.provider_tree [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Updating inventory in ProviderTree for provider bf95157b-a274-42de-9ccf-9851128a44a1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1956.368906] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: 6d80ec06-8559-4964-8577-a2512aa366ed] Reconfiguring VM instance instance-00000021 to attach disk [datastore2] 6d80ec06-8559-4964-8577-a2512aa366ed/6d80ec06-8559-4964-8577-a2512aa366ed.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1956.370235] env[62824]: DEBUG oslo_concurrency.lockutils [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.449s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1956.370787] env[62824]: DEBUG nova.compute.manager [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1956.373686] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d1b8161c-b3fe-4643-afdd-544aaba5e63d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1956.388125] env[62824]: DEBUG oslo_concurrency.lockutils [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.015s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1956.389637] env[62824]: INFO nova.compute.claims [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 487c2c9d-2cd2-4912-8613-e1bfac732c40] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1956.398744] env[62824]: DEBUG oslo_vmware.api [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for the task: (returnval){ [ 1956.398744] env[62824]: value = "task-2145851" [ 1956.398744] env[62824]: _type = "Task" [ 1956.398744] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1956.407106] env[62824]: DEBUG oslo_vmware.api [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2145851, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1956.439968] env[62824]: INFO nova.compute.manager [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Took 43.27 seconds to build instance. [ 1956.611842] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0351681d-3332-4993-bc3c-9563a93d9351 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Lock "b96d1351-f3a7-4bac-998c-a34ab2606041" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 90.836s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1956.701660] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4141e639-5918-4c6f-9a08-ca742556ad4a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1956.723108] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Updating instance 'e158b5d4-c120-4e6c-89c4-7668e097926d' progress to 0 {{(pid=62824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1956.894772] env[62824]: DEBUG nova.compute.utils [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1956.899154] env[62824]: DEBUG nova.compute.manager [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1956.899154] env[62824]: DEBUG nova.network.neutron [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1956.911384] env[62824]: DEBUG oslo_vmware.api [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2145851, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1956.942275] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c67f0538-fba5-4f95-a2b4-6b39d2d5793b tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Lock "b3003c4b-ae5a-48df-8c12-a915a76253f4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 89.555s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1956.968314] env[62824]: DEBUG nova.policy [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2d3d79fc7c0145c0ba97a846553026c3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '102d2f886fd44e1da6780e48e36f1b62', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 1957.116861] env[62824]: DEBUG nova.compute.manager [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1957.150235] env[62824]: DEBUG nova.compute.manager [req-33f77d8d-8104-40cd-828b-3530d75b472c req-9e9c9a77-b996-4f81-80ed-85e2de48c480 service nova] [instance: b96d1351-f3a7-4bac-998c-a34ab2606041] Received event network-changed-c7dfce05-2238-418a-b414-395b1f1fb4f6 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1957.150558] env[62824]: DEBUG nova.compute.manager [req-33f77d8d-8104-40cd-828b-3530d75b472c req-9e9c9a77-b996-4f81-80ed-85e2de48c480 service nova] [instance: b96d1351-f3a7-4bac-998c-a34ab2606041] Refreshing instance network info cache due to event network-changed-c7dfce05-2238-418a-b414-395b1f1fb4f6. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1957.151628] env[62824]: DEBUG oslo_concurrency.lockutils [req-33f77d8d-8104-40cd-828b-3530d75b472c req-9e9c9a77-b996-4f81-80ed-85e2de48c480 service nova] Acquiring lock "refresh_cache-b96d1351-f3a7-4bac-998c-a34ab2606041" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1957.151628] env[62824]: DEBUG oslo_concurrency.lockutils [req-33f77d8d-8104-40cd-828b-3530d75b472c req-9e9c9a77-b996-4f81-80ed-85e2de48c480 service nova] Acquired lock "refresh_cache-b96d1351-f3a7-4bac-998c-a34ab2606041" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1957.151628] env[62824]: DEBUG nova.network.neutron [req-33f77d8d-8104-40cd-828b-3530d75b472c req-9e9c9a77-b996-4f81-80ed-85e2de48c480 service nova] [instance: b96d1351-f3a7-4bac-998c-a34ab2606041] Refreshing network info cache for port c7dfce05-2238-418a-b414-395b1f1fb4f6 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1957.230262] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1957.230262] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1054ef34-8a4b-4cd9-a064-c5b02d4fb71b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1957.240978] env[62824]: DEBUG oslo_vmware.api [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Waiting for the task: (returnval){ [ 1957.240978] env[62824]: value = "task-2145852" [ 1957.240978] env[62824]: _type = "Task" [ 1957.240978] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1957.251582] env[62824]: DEBUG oslo_vmware.api [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Task: {'id': task-2145852, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1957.280032] env[62824]: DEBUG nova.network.neutron [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Successfully created port: 1e74959f-64c2-4c37-85f8-fa8f81a30d60 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1957.397068] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Acquiring lock "6a724332-a165-4a2b-9dd8-85f27e7b7637" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1957.397068] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Lock "6a724332-a165-4a2b-9dd8-85f27e7b7637" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1957.403020] env[62824]: DEBUG nova.compute.manager [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1957.425082] env[62824]: DEBUG oslo_vmware.api [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2145851, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1957.444953] env[62824]: DEBUG nova.compute.manager [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 8bb90a4d-93c9-4f54-b15e-48fe966e6c97] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1957.633027] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1957.703754] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Acquiring lock "21c05725-3dad-4965-98d0-0622ebcc8ece" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1957.703986] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Lock "21c05725-3dad-4965-98d0-0622ebcc8ece" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1957.750117] env[62824]: DEBUG oslo_vmware.api [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Task: {'id': task-2145852, 'name': PowerOffVM_Task, 'duration_secs': 0.362087} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1957.750384] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1957.750567] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Updating instance 'e158b5d4-c120-4e6c-89c4-7668e097926d' progress to 17 {{(pid=62824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1957.914035] env[62824]: DEBUG nova.network.neutron [req-33f77d8d-8104-40cd-828b-3530d75b472c req-9e9c9a77-b996-4f81-80ed-85e2de48c480 service nova] [instance: b96d1351-f3a7-4bac-998c-a34ab2606041] Updated VIF entry in instance network info cache for port c7dfce05-2238-418a-b414-395b1f1fb4f6. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1957.914332] env[62824]: DEBUG nova.network.neutron [req-33f77d8d-8104-40cd-828b-3530d75b472c req-9e9c9a77-b996-4f81-80ed-85e2de48c480 service nova] [instance: b96d1351-f3a7-4bac-998c-a34ab2606041] Updating instance_info_cache with network_info: [{"id": "c7dfce05-2238-418a-b414-395b1f1fb4f6", "address": "fa:16:3e:34:b9:b4", "network": {"id": "0a7184c9-f122-4c98-a7d1-3637da9e3669", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-1696959333-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.142", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4279c92ea7004c74bdbf768b38d9b372", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a2e2e51-010f-4535-ba88-433663275996", "external-id": "nsx-vlan-transportzone-915", "segmentation_id": 915, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc7dfce05-22", "ovs_interfaceid": "c7dfce05-2238-418a-b414-395b1f1fb4f6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1957.924217] env[62824]: DEBUG oslo_vmware.api [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2145851, 'name': ReconfigVM_Task, 'duration_secs': 1.037405} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1957.926933] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: 6d80ec06-8559-4964-8577-a2512aa366ed] Reconfigured VM instance instance-00000021 to attach disk [datastore2] 6d80ec06-8559-4964-8577-a2512aa366ed/6d80ec06-8559-4964-8577-a2512aa366ed.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1957.928471] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1ff24635-6222-4878-b21b-a646ade25045 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1957.935388] env[62824]: DEBUG oslo_vmware.api [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for the task: (returnval){ [ 1957.935388] env[62824]: value = "task-2145853" [ 1957.935388] env[62824]: _type = "Task" [ 1957.935388] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1957.946306] env[62824]: DEBUG oslo_vmware.api [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2145853, 'name': Rename_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1957.968987] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1957.978810] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd16a44f-0d9e-4dfb-bccb-baa443b0453a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1957.986247] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ecae77b-113f-4052-84d4-3416bfccc019 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1958.015893] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0ca3979-590c-485b-8582-31ed0133ffda {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1958.023797] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7029421-b8cc-42b8-9277-207998ca749f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1958.037954] env[62824]: DEBUG nova.compute.provider_tree [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1958.261895] env[62824]: DEBUG nova.virt.hardware [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:24Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1958.262213] env[62824]: DEBUG nova.virt.hardware [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1958.262415] env[62824]: DEBUG nova.virt.hardware [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1958.262744] env[62824]: DEBUG nova.virt.hardware [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1958.262926] env[62824]: DEBUG nova.virt.hardware [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1958.263135] env[62824]: DEBUG nova.virt.hardware [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1958.263390] env[62824]: DEBUG nova.virt.hardware [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1958.263580] env[62824]: DEBUG nova.virt.hardware [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1958.263806] env[62824]: DEBUG nova.virt.hardware [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1958.264023] env[62824]: DEBUG nova.virt.hardware [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1958.264253] env[62824]: DEBUG nova.virt.hardware [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1958.269817] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8211d5af-2427-4cc2-b0a8-e0d71daddc05 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1958.285734] env[62824]: DEBUG oslo_vmware.api [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Waiting for the task: (returnval){ [ 1958.285734] env[62824]: value = "task-2145854" [ 1958.285734] env[62824]: _type = "Task" [ 1958.285734] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1958.294149] env[62824]: DEBUG oslo_vmware.api [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Task: {'id': task-2145854, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1958.417314] env[62824]: DEBUG nova.compute.manager [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1958.419920] env[62824]: DEBUG oslo_concurrency.lockutils [req-33f77d8d-8104-40cd-828b-3530d75b472c req-9e9c9a77-b996-4f81-80ed-85e2de48c480 service nova] Releasing lock "refresh_cache-b96d1351-f3a7-4bac-998c-a34ab2606041" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1958.443744] env[62824]: DEBUG nova.virt.hardware [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1958.443744] env[62824]: DEBUG nova.virt.hardware [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1958.443744] env[62824]: DEBUG nova.virt.hardware [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1958.443976] env[62824]: DEBUG nova.virt.hardware [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1958.443976] env[62824]: DEBUG nova.virt.hardware [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1958.444137] env[62824]: DEBUG nova.virt.hardware [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1958.444344] env[62824]: DEBUG nova.virt.hardware [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1958.444501] env[62824]: DEBUG nova.virt.hardware [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1958.444664] env[62824]: DEBUG nova.virt.hardware [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1958.444885] env[62824]: DEBUG nova.virt.hardware [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1958.445091] env[62824]: DEBUG nova.virt.hardware [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1958.446019] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9abe8c6e-d9cf-4e62-816f-2d955224f9bd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1958.453981] env[62824]: DEBUG oslo_vmware.api [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2145853, 'name': Rename_Task, 'duration_secs': 0.193536} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1958.456229] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: 6d80ec06-8559-4964-8577-a2512aa366ed] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1958.456531] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b03c8f22-505e-4b77-8f9b-afec584977f3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1958.459136] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36d5133b-564d-4495-acd0-b0e9b0ac8168 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1958.475270] env[62824]: DEBUG oslo_vmware.api [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for the task: (returnval){ [ 1958.475270] env[62824]: value = "task-2145855" [ 1958.475270] env[62824]: _type = "Task" [ 1958.475270] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1958.482987] env[62824]: DEBUG oslo_vmware.api [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2145855, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1958.542237] env[62824]: DEBUG nova.scheduler.client.report [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1958.740260] env[62824]: DEBUG nova.compute.manager [req-a5470221-7dae-42ce-8fe8-be6adfc58478 req-e07c750a-2300-463d-bb01-452e6bc8b7da service nova] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Received event network-vif-plugged-1e74959f-64c2-4c37-85f8-fa8f81a30d60 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1958.740482] env[62824]: DEBUG oslo_concurrency.lockutils [req-a5470221-7dae-42ce-8fe8-be6adfc58478 req-e07c750a-2300-463d-bb01-452e6bc8b7da service nova] Acquiring lock "59b5b883-4188-471c-8862-444f3ce08cb0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1958.740660] env[62824]: DEBUG oslo_concurrency.lockutils [req-a5470221-7dae-42ce-8fe8-be6adfc58478 req-e07c750a-2300-463d-bb01-452e6bc8b7da service nova] Lock "59b5b883-4188-471c-8862-444f3ce08cb0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1958.740828] env[62824]: DEBUG oslo_concurrency.lockutils [req-a5470221-7dae-42ce-8fe8-be6adfc58478 req-e07c750a-2300-463d-bb01-452e6bc8b7da service nova] Lock "59b5b883-4188-471c-8862-444f3ce08cb0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1958.740992] env[62824]: DEBUG nova.compute.manager [req-a5470221-7dae-42ce-8fe8-be6adfc58478 req-e07c750a-2300-463d-bb01-452e6bc8b7da service nova] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] No waiting events found dispatching network-vif-plugged-1e74959f-64c2-4c37-85f8-fa8f81a30d60 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1958.741174] env[62824]: WARNING nova.compute.manager [req-a5470221-7dae-42ce-8fe8-be6adfc58478 req-e07c750a-2300-463d-bb01-452e6bc8b7da service nova] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Received unexpected event network-vif-plugged-1e74959f-64c2-4c37-85f8-fa8f81a30d60 for instance with vm_state building and task_state spawning. [ 1958.796153] env[62824]: DEBUG oslo_vmware.api [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Task: {'id': task-2145854, 'name': ReconfigVM_Task, 'duration_secs': 0.265951} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1958.796645] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Updating instance 'e158b5d4-c120-4e6c-89c4-7668e097926d' progress to 33 {{(pid=62824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1958.832386] env[62824]: DEBUG nova.network.neutron [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Successfully updated port: 1e74959f-64c2-4c37-85f8-fa8f81a30d60 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1958.987943] env[62824]: DEBUG oslo_vmware.api [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2145855, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1959.047903] env[62824]: DEBUG oslo_concurrency.lockutils [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.660s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1959.048456] env[62824]: DEBUG nova.compute.manager [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 487c2c9d-2cd2-4912-8613-e1bfac732c40] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1959.051475] env[62824]: DEBUG oslo_concurrency.lockutils [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 33.397s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1959.051661] env[62824]: DEBUG nova.objects.instance [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62824) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1959.302656] env[62824]: DEBUG nova.virt.hardware [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:50:59Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='1bdf5fc7-1e8e-4668-88fe-6c47a0097f11',id=29,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-1536379991',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1959.302977] env[62824]: DEBUG nova.virt.hardware [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1959.303153] env[62824]: DEBUG nova.virt.hardware [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1959.303332] env[62824]: DEBUG nova.virt.hardware [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1959.303476] env[62824]: DEBUG nova.virt.hardware [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1959.303620] env[62824]: DEBUG nova.virt.hardware [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1959.303860] env[62824]: DEBUG nova.virt.hardware [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1959.303980] env[62824]: DEBUG nova.virt.hardware [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1959.304161] env[62824]: DEBUG nova.virt.hardware [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1959.304325] env[62824]: DEBUG nova.virt.hardware [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1959.304495] env[62824]: DEBUG nova.virt.hardware [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1959.310054] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Reconfiguring VM instance instance-00000017 to detach disk 2000 {{(pid=62824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1959.310352] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6feede65-67df-46e0-8410-6d52e846f335 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1959.328926] env[62824]: DEBUG oslo_vmware.api [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Waiting for the task: (returnval){ [ 1959.328926] env[62824]: value = "task-2145856" [ 1959.328926] env[62824]: _type = "Task" [ 1959.328926] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1959.337444] env[62824]: DEBUG oslo_concurrency.lockutils [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Acquiring lock "refresh_cache-59b5b883-4188-471c-8862-444f3ce08cb0" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1959.337585] env[62824]: DEBUG oslo_concurrency.lockutils [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Acquired lock "refresh_cache-59b5b883-4188-471c-8862-444f3ce08cb0" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1959.337731] env[62824]: DEBUG nova.network.neutron [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1959.338838] env[62824]: DEBUG oslo_vmware.api [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Task: {'id': task-2145856, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1959.486366] env[62824]: DEBUG oslo_vmware.api [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2145855, 'name': PowerOnVM_Task} progress is 78%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1959.561684] env[62824]: DEBUG nova.compute.utils [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1959.563332] env[62824]: DEBUG nova.compute.manager [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 487c2c9d-2cd2-4912-8613-e1bfac732c40] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1959.563556] env[62824]: DEBUG nova.network.neutron [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 487c2c9d-2cd2-4912-8613-e1bfac732c40] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1959.605937] env[62824]: DEBUG nova.policy [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c379c824a178444e99bb121716fc7887', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0c11aef7fb6247cb9b5272a6a063cd12', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 1959.844367] env[62824]: DEBUG oslo_vmware.api [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Task: {'id': task-2145856, 'name': ReconfigVM_Task, 'duration_secs': 0.265898} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1959.844666] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Reconfigured VM instance instance-00000017 to detach disk 2000 {{(pid=62824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1959.846297] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d510a8db-92c7-486e-92ed-bee3aa3f84fc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1959.875866] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Reconfiguring VM instance instance-00000017 to attach disk [datastore2] e158b5d4-c120-4e6c-89c4-7668e097926d/e158b5d4-c120-4e6c-89c4-7668e097926d.vmdk or device None with type thin {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1959.875866] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-845b5831-f3d1-4901-8736-3b75c4a74468 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1959.891785] env[62824]: DEBUG nova.network.neutron [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1959.897155] env[62824]: DEBUG nova.network.neutron [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 487c2c9d-2cd2-4912-8613-e1bfac732c40] Successfully created port: bbbe4627-156d-462e-a430-0e4a08a4c2dc {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1959.900474] env[62824]: DEBUG oslo_vmware.api [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Waiting for the task: (returnval){ [ 1959.900474] env[62824]: value = "task-2145857" [ 1959.900474] env[62824]: _type = "Task" [ 1959.900474] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1959.908577] env[62824]: DEBUG oslo_vmware.api [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Task: {'id': task-2145857, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1959.986728] env[62824]: DEBUG oslo_vmware.api [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2145855, 'name': PowerOnVM_Task, 'duration_secs': 1.215496} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1959.987015] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: 6d80ec06-8559-4964-8577-a2512aa366ed] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1959.987237] env[62824]: INFO nova.compute.manager [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: 6d80ec06-8559-4964-8577-a2512aa366ed] Took 8.75 seconds to spawn the instance on the hypervisor. [ 1959.987424] env[62824]: DEBUG nova.compute.manager [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: 6d80ec06-8559-4964-8577-a2512aa366ed] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1959.988215] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c47de15e-1525-4905-94f7-c2b2dc1158f0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1960.065295] env[62824]: DEBUG oslo_concurrency.lockutils [None req-477349c3-97cd-4958-a4fb-b66e89d5cfaa tempest-ServersAdmin275Test-244782444 tempest-ServersAdmin275Test-244782444-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.014s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1960.066456] env[62824]: DEBUG oslo_concurrency.lockutils [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.224s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1960.072169] env[62824]: INFO nova.compute.claims [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] [instance: 69bcac33-d6ac-4ee7-b674-58ab25389149] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1960.074962] env[62824]: DEBUG nova.compute.manager [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 487c2c9d-2cd2-4912-8613-e1bfac732c40] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1960.311923] env[62824]: DEBUG nova.network.neutron [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Updating instance_info_cache with network_info: [{"id": "1e74959f-64c2-4c37-85f8-fa8f81a30d60", "address": "fa:16:3e:6d:26:78", "network": {"id": "ae1dfabd-223a-4d4c-a8c4-fe65a83dd9c0", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-797102932-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "102d2f886fd44e1da6780e48e36f1b62", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "517421c3-bea0-419c-ab0b-987815e5d160", "external-id": "nsx-vlan-transportzone-68", "segmentation_id": 68, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1e74959f-64", "ovs_interfaceid": "1e74959f-64c2-4c37-85f8-fa8f81a30d60", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1960.409707] env[62824]: DEBUG oslo_vmware.api [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Task: {'id': task-2145857, 'name': ReconfigVM_Task, 'duration_secs': 0.357495} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1960.410062] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Reconfigured VM instance instance-00000017 to attach disk [datastore2] e158b5d4-c120-4e6c-89c4-7668e097926d/e158b5d4-c120-4e6c-89c4-7668e097926d.vmdk or device None with type thin {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1960.410273] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Updating instance 'e158b5d4-c120-4e6c-89c4-7668e097926d' progress to 50 {{(pid=62824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1960.506460] env[62824]: INFO nova.compute.manager [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: 6d80ec06-8559-4964-8577-a2512aa366ed] Took 45.85 seconds to build instance. [ 1960.817020] env[62824]: DEBUG oslo_concurrency.lockutils [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Releasing lock "refresh_cache-59b5b883-4188-471c-8862-444f3ce08cb0" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1960.817020] env[62824]: DEBUG nova.compute.manager [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Instance network_info: |[{"id": "1e74959f-64c2-4c37-85f8-fa8f81a30d60", "address": "fa:16:3e:6d:26:78", "network": {"id": "ae1dfabd-223a-4d4c-a8c4-fe65a83dd9c0", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-797102932-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "102d2f886fd44e1da6780e48e36f1b62", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "517421c3-bea0-419c-ab0b-987815e5d160", "external-id": "nsx-vlan-transportzone-68", "segmentation_id": 68, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1e74959f-64", "ovs_interfaceid": "1e74959f-64c2-4c37-85f8-fa8f81a30d60", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1960.817260] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6d:26:78', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '517421c3-bea0-419c-ab0b-987815e5d160', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1e74959f-64c2-4c37-85f8-fa8f81a30d60', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1960.823239] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Creating folder: Project (102d2f886fd44e1da6780e48e36f1b62). Parent ref: group-v438503. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1960.823667] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f3c1e20f-9a4a-4252-8558-c822bbc7fbc0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1960.837120] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Created folder: Project (102d2f886fd44e1da6780e48e36f1b62) in parent group-v438503. [ 1960.837120] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Creating folder: Instances. Parent ref: group-v438573. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1960.837120] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-12d5f746-a36b-4cad-8f91-b1996612b679 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1960.846589] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Created folder: Instances in parent group-v438573. [ 1960.847034] env[62824]: DEBUG oslo.service.loopingcall [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1960.847438] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1960.847785] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3079efab-c1aa-4fc8-a1f4-466878fd3e08 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1960.868137] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1960.868137] env[62824]: value = "task-2145860" [ 1960.868137] env[62824]: _type = "Task" [ 1960.868137] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1960.876212] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145860, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1960.912651] env[62824]: DEBUG nova.compute.manager [req-01aabc16-7335-46bb-8e4d-74571ce2e5dd req-07109705-93ea-4371-9959-eb5b4be5bf8d service nova] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Received event network-changed-1e74959f-64c2-4c37-85f8-fa8f81a30d60 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1960.912651] env[62824]: DEBUG nova.compute.manager [req-01aabc16-7335-46bb-8e4d-74571ce2e5dd req-07109705-93ea-4371-9959-eb5b4be5bf8d service nova] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Refreshing instance network info cache due to event network-changed-1e74959f-64c2-4c37-85f8-fa8f81a30d60. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1960.912875] env[62824]: DEBUG oslo_concurrency.lockutils [req-01aabc16-7335-46bb-8e4d-74571ce2e5dd req-07109705-93ea-4371-9959-eb5b4be5bf8d service nova] Acquiring lock "refresh_cache-59b5b883-4188-471c-8862-444f3ce08cb0" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1960.913032] env[62824]: DEBUG oslo_concurrency.lockutils [req-01aabc16-7335-46bb-8e4d-74571ce2e5dd req-07109705-93ea-4371-9959-eb5b4be5bf8d service nova] Acquired lock "refresh_cache-59b5b883-4188-471c-8862-444f3ce08cb0" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1960.913255] env[62824]: DEBUG nova.network.neutron [req-01aabc16-7335-46bb-8e4d-74571ce2e5dd req-07109705-93ea-4371-9959-eb5b4be5bf8d service nova] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Refreshing network info cache for port 1e74959f-64c2-4c37-85f8-fa8f81a30d60 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1960.917202] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-545b8101-e094-4163-8d46-58c7e530bd5d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1960.938509] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c34cb880-1014-4a27-addc-29c67af3884a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1960.956940] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Updating instance 'e158b5d4-c120-4e6c-89c4-7668e097926d' progress to 67 {{(pid=62824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1961.008348] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ee699d94-97b4-4885-b6ef-f6b8170129bd tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Lock "6d80ec06-8559-4964-8577-a2512aa366ed" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 91.224s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1961.089589] env[62824]: DEBUG nova.compute.manager [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 487c2c9d-2cd2-4912-8613-e1bfac732c40] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1961.117228] env[62824]: DEBUG nova.virt.hardware [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1961.117459] env[62824]: DEBUG nova.virt.hardware [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1961.118087] env[62824]: DEBUG nova.virt.hardware [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1961.118087] env[62824]: DEBUG nova.virt.hardware [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1961.118087] env[62824]: DEBUG nova.virt.hardware [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1961.118087] env[62824]: DEBUG nova.virt.hardware [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1961.118294] env[62824]: DEBUG nova.virt.hardware [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1961.118443] env[62824]: DEBUG nova.virt.hardware [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1961.118687] env[62824]: DEBUG nova.virt.hardware [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1961.119490] env[62824]: DEBUG nova.virt.hardware [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1961.119490] env[62824]: DEBUG nova.virt.hardware [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1961.120049] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c35c5fe-24aa-4007-94a3-1045c00c4b07 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1961.129225] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f3ed6ba-8de5-4c7e-96ea-9ce6f453cd56 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1961.383829] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145860, 'name': CreateVM_Task, 'duration_secs': 0.419738} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1961.386293] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1961.387328] env[62824]: DEBUG oslo_concurrency.lockutils [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1961.387483] env[62824]: DEBUG oslo_concurrency.lockutils [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1961.387785] env[62824]: DEBUG oslo_concurrency.lockutils [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1961.388035] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c976fafc-8bfb-4b36-ab00-d74b781e247b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1961.393177] env[62824]: DEBUG oslo_vmware.api [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Waiting for the task: (returnval){ [ 1961.393177] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5201eed3-11da-fa1a-d956-4e9f34f0e766" [ 1961.393177] env[62824]: _type = "Task" [ 1961.393177] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1961.403388] env[62824]: DEBUG oslo_vmware.api [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5201eed3-11da-fa1a-d956-4e9f34f0e766, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1961.497256] env[62824]: DEBUG nova.network.neutron [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Port 02e90c28-7f36-4d83-84bb-fec194a04dab binding to destination host cpu-1 is already ACTIVE {{(pid=62824) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1961.513067] env[62824]: DEBUG nova.compute.manager [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 2afecec5-763d-4616-b690-41d3101cfc47] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1961.589363] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a729b8ac-84cd-4f93-97dd-c0cf0e821773 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1961.596858] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f056730-18e9-4896-8ba6-efe2c36e3dee {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1961.635882] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac58f606-d0d8-4b9c-a8e2-a35acd1a0fd5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1961.644306] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2b9a48d-17f2-4124-908a-f9476db8ec03 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1961.657980] env[62824]: DEBUG nova.compute.provider_tree [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1961.756839] env[62824]: DEBUG nova.network.neutron [req-01aabc16-7335-46bb-8e4d-74571ce2e5dd req-07109705-93ea-4371-9959-eb5b4be5bf8d service nova] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Updated VIF entry in instance network info cache for port 1e74959f-64c2-4c37-85f8-fa8f81a30d60. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1961.757214] env[62824]: DEBUG nova.network.neutron [req-01aabc16-7335-46bb-8e4d-74571ce2e5dd req-07109705-93ea-4371-9959-eb5b4be5bf8d service nova] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Updating instance_info_cache with network_info: [{"id": "1e74959f-64c2-4c37-85f8-fa8f81a30d60", "address": "fa:16:3e:6d:26:78", "network": {"id": "ae1dfabd-223a-4d4c-a8c4-fe65a83dd9c0", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-797102932-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "102d2f886fd44e1da6780e48e36f1b62", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "517421c3-bea0-419c-ab0b-987815e5d160", "external-id": "nsx-vlan-transportzone-68", "segmentation_id": 68, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1e74959f-64", "ovs_interfaceid": "1e74959f-64c2-4c37-85f8-fa8f81a30d60", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1961.823589] env[62824]: DEBUG nova.network.neutron [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 487c2c9d-2cd2-4912-8613-e1bfac732c40] Successfully updated port: bbbe4627-156d-462e-a430-0e4a08a4c2dc {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1961.873194] env[62824]: INFO nova.compute.manager [None req-86a2990d-e2a8-4265-a7f4-87f7818ab659 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: 6d80ec06-8559-4964-8577-a2512aa366ed] Rescuing [ 1961.873510] env[62824]: DEBUG oslo_concurrency.lockutils [None req-86a2990d-e2a8-4265-a7f4-87f7818ab659 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Acquiring lock "refresh_cache-6d80ec06-8559-4964-8577-a2512aa366ed" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1961.873672] env[62824]: DEBUG oslo_concurrency.lockutils [None req-86a2990d-e2a8-4265-a7f4-87f7818ab659 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Acquired lock "refresh_cache-6d80ec06-8559-4964-8577-a2512aa366ed" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1961.873851] env[62824]: DEBUG nova.network.neutron [None req-86a2990d-e2a8-4265-a7f4-87f7818ab659 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: 6d80ec06-8559-4964-8577-a2512aa366ed] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1961.906119] env[62824]: DEBUG oslo_vmware.api [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5201eed3-11da-fa1a-d956-4e9f34f0e766, 'name': SearchDatastore_Task, 'duration_secs': 0.010293} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1961.906408] env[62824]: DEBUG oslo_concurrency.lockutils [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1961.906631] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1961.906855] env[62824]: DEBUG oslo_concurrency.lockutils [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1961.907054] env[62824]: DEBUG oslo_concurrency.lockutils [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1961.907180] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1961.907475] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a7428366-87b3-41b7-8034-c95eb88a7362 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1961.915584] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1961.915584] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1961.916540] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-60984644-96fd-4942-b61d-8398d87db9d0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1961.921423] env[62824]: DEBUG oslo_vmware.api [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Waiting for the task: (returnval){ [ 1961.921423] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52bef49a-1704-5993-acc4-e7ab941a72a8" [ 1961.921423] env[62824]: _type = "Task" [ 1961.921423] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1961.928948] env[62824]: DEBUG oslo_vmware.api [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52bef49a-1704-5993-acc4-e7ab941a72a8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1962.037086] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1962.161093] env[62824]: DEBUG nova.scheduler.client.report [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1962.260435] env[62824]: DEBUG oslo_concurrency.lockutils [req-01aabc16-7335-46bb-8e4d-74571ce2e5dd req-07109705-93ea-4371-9959-eb5b4be5bf8d service nova] Releasing lock "refresh_cache-59b5b883-4188-471c-8862-444f3ce08cb0" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1962.327036] env[62824]: DEBUG oslo_concurrency.lockutils [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquiring lock "refresh_cache-487c2c9d-2cd2-4912-8613-e1bfac732c40" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1962.327036] env[62824]: DEBUG oslo_concurrency.lockutils [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquired lock "refresh_cache-487c2c9d-2cd2-4912-8613-e1bfac732c40" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1962.327036] env[62824]: DEBUG nova.network.neutron [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 487c2c9d-2cd2-4912-8613-e1bfac732c40] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1962.431849] env[62824]: DEBUG oslo_vmware.api [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52bef49a-1704-5993-acc4-e7ab941a72a8, 'name': SearchDatastore_Task, 'duration_secs': 0.011278} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1962.432654] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-94b755c8-7b9f-4c1e-bf7b-505c862df32a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1962.438136] env[62824]: DEBUG oslo_vmware.api [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Waiting for the task: (returnval){ [ 1962.438136] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52293e41-0517-c0a3-22af-6b11faa1231d" [ 1962.438136] env[62824]: _type = "Task" [ 1962.438136] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1962.445963] env[62824]: DEBUG oslo_vmware.api [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52293e41-0517-c0a3-22af-6b11faa1231d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1962.524118] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Acquiring lock "e158b5d4-c120-4e6c-89c4-7668e097926d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1962.524434] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Lock "e158b5d4-c120-4e6c-89c4-7668e097926d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1962.524557] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Lock "e158b5d4-c120-4e6c-89c4-7668e097926d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1962.595188] env[62824]: DEBUG nova.network.neutron [None req-86a2990d-e2a8-4265-a7f4-87f7818ab659 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: 6d80ec06-8559-4964-8577-a2512aa366ed] Updating instance_info_cache with network_info: [{"id": "24e47f2d-ba7e-48db-9b98-c4f3f6a78930", "address": "fa:16:3e:62:14:0e", "network": {"id": "a410f4f3-d072-4034-b797-8210e1a79093", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1482574391-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e10208040df34eb5adfe8dcbc76043d5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f4399275-8e92-4448-be9e-d4984e93e89c", "external-id": "nsx-vlan-transportzone-192", "segmentation_id": 192, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap24e47f2d-ba", "ovs_interfaceid": "24e47f2d-ba7e-48db-9b98-c4f3f6a78930", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1962.666422] env[62824]: DEBUG oslo_concurrency.lockutils [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.600s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1962.666947] env[62824]: DEBUG nova.compute.manager [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] [instance: 69bcac33-d6ac-4ee7-b674-58ab25389149] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1962.669796] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b27ea9b2-e866-4d82-b8e1-ed0ef4938234 tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 32.014s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1962.670016] env[62824]: DEBUG nova.objects.instance [None req-b27ea9b2-e866-4d82-b8e1-ed0ef4938234 tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Lazy-loading 'resources' on Instance uuid 384a96ea-40ce-43e1-a5f9-82f50b710b0b {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1962.863189] env[62824]: DEBUG nova.network.neutron [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 487c2c9d-2cd2-4912-8613-e1bfac732c40] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1962.944751] env[62824]: DEBUG nova.compute.manager [req-1551207f-8dcd-4029-bb1a-f31aa5e9fe74 req-25958e54-e6cc-46f3-96e6-63548b70c63d service nova] [instance: 487c2c9d-2cd2-4912-8613-e1bfac732c40] Received event network-vif-plugged-bbbe4627-156d-462e-a430-0e4a08a4c2dc {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1962.944979] env[62824]: DEBUG oslo_concurrency.lockutils [req-1551207f-8dcd-4029-bb1a-f31aa5e9fe74 req-25958e54-e6cc-46f3-96e6-63548b70c63d service nova] Acquiring lock "487c2c9d-2cd2-4912-8613-e1bfac732c40-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1962.945201] env[62824]: DEBUG oslo_concurrency.lockutils [req-1551207f-8dcd-4029-bb1a-f31aa5e9fe74 req-25958e54-e6cc-46f3-96e6-63548b70c63d service nova] Lock "487c2c9d-2cd2-4912-8613-e1bfac732c40-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1962.945368] env[62824]: DEBUG oslo_concurrency.lockutils [req-1551207f-8dcd-4029-bb1a-f31aa5e9fe74 req-25958e54-e6cc-46f3-96e6-63548b70c63d service nova] Lock "487c2c9d-2cd2-4912-8613-e1bfac732c40-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1962.945535] env[62824]: DEBUG nova.compute.manager [req-1551207f-8dcd-4029-bb1a-f31aa5e9fe74 req-25958e54-e6cc-46f3-96e6-63548b70c63d service nova] [instance: 487c2c9d-2cd2-4912-8613-e1bfac732c40] No waiting events found dispatching network-vif-plugged-bbbe4627-156d-462e-a430-0e4a08a4c2dc {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1962.945726] env[62824]: WARNING nova.compute.manager [req-1551207f-8dcd-4029-bb1a-f31aa5e9fe74 req-25958e54-e6cc-46f3-96e6-63548b70c63d service nova] [instance: 487c2c9d-2cd2-4912-8613-e1bfac732c40] Received unexpected event network-vif-plugged-bbbe4627-156d-462e-a430-0e4a08a4c2dc for instance with vm_state building and task_state spawning. [ 1962.945866] env[62824]: DEBUG nova.compute.manager [req-1551207f-8dcd-4029-bb1a-f31aa5e9fe74 req-25958e54-e6cc-46f3-96e6-63548b70c63d service nova] [instance: 487c2c9d-2cd2-4912-8613-e1bfac732c40] Received event network-changed-bbbe4627-156d-462e-a430-0e4a08a4c2dc {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1962.946026] env[62824]: DEBUG nova.compute.manager [req-1551207f-8dcd-4029-bb1a-f31aa5e9fe74 req-25958e54-e6cc-46f3-96e6-63548b70c63d service nova] [instance: 487c2c9d-2cd2-4912-8613-e1bfac732c40] Refreshing instance network info cache due to event network-changed-bbbe4627-156d-462e-a430-0e4a08a4c2dc. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1962.946197] env[62824]: DEBUG oslo_concurrency.lockutils [req-1551207f-8dcd-4029-bb1a-f31aa5e9fe74 req-25958e54-e6cc-46f3-96e6-63548b70c63d service nova] Acquiring lock "refresh_cache-487c2c9d-2cd2-4912-8613-e1bfac732c40" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1962.950179] env[62824]: DEBUG oslo_vmware.api [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52293e41-0517-c0a3-22af-6b11faa1231d, 'name': SearchDatastore_Task, 'duration_secs': 0.010163} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1962.950432] env[62824]: DEBUG oslo_concurrency.lockutils [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1962.950670] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 59b5b883-4188-471c-8862-444f3ce08cb0/59b5b883-4188-471c-8862-444f3ce08cb0.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1962.950924] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b0ec1560-1ff8-42c9-94ec-d6a2dec0fbe3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1962.961056] env[62824]: DEBUG oslo_vmware.api [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Waiting for the task: (returnval){ [ 1962.961056] env[62824]: value = "task-2145861" [ 1962.961056] env[62824]: _type = "Task" [ 1962.961056] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1962.970232] env[62824]: DEBUG oslo_vmware.api [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Task: {'id': task-2145861, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1963.012438] env[62824]: DEBUG nova.network.neutron [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 487c2c9d-2cd2-4912-8613-e1bfac732c40] Updating instance_info_cache with network_info: [{"id": "bbbe4627-156d-462e-a430-0e4a08a4c2dc", "address": "fa:16:3e:47:e4:b8", "network": {"id": "b995934c-4934-4469-884c-73cb4de1796f", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1892307918-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c11aef7fb6247cb9b5272a6a063cd12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15165046-2de9-4ada-9e99-0126e20854a9", "external-id": "nsx-vlan-transportzone-974", "segmentation_id": 974, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbbbe4627-15", "ovs_interfaceid": "bbbe4627-156d-462e-a430-0e4a08a4c2dc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1963.097614] env[62824]: DEBUG oslo_concurrency.lockutils [None req-86a2990d-e2a8-4265-a7f4-87f7818ab659 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Releasing lock "refresh_cache-6d80ec06-8559-4964-8577-a2512aa366ed" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1963.173299] env[62824]: DEBUG nova.compute.utils [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1963.174891] env[62824]: DEBUG nova.compute.manager [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] [instance: 69bcac33-d6ac-4ee7-b674-58ab25389149] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1963.175087] env[62824]: DEBUG nova.network.neutron [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] [instance: 69bcac33-d6ac-4ee7-b674-58ab25389149] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1963.227622] env[62824]: DEBUG nova.policy [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fbcc53ffb5a64e939883c9384286add7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c0b48e247f3e4a3b9f42cb780955279f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 1963.476335] env[62824]: DEBUG oslo_vmware.api [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Task: {'id': task-2145861, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.492429} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1963.476598] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 59b5b883-4188-471c-8862-444f3ce08cb0/59b5b883-4188-471c-8862-444f3ce08cb0.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1963.476812] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1963.477122] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4a92a2b1-47a8-4f4a-af08-019da36d5fe3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1963.488030] env[62824]: DEBUG oslo_vmware.api [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Waiting for the task: (returnval){ [ 1963.488030] env[62824]: value = "task-2145862" [ 1963.488030] env[62824]: _type = "Task" [ 1963.488030] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1963.499581] env[62824]: DEBUG oslo_vmware.api [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Task: {'id': task-2145862, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1963.517088] env[62824]: DEBUG oslo_concurrency.lockutils [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Releasing lock "refresh_cache-487c2c9d-2cd2-4912-8613-e1bfac732c40" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1963.517088] env[62824]: DEBUG nova.compute.manager [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 487c2c9d-2cd2-4912-8613-e1bfac732c40] Instance network_info: |[{"id": "bbbe4627-156d-462e-a430-0e4a08a4c2dc", "address": "fa:16:3e:47:e4:b8", "network": {"id": "b995934c-4934-4469-884c-73cb4de1796f", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1892307918-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c11aef7fb6247cb9b5272a6a063cd12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15165046-2de9-4ada-9e99-0126e20854a9", "external-id": "nsx-vlan-transportzone-974", "segmentation_id": 974, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbbbe4627-15", "ovs_interfaceid": "bbbe4627-156d-462e-a430-0e4a08a4c2dc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1963.517744] env[62824]: DEBUG oslo_concurrency.lockutils [req-1551207f-8dcd-4029-bb1a-f31aa5e9fe74 req-25958e54-e6cc-46f3-96e6-63548b70c63d service nova] Acquired lock "refresh_cache-487c2c9d-2cd2-4912-8613-e1bfac732c40" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1963.517744] env[62824]: DEBUG nova.network.neutron [req-1551207f-8dcd-4029-bb1a-f31aa5e9fe74 req-25958e54-e6cc-46f3-96e6-63548b70c63d service nova] [instance: 487c2c9d-2cd2-4912-8613-e1bfac732c40] Refreshing network info cache for port bbbe4627-156d-462e-a430-0e4a08a4c2dc {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1963.517812] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 487c2c9d-2cd2-4912-8613-e1bfac732c40] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:47:e4:b8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '15165046-2de9-4ada-9e99-0126e20854a9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bbbe4627-156d-462e-a430-0e4a08a4c2dc', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1963.528950] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Creating folder: Project (0c11aef7fb6247cb9b5272a6a063cd12). Parent ref: group-v438503. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1963.538437] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fec28b1a-8db7-41fe-9594-bc9239cafc1a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1963.549248] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Created folder: Project (0c11aef7fb6247cb9b5272a6a063cd12) in parent group-v438503. [ 1963.549440] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Creating folder: Instances. Parent ref: group-v438576. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1963.550156] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-52b86637-bfe0-43cc-b633-0e82185b08e8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1963.562336] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Created folder: Instances in parent group-v438576. [ 1963.562569] env[62824]: DEBUG oslo.service.loopingcall [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1963.562751] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 487c2c9d-2cd2-4912-8613-e1bfac732c40] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1963.562947] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1d123d90-2476-41de-93ff-207ae9aa81a4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1963.582850] env[62824]: DEBUG nova.network.neutron [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] [instance: 69bcac33-d6ac-4ee7-b674-58ab25389149] Successfully created port: 08901969-e1cd-4351-9470-7c2ceabab5c4 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1963.590410] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1963.590410] env[62824]: value = "task-2145865" [ 1963.590410] env[62824]: _type = "Task" [ 1963.590410] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1963.600182] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145865, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1963.620690] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Acquiring lock "refresh_cache-e158b5d4-c120-4e6c-89c4-7668e097926d" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1963.620690] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Acquired lock "refresh_cache-e158b5d4-c120-4e6c-89c4-7668e097926d" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1963.620690] env[62824]: DEBUG nova.network.neutron [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1963.683228] env[62824]: DEBUG nova.compute.manager [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] [instance: 69bcac33-d6ac-4ee7-b674-58ab25389149] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1963.800974] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0cce7ed-a76d-44cf-8c99-a773cab67873 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1963.808872] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc7fd497-4bb3-4982-b192-db0ed05013fa {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1963.845096] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffb2f28a-7176-4cb4-a7a9-abc5018497c2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1963.852299] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d191d80-52cf-4b23-9d54-fb254c0994cc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1963.867723] env[62824]: DEBUG nova.compute.provider_tree [None req-b27ea9b2-e866-4d82-b8e1-ed0ef4938234 tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1963.920311] env[62824]: DEBUG nova.network.neutron [req-1551207f-8dcd-4029-bb1a-f31aa5e9fe74 req-25958e54-e6cc-46f3-96e6-63548b70c63d service nova] [instance: 487c2c9d-2cd2-4912-8613-e1bfac732c40] Updated VIF entry in instance network info cache for port bbbe4627-156d-462e-a430-0e4a08a4c2dc. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1963.920710] env[62824]: DEBUG nova.network.neutron [req-1551207f-8dcd-4029-bb1a-f31aa5e9fe74 req-25958e54-e6cc-46f3-96e6-63548b70c63d service nova] [instance: 487c2c9d-2cd2-4912-8613-e1bfac732c40] Updating instance_info_cache with network_info: [{"id": "bbbe4627-156d-462e-a430-0e4a08a4c2dc", "address": "fa:16:3e:47:e4:b8", "network": {"id": "b995934c-4934-4469-884c-73cb4de1796f", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1892307918-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c11aef7fb6247cb9b5272a6a063cd12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15165046-2de9-4ada-9e99-0126e20854a9", "external-id": "nsx-vlan-transportzone-974", "segmentation_id": 974, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbbbe4627-15", "ovs_interfaceid": "bbbe4627-156d-462e-a430-0e4a08a4c2dc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1963.997700] env[62824]: DEBUG oslo_vmware.api [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Task: {'id': task-2145862, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07189} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1963.997959] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1963.999070] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a44714b8-c1cd-419e-a478-6968f093c8e0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1964.020750] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Reconfiguring VM instance instance-00000022 to attach disk [datastore1] 59b5b883-4188-471c-8862-444f3ce08cb0/59b5b883-4188-471c-8862-444f3ce08cb0.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1964.021064] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6748e1bf-8dd2-4e13-91ae-b16ee762e54a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1964.040436] env[62824]: DEBUG oslo_vmware.api [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Waiting for the task: (returnval){ [ 1964.040436] env[62824]: value = "task-2145866" [ 1964.040436] env[62824]: _type = "Task" [ 1964.040436] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1964.048443] env[62824]: DEBUG oslo_vmware.api [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Task: {'id': task-2145866, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1964.099624] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145865, 'name': CreateVM_Task, 'duration_secs': 0.387688} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1964.099792] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 487c2c9d-2cd2-4912-8613-e1bfac732c40] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1964.100543] env[62824]: DEBUG oslo_concurrency.lockutils [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1964.100724] env[62824]: DEBUG oslo_concurrency.lockutils [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1964.101046] env[62824]: DEBUG oslo_concurrency.lockutils [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1964.101368] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-724bba31-c8c2-4922-8bfc-b443426f3b89 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1964.105820] env[62824]: DEBUG oslo_vmware.api [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 1964.105820] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52030ac1-dd7d-32cc-488c-691bf4ef8828" [ 1964.105820] env[62824]: _type = "Task" [ 1964.105820] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1964.113597] env[62824]: DEBUG oslo_vmware.api [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52030ac1-dd7d-32cc-488c-691bf4ef8828, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1964.323510] env[62824]: DEBUG nova.network.neutron [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Updating instance_info_cache with network_info: [{"id": "02e90c28-7f36-4d83-84bb-fec194a04dab", "address": "fa:16:3e:53:1a:4c", "network": {"id": "b6403dc3-30ef-4c76-b521-380c179f345d", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.19", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "22c5c40c2e004babaa15896c89307303", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1002b79b-224e-41e3-a484-4245a767147a", "external-id": "nsx-vlan-transportzone-353", "segmentation_id": 353, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap02e90c28-7f", "ovs_interfaceid": "02e90c28-7f36-4d83-84bb-fec194a04dab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1964.370637] env[62824]: DEBUG nova.scheduler.client.report [None req-b27ea9b2-e866-4d82-b8e1-ed0ef4938234 tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1964.423416] env[62824]: DEBUG oslo_concurrency.lockutils [req-1551207f-8dcd-4029-bb1a-f31aa5e9fe74 req-25958e54-e6cc-46f3-96e6-63548b70c63d service nova] Releasing lock "refresh_cache-487c2c9d-2cd2-4912-8613-e1bfac732c40" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1964.552910] env[62824]: DEBUG oslo_vmware.api [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Task: {'id': task-2145866, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1964.616252] env[62824]: DEBUG oslo_vmware.api [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52030ac1-dd7d-32cc-488c-691bf4ef8828, 'name': SearchDatastore_Task, 'duration_secs': 0.009344} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1964.616613] env[62824]: DEBUG oslo_concurrency.lockutils [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1964.616911] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 487c2c9d-2cd2-4912-8613-e1bfac732c40] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1964.617131] env[62824]: DEBUG oslo_concurrency.lockutils [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1964.617444] env[62824]: DEBUG oslo_concurrency.lockutils [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1964.617498] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1964.617871] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fb4af8f3-89cf-49d7-9903-e9bad0190b02 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1964.626892] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1964.627123] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1964.627892] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-666c5e57-b0ca-4726-bcca-db25e6c86005 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1964.633558] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-86a2990d-e2a8-4265-a7f4-87f7818ab659 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: 6d80ec06-8559-4964-8577-a2512aa366ed] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1964.633887] env[62824]: DEBUG oslo_vmware.api [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 1964.633887] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]523e055d-1989-64fd-11af-fc4ec9d1c443" [ 1964.633887] env[62824]: _type = "Task" [ 1964.633887] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1964.634119] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-20a3d4cd-97f4-4afd-8ca3-995415c1ccf0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1964.644606] env[62824]: DEBUG oslo_vmware.api [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]523e055d-1989-64fd-11af-fc4ec9d1c443, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1964.645877] env[62824]: DEBUG oslo_vmware.api [None req-86a2990d-e2a8-4265-a7f4-87f7818ab659 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for the task: (returnval){ [ 1964.645877] env[62824]: value = "task-2145867" [ 1964.645877] env[62824]: _type = "Task" [ 1964.645877] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1964.654166] env[62824]: DEBUG oslo_vmware.api [None req-86a2990d-e2a8-4265-a7f4-87f7818ab659 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2145867, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1964.696624] env[62824]: DEBUG nova.compute.manager [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] [instance: 69bcac33-d6ac-4ee7-b674-58ab25389149] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1964.726887] env[62824]: DEBUG nova.virt.hardware [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1964.727181] env[62824]: DEBUG nova.virt.hardware [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1964.727358] env[62824]: DEBUG nova.virt.hardware [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1964.727540] env[62824]: DEBUG nova.virt.hardware [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1964.727704] env[62824]: DEBUG nova.virt.hardware [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1964.727884] env[62824]: DEBUG nova.virt.hardware [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1964.728196] env[62824]: DEBUG nova.virt.hardware [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1964.728417] env[62824]: DEBUG nova.virt.hardware [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1964.728647] env[62824]: DEBUG nova.virt.hardware [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1964.728869] env[62824]: DEBUG nova.virt.hardware [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1964.729123] env[62824]: DEBUG nova.virt.hardware [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1964.730051] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-723ea9c1-2ba4-4efa-b36a-5ce96be8a4c2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1964.737674] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11b87439-077a-4abd-8578-ca06f9991a65 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1964.828795] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Releasing lock "refresh_cache-e158b5d4-c120-4e6c-89c4-7668e097926d" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1964.877089] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b27ea9b2-e866-4d82-b8e1-ed0ef4938234 tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.207s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1964.879662] env[62824]: DEBUG oslo_concurrency.lockutils [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 33.896s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1964.881621] env[62824]: INFO nova.compute.claims [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] [instance: 5f1237ff-5841-41f1-8d77-df9fabaa62b6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1964.901888] env[62824]: INFO nova.scheduler.client.report [None req-b27ea9b2-e866-4d82-b8e1-ed0ef4938234 tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Deleted allocations for instance 384a96ea-40ce-43e1-a5f9-82f50b710b0b [ 1965.051362] env[62824]: DEBUG oslo_vmware.api [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Task: {'id': task-2145866, 'name': ReconfigVM_Task, 'duration_secs': 0.728932} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1965.051483] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Reconfigured VM instance instance-00000022 to attach disk [datastore1] 59b5b883-4188-471c-8862-444f3ce08cb0/59b5b883-4188-471c-8862-444f3ce08cb0.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1965.052124] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6895868c-72a3-4326-b1fd-642e08834a89 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1965.058948] env[62824]: DEBUG oslo_vmware.api [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Waiting for the task: (returnval){ [ 1965.058948] env[62824]: value = "task-2145868" [ 1965.058948] env[62824]: _type = "Task" [ 1965.058948] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1965.067036] env[62824]: DEBUG oslo_vmware.api [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Task: {'id': task-2145868, 'name': Rename_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1965.147802] env[62824]: DEBUG oslo_vmware.api [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]523e055d-1989-64fd-11af-fc4ec9d1c443, 'name': SearchDatastore_Task, 'duration_secs': 0.010174} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1965.152356] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3089db18-8cb4-4ee1-ac5e-0e892c9e20bf {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1965.157794] env[62824]: DEBUG nova.compute.manager [req-43fc2466-1f1c-4d09-bd00-cc47a2b0eb28 req-76ac4a3e-4bc8-492a-8ad8-7c233ee4a233 service nova] [instance: 69bcac33-d6ac-4ee7-b674-58ab25389149] Received event network-vif-plugged-08901969-e1cd-4351-9470-7c2ceabab5c4 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1965.157794] env[62824]: DEBUG oslo_concurrency.lockutils [req-43fc2466-1f1c-4d09-bd00-cc47a2b0eb28 req-76ac4a3e-4bc8-492a-8ad8-7c233ee4a233 service nova] Acquiring lock "69bcac33-d6ac-4ee7-b674-58ab25389149-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1965.157977] env[62824]: DEBUG oslo_concurrency.lockutils [req-43fc2466-1f1c-4d09-bd00-cc47a2b0eb28 req-76ac4a3e-4bc8-492a-8ad8-7c233ee4a233 service nova] Lock "69bcac33-d6ac-4ee7-b674-58ab25389149-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1965.158188] env[62824]: DEBUG oslo_concurrency.lockutils [req-43fc2466-1f1c-4d09-bd00-cc47a2b0eb28 req-76ac4a3e-4bc8-492a-8ad8-7c233ee4a233 service nova] Lock "69bcac33-d6ac-4ee7-b674-58ab25389149-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1965.158361] env[62824]: DEBUG nova.compute.manager [req-43fc2466-1f1c-4d09-bd00-cc47a2b0eb28 req-76ac4a3e-4bc8-492a-8ad8-7c233ee4a233 service nova] [instance: 69bcac33-d6ac-4ee7-b674-58ab25389149] No waiting events found dispatching network-vif-plugged-08901969-e1cd-4351-9470-7c2ceabab5c4 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1965.158525] env[62824]: WARNING nova.compute.manager [req-43fc2466-1f1c-4d09-bd00-cc47a2b0eb28 req-76ac4a3e-4bc8-492a-8ad8-7c233ee4a233 service nova] [instance: 69bcac33-d6ac-4ee7-b674-58ab25389149] Received unexpected event network-vif-plugged-08901969-e1cd-4351-9470-7c2ceabab5c4 for instance with vm_state building and task_state spawning. [ 1965.164658] env[62824]: DEBUG oslo_vmware.api [None req-86a2990d-e2a8-4265-a7f4-87f7818ab659 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2145867, 'name': PowerOffVM_Task, 'duration_secs': 0.168041} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1965.165571] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-86a2990d-e2a8-4265-a7f4-87f7818ab659 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: 6d80ec06-8559-4964-8577-a2512aa366ed] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1965.165922] env[62824]: DEBUG oslo_vmware.api [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 1965.165922] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52a7a655-66a0-62ff-9eb6-8a1fa0c7a402" [ 1965.165922] env[62824]: _type = "Task" [ 1965.165922] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1965.166612] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c5c3d97-7c87-4300-9bdc-d60f8a41f8e9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1965.177924] env[62824]: DEBUG oslo_vmware.api [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52a7a655-66a0-62ff-9eb6-8a1fa0c7a402, 'name': SearchDatastore_Task, 'duration_secs': 0.010071} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1965.191718] env[62824]: DEBUG oslo_concurrency.lockutils [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1965.192028] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 487c2c9d-2cd2-4912-8613-e1bfac732c40/487c2c9d-2cd2-4912-8613-e1bfac732c40.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1965.192432] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-274a7f58-9396-4dcb-9ee1-c97e74cffd70 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1965.195144] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a76daf82-60cb-431d-9d27-e02d2fcfa747 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1965.206659] env[62824]: DEBUG oslo_vmware.api [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 1965.206659] env[62824]: value = "task-2145869" [ 1965.206659] env[62824]: _type = "Task" [ 1965.206659] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1965.214644] env[62824]: DEBUG oslo_vmware.api [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2145869, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1965.237675] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-86a2990d-e2a8-4265-a7f4-87f7818ab659 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: 6d80ec06-8559-4964-8577-a2512aa366ed] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1965.237975] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-278293da-2572-48bb-868f-615a7d05e97f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1965.243874] env[62824]: DEBUG oslo_vmware.api [None req-86a2990d-e2a8-4265-a7f4-87f7818ab659 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for the task: (returnval){ [ 1965.243874] env[62824]: value = "task-2145870" [ 1965.243874] env[62824]: _type = "Task" [ 1965.243874] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1965.252665] env[62824]: DEBUG oslo_vmware.api [None req-86a2990d-e2a8-4265-a7f4-87f7818ab659 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2145870, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1965.260105] env[62824]: DEBUG nova.network.neutron [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] [instance: 69bcac33-d6ac-4ee7-b674-58ab25389149] Successfully updated port: 08901969-e1cd-4351-9470-7c2ceabab5c4 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1965.352183] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb381a1c-857e-4029-892f-f4eeccd67cca {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1965.374682] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2df1f75-9051-432e-8ebb-2f18677ed9e5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1965.384188] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Updating instance 'e158b5d4-c120-4e6c-89c4-7668e097926d' progress to 83 {{(pid=62824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1965.410028] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b27ea9b2-e866-4d82-b8e1-ed0ef4938234 tempest-ServersAdmin275Test-162219430 tempest-ServersAdmin275Test-162219430-project-member] Lock "384a96ea-40ce-43e1-a5f9-82f50b710b0b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 38.993s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1965.570930] env[62824]: DEBUG oslo_vmware.api [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Task: {'id': task-2145868, 'name': Rename_Task, 'duration_secs': 0.157308} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1965.572401] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1965.572756] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a0287f62-ced6-4ac0-be23-7201423028d0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1965.579340] env[62824]: DEBUG oslo_vmware.api [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Waiting for the task: (returnval){ [ 1965.579340] env[62824]: value = "task-2145871" [ 1965.579340] env[62824]: _type = "Task" [ 1965.579340] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1965.588361] env[62824]: DEBUG oslo_vmware.api [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Task: {'id': task-2145871, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1965.718426] env[62824]: DEBUG oslo_vmware.api [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2145869, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.465477} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1965.718426] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 487c2c9d-2cd2-4912-8613-e1bfac732c40/487c2c9d-2cd2-4912-8613-e1bfac732c40.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1965.718426] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 487c2c9d-2cd2-4912-8613-e1bfac732c40] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1965.718426] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3336887f-0de2-4e3a-a20a-3db158aa96e4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1965.725853] env[62824]: DEBUG oslo_vmware.api [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 1965.725853] env[62824]: value = "task-2145872" [ 1965.725853] env[62824]: _type = "Task" [ 1965.725853] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1965.741991] env[62824]: DEBUG oslo_vmware.api [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2145872, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1965.761020] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-86a2990d-e2a8-4265-a7f4-87f7818ab659 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: 6d80ec06-8559-4964-8577-a2512aa366ed] VM already powered off {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1965.761020] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-86a2990d-e2a8-4265-a7f4-87f7818ab659 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: 6d80ec06-8559-4964-8577-a2512aa366ed] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1965.761020] env[62824]: DEBUG oslo_concurrency.lockutils [None req-86a2990d-e2a8-4265-a7f4-87f7818ab659 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1965.761020] env[62824]: DEBUG oslo_concurrency.lockutils [None req-86a2990d-e2a8-4265-a7f4-87f7818ab659 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1965.761316] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-86a2990d-e2a8-4265-a7f4-87f7818ab659 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1965.761316] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dbc246c3-91f3-4541-b8d1-3a923bc2b2cb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1965.762930] env[62824]: DEBUG oslo_concurrency.lockutils [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Acquiring lock "refresh_cache-69bcac33-d6ac-4ee7-b674-58ab25389149" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1965.763188] env[62824]: DEBUG oslo_concurrency.lockutils [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Acquired lock "refresh_cache-69bcac33-d6ac-4ee7-b674-58ab25389149" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1965.763357] env[62824]: DEBUG nova.network.neutron [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] [instance: 69bcac33-d6ac-4ee7-b674-58ab25389149] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1965.772856] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-86a2990d-e2a8-4265-a7f4-87f7818ab659 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1965.772856] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-86a2990d-e2a8-4265-a7f4-87f7818ab659 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1965.773989] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6ae37c21-c604-4f2f-9489-8abd818e47d8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1965.780652] env[62824]: DEBUG oslo_vmware.api [None req-86a2990d-e2a8-4265-a7f4-87f7818ab659 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for the task: (returnval){ [ 1965.780652] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]524af9ff-2abe-68d0-0d68-42b89e47a12d" [ 1965.780652] env[62824]: _type = "Task" [ 1965.780652] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1965.789339] env[62824]: DEBUG oslo_vmware.api [None req-86a2990d-e2a8-4265-a7f4-87f7818ab659 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]524af9ff-2abe-68d0-0d68-42b89e47a12d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1965.891737] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1965.894572] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8081d918-ba7c-4a69-b73e-66d047777858 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1965.902077] env[62824]: DEBUG oslo_vmware.api [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Waiting for the task: (returnval){ [ 1965.902077] env[62824]: value = "task-2145873" [ 1965.902077] env[62824]: _type = "Task" [ 1965.902077] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1965.910622] env[62824]: DEBUG oslo_vmware.api [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Task: {'id': task-2145873, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1966.091666] env[62824]: DEBUG oslo_vmware.api [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Task: {'id': task-2145871, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1966.233675] env[62824]: DEBUG oslo_vmware.api [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2145872, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068735} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1966.236311] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 487c2c9d-2cd2-4912-8613-e1bfac732c40] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1966.237365] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7c4797b-6266-46a8-a75e-6dac02c1fd2d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1966.259993] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 487c2c9d-2cd2-4912-8613-e1bfac732c40] Reconfiguring VM instance instance-00000023 to attach disk [datastore1] 487c2c9d-2cd2-4912-8613-e1bfac732c40/487c2c9d-2cd2-4912-8613-e1bfac732c40.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1966.262855] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-53098653-995b-4876-b572-ecbacb69aa44 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1966.286675] env[62824]: DEBUG oslo_vmware.api [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 1966.286675] env[62824]: value = "task-2145874" [ 1966.286675] env[62824]: _type = "Task" [ 1966.286675] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1966.297035] env[62824]: DEBUG oslo_vmware.api [None req-86a2990d-e2a8-4265-a7f4-87f7818ab659 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]524af9ff-2abe-68d0-0d68-42b89e47a12d, 'name': SearchDatastore_Task, 'duration_secs': 0.009421} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1966.301093] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a1416b0c-a27d-45fd-8c3c-252b4097c0f5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1966.303758] env[62824]: DEBUG oslo_vmware.api [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2145874, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1966.310109] env[62824]: DEBUG oslo_vmware.api [None req-86a2990d-e2a8-4265-a7f4-87f7818ab659 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for the task: (returnval){ [ 1966.310109] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]524b6c88-5bd5-7802-e846-9dec032308eb" [ 1966.310109] env[62824]: _type = "Task" [ 1966.310109] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1966.310405] env[62824]: DEBUG nova.network.neutron [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] [instance: 69bcac33-d6ac-4ee7-b674-58ab25389149] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1966.321883] env[62824]: DEBUG oslo_vmware.api [None req-86a2990d-e2a8-4265-a7f4-87f7818ab659 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]524b6c88-5bd5-7802-e846-9dec032308eb, 'name': SearchDatastore_Task, 'duration_secs': 0.012232} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1966.324253] env[62824]: DEBUG oslo_concurrency.lockutils [None req-86a2990d-e2a8-4265-a7f4-87f7818ab659 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1966.324519] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-86a2990d-e2a8-4265-a7f4-87f7818ab659 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 6d80ec06-8559-4964-8577-a2512aa366ed/9e2a7d30-212d-4ab8-9606-c5c6d52629e8-rescue.vmdk. {{(pid=62824) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1966.325098] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5e20054c-4082-4eed-a4b5-d251c5caa3ec {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1966.330775] env[62824]: DEBUG oslo_vmware.api [None req-86a2990d-e2a8-4265-a7f4-87f7818ab659 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for the task: (returnval){ [ 1966.330775] env[62824]: value = "task-2145875" [ 1966.330775] env[62824]: _type = "Task" [ 1966.330775] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1966.342141] env[62824]: DEBUG oslo_vmware.api [None req-86a2990d-e2a8-4265-a7f4-87f7818ab659 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2145875, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1966.414690] env[62824]: DEBUG oslo_vmware.api [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Task: {'id': task-2145873, 'name': PowerOnVM_Task, 'duration_secs': 0.397203} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1966.415234] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1966.415425] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-1f407b7d-ea33-4f8b-bc06-00fc6c3483ee tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Updating instance 'e158b5d4-c120-4e6c-89c4-7668e097926d' progress to 100 {{(pid=62824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1966.430562] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e580d41-17ad-4936-9d89-dfa836a66760 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1966.438060] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7833792c-3178-4e76-915f-0eb38a809a2b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1966.471595] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54fe76ec-cf40-4b93-92ba-02ebd6b58dd7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1966.483745] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d00ea827-6e38-4d7d-bf1d-fcb3a52f48ca {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1966.494410] env[62824]: DEBUG nova.compute.provider_tree [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1966.514130] env[62824]: DEBUG nova.network.neutron [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] [instance: 69bcac33-d6ac-4ee7-b674-58ab25389149] Updating instance_info_cache with network_info: [{"id": "08901969-e1cd-4351-9470-7c2ceabab5c4", "address": "fa:16:3e:e6:13:e2", "network": {"id": "79079501-33e5-4c53-b192-339eaa476dab", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-50257271-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c0b48e247f3e4a3b9f42cb780955279f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69f65356-c85e-4b7f-ad28-7c7b5e8cf50c", "external-id": "nsx-vlan-transportzone-281", "segmentation_id": 281, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap08901969-e1", "ovs_interfaceid": "08901969-e1cd-4351-9470-7c2ceabab5c4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1966.597056] env[62824]: DEBUG oslo_vmware.api [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Task: {'id': task-2145871, 'name': PowerOnVM_Task, 'duration_secs': 0.656226} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1966.597489] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1966.597755] env[62824]: INFO nova.compute.manager [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Took 8.18 seconds to spawn the instance on the hypervisor. [ 1966.598010] env[62824]: DEBUG nova.compute.manager [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1966.598957] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28e62b82-17b9-47ef-a011-37f2e3683102 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1966.798131] env[62824]: DEBUG oslo_vmware.api [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2145874, 'name': ReconfigVM_Task, 'duration_secs': 0.307894} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1966.798574] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 487c2c9d-2cd2-4912-8613-e1bfac732c40] Reconfigured VM instance instance-00000023 to attach disk [datastore1] 487c2c9d-2cd2-4912-8613-e1bfac732c40/487c2c9d-2cd2-4912-8613-e1bfac732c40.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1966.799837] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-69256936-2a85-42cb-aa2c-a213534a9f3b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1966.808790] env[62824]: DEBUG oslo_vmware.api [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 1966.808790] env[62824]: value = "task-2145876" [ 1966.808790] env[62824]: _type = "Task" [ 1966.808790] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1966.818607] env[62824]: DEBUG oslo_vmware.api [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2145876, 'name': Rename_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1966.841739] env[62824]: DEBUG oslo_vmware.api [None req-86a2990d-e2a8-4265-a7f4-87f7818ab659 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2145875, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.496782} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1966.842434] env[62824]: INFO nova.virt.vmwareapi.ds_util [None req-86a2990d-e2a8-4265-a7f4-87f7818ab659 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 6d80ec06-8559-4964-8577-a2512aa366ed/9e2a7d30-212d-4ab8-9606-c5c6d52629e8-rescue.vmdk. [ 1966.844843] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c02d40f-929a-48f2-a3d2-b19cef8dce70 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1966.878019] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-86a2990d-e2a8-4265-a7f4-87f7818ab659 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: 6d80ec06-8559-4964-8577-a2512aa366ed] Reconfiguring VM instance instance-00000021 to attach disk [datastore2] 6d80ec06-8559-4964-8577-a2512aa366ed/9e2a7d30-212d-4ab8-9606-c5c6d52629e8-rescue.vmdk or device None with type thin {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1966.878381] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cccae8b1-512d-4330-92f6-453475a41e11 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1966.897785] env[62824]: DEBUG oslo_vmware.api [None req-86a2990d-e2a8-4265-a7f4-87f7818ab659 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for the task: (returnval){ [ 1966.897785] env[62824]: value = "task-2145877" [ 1966.897785] env[62824]: _type = "Task" [ 1966.897785] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1966.906218] env[62824]: DEBUG oslo_vmware.api [None req-86a2990d-e2a8-4265-a7f4-87f7818ab659 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2145877, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1966.998276] env[62824]: DEBUG nova.scheduler.client.report [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1967.016613] env[62824]: DEBUG oslo_concurrency.lockutils [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Releasing lock "refresh_cache-69bcac33-d6ac-4ee7-b674-58ab25389149" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1967.016985] env[62824]: DEBUG nova.compute.manager [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] [instance: 69bcac33-d6ac-4ee7-b674-58ab25389149] Instance network_info: |[{"id": "08901969-e1cd-4351-9470-7c2ceabab5c4", "address": "fa:16:3e:e6:13:e2", "network": {"id": "79079501-33e5-4c53-b192-339eaa476dab", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-50257271-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c0b48e247f3e4a3b9f42cb780955279f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69f65356-c85e-4b7f-ad28-7c7b5e8cf50c", "external-id": "nsx-vlan-transportzone-281", "segmentation_id": 281, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap08901969-e1", "ovs_interfaceid": "08901969-e1cd-4351-9470-7c2ceabab5c4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1967.017747] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] [instance: 69bcac33-d6ac-4ee7-b674-58ab25389149] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e6:13:e2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '69f65356-c85e-4b7f-ad28-7c7b5e8cf50c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '08901969-e1cd-4351-9470-7c2ceabab5c4', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1967.025727] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Creating folder: Project (c0b48e247f3e4a3b9f42cb780955279f). Parent ref: group-v438503. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1967.026029] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b83dd9ff-0db1-40b7-9f3c-0c84d0ebd22d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1967.036900] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Created folder: Project (c0b48e247f3e4a3b9f42cb780955279f) in parent group-v438503. [ 1967.037166] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Creating folder: Instances. Parent ref: group-v438579. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1967.037408] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0ad99ebc-3b8e-4d43-8c66-42e641c31ad3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1967.046104] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Created folder: Instances in parent group-v438579. [ 1967.046702] env[62824]: DEBUG oslo.service.loopingcall [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1967.046702] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 69bcac33-d6ac-4ee7-b674-58ab25389149] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1967.046702] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-84f20912-ab6c-4a98-9e60-6a8c70fb20aa {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1967.066610] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1967.066610] env[62824]: value = "task-2145880" [ 1967.066610] env[62824]: _type = "Task" [ 1967.066610] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1967.073566] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145880, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1967.124400] env[62824]: INFO nova.compute.manager [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Took 43.15 seconds to build instance. [ 1967.234018] env[62824]: DEBUG nova.compute.manager [req-6f628b28-9e3a-4662-a386-7ad95999d527 req-c6590fe5-7d5b-48a9-b185-d747c4281284 service nova] [instance: 69bcac33-d6ac-4ee7-b674-58ab25389149] Received event network-changed-08901969-e1cd-4351-9470-7c2ceabab5c4 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1967.234266] env[62824]: DEBUG nova.compute.manager [req-6f628b28-9e3a-4662-a386-7ad95999d527 req-c6590fe5-7d5b-48a9-b185-d747c4281284 service nova] [instance: 69bcac33-d6ac-4ee7-b674-58ab25389149] Refreshing instance network info cache due to event network-changed-08901969-e1cd-4351-9470-7c2ceabab5c4. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1967.234487] env[62824]: DEBUG oslo_concurrency.lockutils [req-6f628b28-9e3a-4662-a386-7ad95999d527 req-c6590fe5-7d5b-48a9-b185-d747c4281284 service nova] Acquiring lock "refresh_cache-69bcac33-d6ac-4ee7-b674-58ab25389149" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1967.234653] env[62824]: DEBUG oslo_concurrency.lockutils [req-6f628b28-9e3a-4662-a386-7ad95999d527 req-c6590fe5-7d5b-48a9-b185-d747c4281284 service nova] Acquired lock "refresh_cache-69bcac33-d6ac-4ee7-b674-58ab25389149" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1967.234896] env[62824]: DEBUG nova.network.neutron [req-6f628b28-9e3a-4662-a386-7ad95999d527 req-c6590fe5-7d5b-48a9-b185-d747c4281284 service nova] [instance: 69bcac33-d6ac-4ee7-b674-58ab25389149] Refreshing network info cache for port 08901969-e1cd-4351-9470-7c2ceabab5c4 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1967.320591] env[62824]: DEBUG oslo_vmware.api [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2145876, 'name': Rename_Task, 'duration_secs': 0.153022} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1967.320892] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 487c2c9d-2cd2-4912-8613-e1bfac732c40] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1967.321150] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f1f2c501-ecf0-4241-9053-c53c8231861c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1967.328519] env[62824]: DEBUG oslo_vmware.api [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 1967.328519] env[62824]: value = "task-2145881" [ 1967.328519] env[62824]: _type = "Task" [ 1967.328519] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1967.337897] env[62824]: DEBUG oslo_vmware.api [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2145881, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1967.409385] env[62824]: DEBUG oslo_vmware.api [None req-86a2990d-e2a8-4265-a7f4-87f7818ab659 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2145877, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1967.503224] env[62824]: DEBUG oslo_concurrency.lockutils [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.623s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1967.504341] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 33.991s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1967.506589] env[62824]: INFO nova.compute.claims [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: dc3317e4-6958-4df6-8e39-c89b4a5d861c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1967.576493] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145880, 'name': CreateVM_Task, 'duration_secs': 0.412546} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1967.576695] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 69bcac33-d6ac-4ee7-b674-58ab25389149] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1967.577417] env[62824]: DEBUG oslo_concurrency.lockutils [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1967.577579] env[62824]: DEBUG oslo_concurrency.lockutils [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1967.577902] env[62824]: DEBUG oslo_concurrency.lockutils [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1967.578220] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e21dd423-c8c4-4c49-b775-0e5f19a2c3ff {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1967.582952] env[62824]: DEBUG oslo_vmware.api [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Waiting for the task: (returnval){ [ 1967.582952] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]529f1c0a-491f-46e2-abbd-1d04ceaaffa4" [ 1967.582952] env[62824]: _type = "Task" [ 1967.582952] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1967.590412] env[62824]: DEBUG oslo_vmware.api [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]529f1c0a-491f-46e2-abbd-1d04ceaaffa4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1967.627681] env[62824]: DEBUG oslo_concurrency.lockutils [None req-031f1722-85f3-4dd8-a784-2356ef8d797c tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Lock "59b5b883-4188-471c-8862-444f3ce08cb0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 96.690s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1967.846028] env[62824]: DEBUG oslo_vmware.api [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2145881, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1967.911187] env[62824]: DEBUG oslo_vmware.api [None req-86a2990d-e2a8-4265-a7f4-87f7818ab659 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2145877, 'name': ReconfigVM_Task, 'duration_secs': 0.969577} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1967.913891] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-86a2990d-e2a8-4265-a7f4-87f7818ab659 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: 6d80ec06-8559-4964-8577-a2512aa366ed] Reconfigured VM instance instance-00000021 to attach disk [datastore2] 6d80ec06-8559-4964-8577-a2512aa366ed/9e2a7d30-212d-4ab8-9606-c5c6d52629e8-rescue.vmdk or device None with type thin {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1967.913891] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-116075d2-4376-40e1-8799-06d55ca06fee {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1967.954185] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a069ce97-382c-45dd-9c60-af2d17a09f4e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1967.973831] env[62824]: DEBUG oslo_vmware.api [None req-86a2990d-e2a8-4265-a7f4-87f7818ab659 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for the task: (returnval){ [ 1967.973831] env[62824]: value = "task-2145882" [ 1967.973831] env[62824]: _type = "Task" [ 1967.973831] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1967.985700] env[62824]: DEBUG oslo_vmware.api [None req-86a2990d-e2a8-4265-a7f4-87f7818ab659 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2145882, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1968.011275] env[62824]: DEBUG oslo_concurrency.lockutils [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Acquiring lock "5ab8007d-df40-4840-ba8f-d84c3d00d195" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1968.011275] env[62824]: DEBUG oslo_concurrency.lockutils [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Lock "5ab8007d-df40-4840-ba8f-d84c3d00d195" acquired by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1968.022544] env[62824]: DEBUG oslo_concurrency.lockutils [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Lock "5ab8007d-df40-4840-ba8f-d84c3d00d195" "released" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: held 0.011s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1968.022883] env[62824]: DEBUG nova.compute.manager [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] [instance: 5f1237ff-5841-41f1-8d77-df9fabaa62b6] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1968.095669] env[62824]: DEBUG oslo_vmware.api [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]529f1c0a-491f-46e2-abbd-1d04ceaaffa4, 'name': SearchDatastore_Task, 'duration_secs': 0.016438} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1968.096531] env[62824]: DEBUG oslo_concurrency.lockutils [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1968.096531] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] [instance: 69bcac33-d6ac-4ee7-b674-58ab25389149] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1968.096531] env[62824]: DEBUG oslo_concurrency.lockutils [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1968.096713] env[62824]: DEBUG oslo_concurrency.lockutils [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1968.096839] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1968.097110] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-443b4119-bc67-4c2b-bbba-f9bb8b8a675f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1968.108729] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1968.109018] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1968.109822] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7a0c49b9-06c0-4e1d-8633-b007abf8621f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1968.116395] env[62824]: DEBUG oslo_vmware.api [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Waiting for the task: (returnval){ [ 1968.116395] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]527f7438-b40b-3b82-c0bb-8a04d5b8329c" [ 1968.116395] env[62824]: _type = "Task" [ 1968.116395] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1968.124594] env[62824]: DEBUG oslo_vmware.api [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]527f7438-b40b-3b82-c0bb-8a04d5b8329c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1968.132181] env[62824]: DEBUG nova.compute.manager [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 41f068d3-2c8d-46f0-8d84-78531d28b0dd] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1968.175477] env[62824]: DEBUG nova.network.neutron [req-6f628b28-9e3a-4662-a386-7ad95999d527 req-c6590fe5-7d5b-48a9-b185-d747c4281284 service nova] [instance: 69bcac33-d6ac-4ee7-b674-58ab25389149] Updated VIF entry in instance network info cache for port 08901969-e1cd-4351-9470-7c2ceabab5c4. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1968.176161] env[62824]: DEBUG nova.network.neutron [req-6f628b28-9e3a-4662-a386-7ad95999d527 req-c6590fe5-7d5b-48a9-b185-d747c4281284 service nova] [instance: 69bcac33-d6ac-4ee7-b674-58ab25389149] Updating instance_info_cache with network_info: [{"id": "08901969-e1cd-4351-9470-7c2ceabab5c4", "address": "fa:16:3e:e6:13:e2", "network": {"id": "79079501-33e5-4c53-b192-339eaa476dab", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-50257271-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c0b48e247f3e4a3b9f42cb780955279f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69f65356-c85e-4b7f-ad28-7c7b5e8cf50c", "external-id": "nsx-vlan-transportzone-281", "segmentation_id": 281, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap08901969-e1", "ovs_interfaceid": "08901969-e1cd-4351-9470-7c2ceabab5c4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1968.339729] env[62824]: DEBUG oslo_vmware.api [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2145881, 'name': PowerOnVM_Task, 'duration_secs': 0.555698} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1968.340015] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 487c2c9d-2cd2-4912-8613-e1bfac732c40] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1968.340225] env[62824]: INFO nova.compute.manager [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 487c2c9d-2cd2-4912-8613-e1bfac732c40] Took 7.25 seconds to spawn the instance on the hypervisor. [ 1968.340575] env[62824]: DEBUG nova.compute.manager [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 487c2c9d-2cd2-4912-8613-e1bfac732c40] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1968.341213] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-debb3e98-fa9f-48fd-825c-3a7aa37afd06 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1968.486449] env[62824]: DEBUG oslo_vmware.api [None req-86a2990d-e2a8-4265-a7f4-87f7818ab659 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2145882, 'name': ReconfigVM_Task, 'duration_secs': 0.512482} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1968.486724] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-86a2990d-e2a8-4265-a7f4-87f7818ab659 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: 6d80ec06-8559-4964-8577-a2512aa366ed] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1968.486985] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-72c9cacb-299e-4b1e-84d3-b5ea92b5d4bd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1968.493312] env[62824]: DEBUG oslo_vmware.api [None req-86a2990d-e2a8-4265-a7f4-87f7818ab659 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for the task: (returnval){ [ 1968.493312] env[62824]: value = "task-2145883" [ 1968.493312] env[62824]: _type = "Task" [ 1968.493312] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1968.503240] env[62824]: DEBUG oslo_vmware.api [None req-86a2990d-e2a8-4265-a7f4-87f7818ab659 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2145883, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1968.529328] env[62824]: DEBUG nova.compute.utils [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1968.534827] env[62824]: DEBUG nova.compute.manager [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] [instance: 5f1237ff-5841-41f1-8d77-df9fabaa62b6] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1968.535056] env[62824]: DEBUG nova.network.neutron [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] [instance: 5f1237ff-5841-41f1-8d77-df9fabaa62b6] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1968.618610] env[62824]: DEBUG nova.policy [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c21d871b41044a1b9e4c33c5b2265d1b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '87eb5358b5fe4919b0d657d3e0d008f0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 1968.642496] env[62824]: DEBUG oslo_vmware.api [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]527f7438-b40b-3b82-c0bb-8a04d5b8329c, 'name': SearchDatastore_Task, 'duration_secs': 0.030505} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1968.647814] env[62824]: DEBUG nova.compute.manager [req-60c1b825-93bf-459b-86ee-5ba32fcd624b req-4a49f26f-caca-428d-a3fd-e47184a448ca service nova] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Received event network-changed-1e74959f-64c2-4c37-85f8-fa8f81a30d60 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1968.647904] env[62824]: DEBUG nova.compute.manager [req-60c1b825-93bf-459b-86ee-5ba32fcd624b req-4a49f26f-caca-428d-a3fd-e47184a448ca service nova] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Refreshing instance network info cache due to event network-changed-1e74959f-64c2-4c37-85f8-fa8f81a30d60. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1968.648138] env[62824]: DEBUG oslo_concurrency.lockutils [req-60c1b825-93bf-459b-86ee-5ba32fcd624b req-4a49f26f-caca-428d-a3fd-e47184a448ca service nova] Acquiring lock "refresh_cache-59b5b883-4188-471c-8862-444f3ce08cb0" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1968.649063] env[62824]: DEBUG oslo_concurrency.lockutils [req-60c1b825-93bf-459b-86ee-5ba32fcd624b req-4a49f26f-caca-428d-a3fd-e47184a448ca service nova] Acquired lock "refresh_cache-59b5b883-4188-471c-8862-444f3ce08cb0" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1968.649063] env[62824]: DEBUG nova.network.neutron [req-60c1b825-93bf-459b-86ee-5ba32fcd624b req-4a49f26f-caca-428d-a3fd-e47184a448ca service nova] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Refreshing network info cache for port 1e74959f-64c2-4c37-85f8-fa8f81a30d60 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1968.652229] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e18f8810-b2f8-47a1-93cc-a3f2c308a377 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1968.659534] env[62824]: DEBUG oslo_vmware.api [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Waiting for the task: (returnval){ [ 1968.659534] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]521b9dfa-0ab6-053a-f769-34b623496dbc" [ 1968.659534] env[62824]: _type = "Task" [ 1968.659534] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1968.677197] env[62824]: DEBUG oslo_concurrency.lockutils [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1968.680220] env[62824]: DEBUG oslo_concurrency.lockutils [req-6f628b28-9e3a-4662-a386-7ad95999d527 req-c6590fe5-7d5b-48a9-b185-d747c4281284 service nova] Releasing lock "refresh_cache-69bcac33-d6ac-4ee7-b674-58ab25389149" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1968.680845] env[62824]: DEBUG oslo_vmware.api [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]521b9dfa-0ab6-053a-f769-34b623496dbc, 'name': SearchDatastore_Task, 'duration_secs': 0.013642} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1968.684131] env[62824]: DEBUG oslo_concurrency.lockutils [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1968.684417] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 69bcac33-d6ac-4ee7-b674-58ab25389149/69bcac33-d6ac-4ee7-b674-58ab25389149.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1968.684925] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-978e54c7-565d-4b40-9283-ccc2d2798a57 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1968.692995] env[62824]: DEBUG oslo_vmware.api [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Waiting for the task: (returnval){ [ 1968.692995] env[62824]: value = "task-2145884" [ 1968.692995] env[62824]: _type = "Task" [ 1968.692995] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1968.705144] env[62824]: DEBUG oslo_vmware.api [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Task: {'id': task-2145884, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1968.860274] env[62824]: INFO nova.compute.manager [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 487c2c9d-2cd2-4912-8613-e1bfac732c40] Took 43.50 seconds to build instance. [ 1969.012586] env[62824]: DEBUG oslo_vmware.api [None req-86a2990d-e2a8-4265-a7f4-87f7818ab659 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2145883, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1969.035439] env[62824]: DEBUG nova.compute.manager [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] [instance: 5f1237ff-5841-41f1-8d77-df9fabaa62b6] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1969.210143] env[62824]: DEBUG oslo_vmware.api [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Task: {'id': task-2145884, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1969.273385] env[62824]: DEBUG nova.network.neutron [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Port 02e90c28-7f36-4d83-84bb-fec194a04dab binding to destination host cpu-1 is already ACTIVE {{(pid=62824) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1969.273645] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Acquiring lock "refresh_cache-e158b5d4-c120-4e6c-89c4-7668e097926d" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1969.273796] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Acquired lock "refresh_cache-e158b5d4-c120-4e6c-89c4-7668e097926d" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1969.273960] env[62824]: DEBUG nova.network.neutron [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1969.297688] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96b04889-b67f-458c-9392-6aae0bc4ee03 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1969.306370] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c97d2cee-8fd9-4146-ba50-c1826af162dd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1969.336442] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-baaec130-841f-43d4-be35-ab7898cbfcf7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1969.344045] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42b063a5-1eb3-4417-881e-6d9d7a05f640 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1969.359835] env[62824]: DEBUG nova.compute.provider_tree [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1969.366450] env[62824]: DEBUG oslo_concurrency.lockutils [None req-731dafa5-dccc-4f30-a31c-ff8b030baa7d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "487c2c9d-2cd2-4912-8613-e1bfac732c40" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 94.849s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1969.509798] env[62824]: DEBUG oslo_vmware.api [None req-86a2990d-e2a8-4265-a7f4-87f7818ab659 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2145883, 'name': PowerOnVM_Task, 'duration_secs': 0.519898} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1969.509798] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-86a2990d-e2a8-4265-a7f4-87f7818ab659 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: 6d80ec06-8559-4964-8577-a2512aa366ed] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1969.509999] env[62824]: DEBUG nova.compute.manager [None req-86a2990d-e2a8-4265-a7f4-87f7818ab659 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: 6d80ec06-8559-4964-8577-a2512aa366ed] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1969.510837] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44ff66bc-6087-4edc-8d98-5881c32c4530 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1969.520096] env[62824]: DEBUG nova.network.neutron [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] [instance: 5f1237ff-5841-41f1-8d77-df9fabaa62b6] Successfully created port: 660b16ae-ebd9-43c6-a3dd-3338c7b75f72 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1969.576446] env[62824]: DEBUG nova.network.neutron [req-60c1b825-93bf-459b-86ee-5ba32fcd624b req-4a49f26f-caca-428d-a3fd-e47184a448ca service nova] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Updated VIF entry in instance network info cache for port 1e74959f-64c2-4c37-85f8-fa8f81a30d60. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1969.576446] env[62824]: DEBUG nova.network.neutron [req-60c1b825-93bf-459b-86ee-5ba32fcd624b req-4a49f26f-caca-428d-a3fd-e47184a448ca service nova] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Updating instance_info_cache with network_info: [{"id": "1e74959f-64c2-4c37-85f8-fa8f81a30d60", "address": "fa:16:3e:6d:26:78", "network": {"id": "ae1dfabd-223a-4d4c-a8c4-fe65a83dd9c0", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-797102932-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.148", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "102d2f886fd44e1da6780e48e36f1b62", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "517421c3-bea0-419c-ab0b-987815e5d160", "external-id": "nsx-vlan-transportzone-68", "segmentation_id": 68, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1e74959f-64", "ovs_interfaceid": "1e74959f-64c2-4c37-85f8-fa8f81a30d60", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1969.709045] env[62824]: DEBUG oslo_vmware.api [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Task: {'id': task-2145884, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.609281} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1969.709335] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 69bcac33-d6ac-4ee7-b674-58ab25389149/69bcac33-d6ac-4ee7-b674-58ab25389149.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1969.709617] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] [instance: 69bcac33-d6ac-4ee7-b674-58ab25389149] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1969.709805] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e808d46c-a26e-4591-81e2-1b0e7e797eac {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1969.716481] env[62824]: DEBUG oslo_vmware.api [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Waiting for the task: (returnval){ [ 1969.716481] env[62824]: value = "task-2145885" [ 1969.716481] env[62824]: _type = "Task" [ 1969.716481] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1969.730082] env[62824]: DEBUG oslo_vmware.api [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Task: {'id': task-2145885, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1969.747217] env[62824]: DEBUG oslo_concurrency.lockutils [None req-847fb409-5740-4b52-92bf-c077485b2771 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquiring lock "487c2c9d-2cd2-4912-8613-e1bfac732c40" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1969.747480] env[62824]: DEBUG oslo_concurrency.lockutils [None req-847fb409-5740-4b52-92bf-c077485b2771 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "487c2c9d-2cd2-4912-8613-e1bfac732c40" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1969.747688] env[62824]: DEBUG oslo_concurrency.lockutils [None req-847fb409-5740-4b52-92bf-c077485b2771 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquiring lock "487c2c9d-2cd2-4912-8613-e1bfac732c40-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1969.747866] env[62824]: DEBUG oslo_concurrency.lockutils [None req-847fb409-5740-4b52-92bf-c077485b2771 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "487c2c9d-2cd2-4912-8613-e1bfac732c40-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1969.750822] env[62824]: DEBUG oslo_concurrency.lockutils [None req-847fb409-5740-4b52-92bf-c077485b2771 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "487c2c9d-2cd2-4912-8613-e1bfac732c40-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1969.750822] env[62824]: INFO nova.compute.manager [None req-847fb409-5740-4b52-92bf-c077485b2771 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 487c2c9d-2cd2-4912-8613-e1bfac732c40] Terminating instance [ 1969.866143] env[62824]: DEBUG nova.compute.manager [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1969.869378] env[62824]: DEBUG nova.scheduler.client.report [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1970.049566] env[62824]: DEBUG nova.compute.manager [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] [instance: 5f1237ff-5841-41f1-8d77-df9fabaa62b6] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1970.077019] env[62824]: DEBUG nova.virt.hardware [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1970.077019] env[62824]: DEBUG nova.virt.hardware [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1970.077019] env[62824]: DEBUG nova.virt.hardware [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1970.077019] env[62824]: DEBUG nova.virt.hardware [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1970.077301] env[62824]: DEBUG nova.virt.hardware [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1970.077947] env[62824]: DEBUG nova.virt.hardware [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1970.078795] env[62824]: DEBUG nova.virt.hardware [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1970.079043] env[62824]: DEBUG nova.virt.hardware [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1970.079268] env[62824]: DEBUG nova.virt.hardware [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1970.079470] env[62824]: DEBUG nova.virt.hardware [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1970.080892] env[62824]: DEBUG nova.virt.hardware [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1970.080892] env[62824]: DEBUG oslo_concurrency.lockutils [req-60c1b825-93bf-459b-86ee-5ba32fcd624b req-4a49f26f-caca-428d-a3fd-e47184a448ca service nova] Releasing lock "refresh_cache-59b5b883-4188-471c-8862-444f3ce08cb0" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1970.081962] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7918caaf-293b-400a-bb28-8a69c38778c8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1970.091645] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85458d5c-1253-4baa-afdd-e1d362125b10 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1970.139302] env[62824]: DEBUG nova.network.neutron [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Updating instance_info_cache with network_info: [{"id": "02e90c28-7f36-4d83-84bb-fec194a04dab", "address": "fa:16:3e:53:1a:4c", "network": {"id": "b6403dc3-30ef-4c76-b521-380c179f345d", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.19", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "22c5c40c2e004babaa15896c89307303", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1002b79b-224e-41e3-a484-4245a767147a", "external-id": "nsx-vlan-transportzone-353", "segmentation_id": 353, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap02e90c28-7f", "ovs_interfaceid": "02e90c28-7f36-4d83-84bb-fec194a04dab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1970.231398] env[62824]: DEBUG oslo_vmware.api [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Task: {'id': task-2145885, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.214518} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1970.232183] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] [instance: 69bcac33-d6ac-4ee7-b674-58ab25389149] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1970.233330] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-184d4060-89d8-4cf4-b64c-4c98582be0f9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1970.254112] env[62824]: DEBUG nova.compute.manager [None req-847fb409-5740-4b52-92bf-c077485b2771 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 487c2c9d-2cd2-4912-8613-e1bfac732c40] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1970.254400] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-847fb409-5740-4b52-92bf-c077485b2771 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 487c2c9d-2cd2-4912-8613-e1bfac732c40] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1970.264976] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] [instance: 69bcac33-d6ac-4ee7-b674-58ab25389149] Reconfiguring VM instance instance-00000024 to attach disk [datastore2] 69bcac33-d6ac-4ee7-b674-58ab25389149/69bcac33-d6ac-4ee7-b674-58ab25389149.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1970.265794] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3536311d-7ed7-4e9c-91e8-257fa8f4ab33 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1970.268802] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9975c755-3ec1-4044-8c10-4f588c2139e0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1970.289997] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-847fb409-5740-4b52-92bf-c077485b2771 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 487c2c9d-2cd2-4912-8613-e1bfac732c40] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1970.292995] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0c0dae2a-ad24-4d8a-8581-bc12fa978d24 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1970.293399] env[62824]: DEBUG oslo_vmware.api [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Waiting for the task: (returnval){ [ 1970.293399] env[62824]: value = "task-2145886" [ 1970.293399] env[62824]: _type = "Task" [ 1970.293399] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1970.297751] env[62824]: DEBUG oslo_vmware.api [None req-847fb409-5740-4b52-92bf-c077485b2771 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 1970.297751] env[62824]: value = "task-2145887" [ 1970.297751] env[62824]: _type = "Task" [ 1970.297751] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1970.305102] env[62824]: DEBUG oslo_vmware.api [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Task: {'id': task-2145886, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1970.310953] env[62824]: DEBUG oslo_vmware.api [None req-847fb409-5740-4b52-92bf-c077485b2771 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2145887, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1970.377474] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.873s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1970.377954] env[62824]: DEBUG nova.compute.manager [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: dc3317e4-6958-4df6-8e39-c89b4a5d861c] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1970.383531] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 35.093s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1970.385604] env[62824]: INFO nova.compute.claims [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1970.404214] env[62824]: DEBUG oslo_concurrency.lockutils [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1970.642086] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Releasing lock "refresh_cache-e158b5d4-c120-4e6c-89c4-7668e097926d" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1970.806085] env[62824]: DEBUG oslo_vmware.api [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Task: {'id': task-2145886, 'name': ReconfigVM_Task, 'duration_secs': 0.386901} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1970.806865] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] [instance: 69bcac33-d6ac-4ee7-b674-58ab25389149] Reconfigured VM instance instance-00000024 to attach disk [datastore2] 69bcac33-d6ac-4ee7-b674-58ab25389149/69bcac33-d6ac-4ee7-b674-58ab25389149.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1970.808647] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cac17d03-7180-47ff-be03-de9b54918470 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1970.814183] env[62824]: DEBUG oslo_vmware.api [None req-847fb409-5740-4b52-92bf-c077485b2771 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2145887, 'name': PowerOffVM_Task, 'duration_secs': 0.202298} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1970.814183] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-847fb409-5740-4b52-92bf-c077485b2771 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 487c2c9d-2cd2-4912-8613-e1bfac732c40] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1970.814860] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-847fb409-5740-4b52-92bf-c077485b2771 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 487c2c9d-2cd2-4912-8613-e1bfac732c40] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1970.815473] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-855f1d2d-3a6c-40e1-8f42-2e1db9fd0b52 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1970.827169] env[62824]: DEBUG oslo_vmware.api [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Waiting for the task: (returnval){ [ 1970.827169] env[62824]: value = "task-2145888" [ 1970.827169] env[62824]: _type = "Task" [ 1970.827169] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1970.833071] env[62824]: DEBUG oslo_vmware.api [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Task: {'id': task-2145888, 'name': Rename_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1970.891322] env[62824]: DEBUG nova.compute.utils [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1970.900223] env[62824]: DEBUG nova.compute.manager [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: dc3317e4-6958-4df6-8e39-c89b4a5d861c] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1970.900223] env[62824]: DEBUG nova.network.neutron [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: dc3317e4-6958-4df6-8e39-c89b4a5d861c] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1970.901506] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-847fb409-5740-4b52-92bf-c077485b2771 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 487c2c9d-2cd2-4912-8613-e1bfac732c40] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1970.901711] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-847fb409-5740-4b52-92bf-c077485b2771 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 487c2c9d-2cd2-4912-8613-e1bfac732c40] Deleting contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1970.901881] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-847fb409-5740-4b52-92bf-c077485b2771 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Deleting the datastore file [datastore1] 487c2c9d-2cd2-4912-8613-e1bfac732c40 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1970.902378] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9d21ebc7-3b1c-4e9d-812b-cc5efe1bf969 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1970.908687] env[62824]: DEBUG oslo_vmware.api [None req-847fb409-5740-4b52-92bf-c077485b2771 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 1970.908687] env[62824]: value = "task-2145890" [ 1970.908687] env[62824]: _type = "Task" [ 1970.908687] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1970.919685] env[62824]: DEBUG oslo_vmware.api [None req-847fb409-5740-4b52-92bf-c077485b2771 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2145890, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1970.988521] env[62824]: DEBUG nova.policy [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c1978c29e9d54f73bf6fb9505ddf76d0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4cb963b19c1549ab8a11e5a77e4f202d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 1971.148422] env[62824]: DEBUG nova.compute.manager [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=62824) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:901}} [ 1971.148664] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1971.336372] env[62824]: DEBUG oslo_vmware.api [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Task: {'id': task-2145888, 'name': Rename_Task, 'duration_secs': 0.170301} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1971.336671] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] [instance: 69bcac33-d6ac-4ee7-b674-58ab25389149] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1971.336925] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-34ef0118-5277-4486-ae3b-aeb63dc852bb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1971.344707] env[62824]: DEBUG oslo_vmware.api [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Waiting for the task: (returnval){ [ 1971.344707] env[62824]: value = "task-2145891" [ 1971.344707] env[62824]: _type = "Task" [ 1971.344707] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1971.352927] env[62824]: DEBUG oslo_vmware.api [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Task: {'id': task-2145891, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1971.402778] env[62824]: DEBUG nova.compute.manager [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: dc3317e4-6958-4df6-8e39-c89b4a5d861c] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1971.428286] env[62824]: DEBUG oslo_vmware.api [None req-847fb409-5740-4b52-92bf-c077485b2771 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2145890, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.153352} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1971.428286] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-847fb409-5740-4b52-92bf-c077485b2771 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1971.428286] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-847fb409-5740-4b52-92bf-c077485b2771 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 487c2c9d-2cd2-4912-8613-e1bfac732c40] Deleted contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1971.428286] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-847fb409-5740-4b52-92bf-c077485b2771 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 487c2c9d-2cd2-4912-8613-e1bfac732c40] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1971.428286] env[62824]: INFO nova.compute.manager [None req-847fb409-5740-4b52-92bf-c077485b2771 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 487c2c9d-2cd2-4912-8613-e1bfac732c40] Took 1.17 seconds to destroy the instance on the hypervisor. [ 1971.428656] env[62824]: DEBUG oslo.service.loopingcall [None req-847fb409-5740-4b52-92bf-c077485b2771 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1971.428656] env[62824]: DEBUG nova.compute.manager [-] [instance: 487c2c9d-2cd2-4912-8613-e1bfac732c40] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1971.428656] env[62824]: DEBUG nova.network.neutron [-] [instance: 487c2c9d-2cd2-4912-8613-e1bfac732c40] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1971.448663] env[62824]: DEBUG nova.network.neutron [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: dc3317e4-6958-4df6-8e39-c89b4a5d861c] Successfully created port: ee0328da-cfb9-4f03-94fc-1acb84f8c553 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1971.858045] env[62824]: DEBUG oslo_vmware.api [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Task: {'id': task-2145891, 'name': PowerOnVM_Task, 'duration_secs': 0.470941} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1971.862554] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] [instance: 69bcac33-d6ac-4ee7-b674-58ab25389149] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1971.863037] env[62824]: INFO nova.compute.manager [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] [instance: 69bcac33-d6ac-4ee7-b674-58ab25389149] Took 7.17 seconds to spawn the instance on the hypervisor. [ 1971.863339] env[62824]: DEBUG nova.compute.manager [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] [instance: 69bcac33-d6ac-4ee7-b674-58ab25389149] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1971.868941] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e065981f-0f20-48df-b2c3-d17dc0db4f98 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1971.918152] env[62824]: DEBUG nova.compute.manager [req-fde62699-0b55-46cb-b826-f73b9e884c48 req-0af47d27-0175-49a0-b1e1-c4e9c0a5a954 service nova] [instance: 487c2c9d-2cd2-4912-8613-e1bfac732c40] Received event network-vif-deleted-bbbe4627-156d-462e-a430-0e4a08a4c2dc {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1971.918228] env[62824]: INFO nova.compute.manager [req-fde62699-0b55-46cb-b826-f73b9e884c48 req-0af47d27-0175-49a0-b1e1-c4e9c0a5a954 service nova] [instance: 487c2c9d-2cd2-4912-8613-e1bfac732c40] Neutron deleted interface bbbe4627-156d-462e-a430-0e4a08a4c2dc; detaching it from the instance and deleting it from the info cache [ 1971.918513] env[62824]: DEBUG nova.network.neutron [req-fde62699-0b55-46cb-b826-f73b9e884c48 req-0af47d27-0175-49a0-b1e1-c4e9c0a5a954 service nova] [instance: 487c2c9d-2cd2-4912-8613-e1bfac732c40] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1972.003916] env[62824]: DEBUG nova.network.neutron [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: dc3317e4-6958-4df6-8e39-c89b4a5d861c] Successfully created port: 4c03a7c0-a685-4fe9-9dce-635f6dda43b0 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1972.094260] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7abe114-0df0-4360-b169-c3ce51279c0e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1972.102490] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-215d67fc-b0f0-44ab-96bb-31bca69e8f91 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1972.137680] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e80e2817-4163-4333-9134-711a5384fd0f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1972.142849] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e34191fa-ec15-4945-89f8-85b7b11620c8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1972.157282] env[62824]: DEBUG nova.compute.provider_tree [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1972.261781] env[62824]: DEBUG nova.network.neutron [-] [instance: 487c2c9d-2cd2-4912-8613-e1bfac732c40] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1972.392674] env[62824]: INFO nova.compute.manager [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] [instance: 69bcac33-d6ac-4ee7-b674-58ab25389149] Took 42.56 seconds to build instance. [ 1972.414787] env[62824]: DEBUG nova.compute.manager [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: dc3317e4-6958-4df6-8e39-c89b4a5d861c] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1972.421169] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e1009236-285e-4293-9589-efbcc22f0cfc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1972.432306] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0633066-3f27-4286-8508-712637caaa58 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1972.460007] env[62824]: DEBUG nova.virt.hardware [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1972.460278] env[62824]: DEBUG nova.virt.hardware [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1972.460454] env[62824]: DEBUG nova.virt.hardware [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1972.460692] env[62824]: DEBUG nova.virt.hardware [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1972.460848] env[62824]: DEBUG nova.virt.hardware [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1972.460999] env[62824]: DEBUG nova.virt.hardware [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1972.461221] env[62824]: DEBUG nova.virt.hardware [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1972.461382] env[62824]: DEBUG nova.virt.hardware [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1972.462185] env[62824]: DEBUG nova.virt.hardware [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1972.462185] env[62824]: DEBUG nova.virt.hardware [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1972.462185] env[62824]: DEBUG nova.virt.hardware [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1972.462746] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bf5efdb-8da3-4094-94e7-1fbc6873355b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1972.476713] env[62824]: DEBUG nova.compute.manager [req-fde62699-0b55-46cb-b826-f73b9e884c48 req-0af47d27-0175-49a0-b1e1-c4e9c0a5a954 service nova] [instance: 487c2c9d-2cd2-4912-8613-e1bfac732c40] Detach interface failed, port_id=bbbe4627-156d-462e-a430-0e4a08a4c2dc, reason: Instance 487c2c9d-2cd2-4912-8613-e1bfac732c40 could not be found. {{(pid=62824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11482}} [ 1972.482293] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebffc5d4-1bd1-4761-a574-4b40c8dc5346 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1972.520265] env[62824]: INFO nova.compute.manager [None req-ed4db4f0-e595-4eab-9900-a79cf7eef489 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Rescuing [ 1972.520536] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ed4db4f0-e595-4eab-9900-a79cf7eef489 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Acquiring lock "refresh_cache-b3003c4b-ae5a-48df-8c12-a915a76253f4" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1972.520689] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ed4db4f0-e595-4eab-9900-a79cf7eef489 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Acquired lock "refresh_cache-b3003c4b-ae5a-48df-8c12-a915a76253f4" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1972.520857] env[62824]: DEBUG nova.network.neutron [None req-ed4db4f0-e595-4eab-9900-a79cf7eef489 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1972.661050] env[62824]: DEBUG nova.scheduler.client.report [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1972.765511] env[62824]: INFO nova.compute.manager [-] [instance: 487c2c9d-2cd2-4912-8613-e1bfac732c40] Took 1.34 seconds to deallocate network for instance. [ 1972.895216] env[62824]: DEBUG oslo_concurrency.lockutils [None req-badc1967-1ac2-4b85-bd96-337827f554fa tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Lock "69bcac33-d6ac-4ee7-b674-58ab25389149" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 94.932s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1973.166343] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.783s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1973.166882] env[62824]: DEBUG nova.compute.manager [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1973.169812] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8453c213-135b-48f6-a29c-da7b36504504 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 36.531s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1973.170054] env[62824]: DEBUG nova.objects.instance [None req-8453c213-135b-48f6-a29c-da7b36504504 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Lazy-loading 'resources' on Instance uuid ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1973.270868] env[62824]: DEBUG nova.network.neutron [None req-ed4db4f0-e595-4eab-9900-a79cf7eef489 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Updating instance_info_cache with network_info: [{"id": "4e0d51df-4419-4059-9dab-be44957e4f77", "address": "fa:16:3e:15:1a:78", "network": {"id": "a410f4f3-d072-4034-b797-8210e1a79093", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1482574391-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e10208040df34eb5adfe8dcbc76043d5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f4399275-8e92-4448-be9e-d4984e93e89c", "external-id": "nsx-vlan-transportzone-192", "segmentation_id": 192, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e0d51df-44", "ovs_interfaceid": "4e0d51df-4419-4059-9dab-be44957e4f77", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1973.272937] env[62824]: DEBUG oslo_concurrency.lockutils [None req-847fb409-5740-4b52-92bf-c077485b2771 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1973.402208] env[62824]: DEBUG nova.compute.manager [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] [instance: db018af0-7bdf-40c5-b129-fde15dfb8ab4] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1973.676023] env[62824]: DEBUG nova.compute.utils [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1973.676023] env[62824]: DEBUG nova.compute.manager [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1973.676737] env[62824]: DEBUG nova.network.neutron [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1973.769637] env[62824]: DEBUG nova.policy [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '31bf68236cfe4cfeb562a22258f003d9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1ab99ed209af4474a4c8a0a6e3844b2e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 1973.774449] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ed4db4f0-e595-4eab-9900-a79cf7eef489 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Releasing lock "refresh_cache-b3003c4b-ae5a-48df-8c12-a915a76253f4" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1973.919985] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0a53b27b-71f4-4fee-b778-9e0958eeed63 tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Acquiring lock "69bcac33-d6ac-4ee7-b674-58ab25389149" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1973.920295] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0a53b27b-71f4-4fee-b778-9e0958eeed63 tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Lock "69bcac33-d6ac-4ee7-b674-58ab25389149" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1973.920715] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0a53b27b-71f4-4fee-b778-9e0958eeed63 tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Acquiring lock "69bcac33-d6ac-4ee7-b674-58ab25389149-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1973.920715] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0a53b27b-71f4-4fee-b778-9e0958eeed63 tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Lock "69bcac33-d6ac-4ee7-b674-58ab25389149-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1973.920845] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0a53b27b-71f4-4fee-b778-9e0958eeed63 tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Lock "69bcac33-d6ac-4ee7-b674-58ab25389149-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1973.923262] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1973.923711] env[62824]: INFO nova.compute.manager [None req-0a53b27b-71f4-4fee-b778-9e0958eeed63 tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] [instance: 69bcac33-d6ac-4ee7-b674-58ab25389149] Terminating instance [ 1974.176378] env[62824]: DEBUG nova.network.neutron [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Successfully created port: bd32bec1-cb61-4462-8231-5f4a4ec2ed9d {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1974.182284] env[62824]: DEBUG nova.compute.manager [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1974.374562] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e76aadd1-c0b1-4be2-9cf2-a3b306d464cf {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1974.385041] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f17c9b85-927b-405d-8002-29b6684d1ef0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1974.416834] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c81b23b9-e983-46d7-92a9-7a783ef7f2d6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1974.424283] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64269641-f256-40e6-b64e-7624de83c40f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1974.429933] env[62824]: DEBUG nova.compute.manager [None req-0a53b27b-71f4-4fee-b778-9e0958eeed63 tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] [instance: 69bcac33-d6ac-4ee7-b674-58ab25389149] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1974.432833] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-0a53b27b-71f4-4fee-b778-9e0958eeed63 tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] [instance: 69bcac33-d6ac-4ee7-b674-58ab25389149] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1974.432833] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1ae8ce3-95e9-42ec-af04-255a87f85efd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1974.448537] env[62824]: DEBUG nova.compute.provider_tree [None req-8453c213-135b-48f6-a29c-da7b36504504 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1974.452963] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a53b27b-71f4-4fee-b778-9e0958eeed63 tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] [instance: 69bcac33-d6ac-4ee7-b674-58ab25389149] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1974.452963] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fa660407-1d12-4aae-b28c-b72e3340d5c0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1974.462264] env[62824]: DEBUG oslo_vmware.api [None req-0a53b27b-71f4-4fee-b778-9e0958eeed63 tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Waiting for the task: (returnval){ [ 1974.462264] env[62824]: value = "task-2145896" [ 1974.462264] env[62824]: _type = "Task" [ 1974.462264] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1974.467653] env[62824]: DEBUG oslo_vmware.api [None req-0a53b27b-71f4-4fee-b778-9e0958eeed63 tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Task: {'id': task-2145896, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1974.953390] env[62824]: DEBUG nova.scheduler.client.report [None req-8453c213-135b-48f6-a29c-da7b36504504 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1974.968055] env[62824]: DEBUG oslo_vmware.api [None req-0a53b27b-71f4-4fee-b778-9e0958eeed63 tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Task: {'id': task-2145896, 'name': PowerOffVM_Task, 'duration_secs': 0.200356} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1974.969322] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a53b27b-71f4-4fee-b778-9e0958eeed63 tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] [instance: 69bcac33-d6ac-4ee7-b674-58ab25389149] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1974.969528] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-0a53b27b-71f4-4fee-b778-9e0958eeed63 tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] [instance: 69bcac33-d6ac-4ee7-b674-58ab25389149] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1974.969986] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-27a4f4d6-1067-4be0-b692-618df3d0d114 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1975.201705] env[62824]: DEBUG nova.compute.manager [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1975.229904] env[62824]: DEBUG nova.virt.hardware [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1975.230195] env[62824]: DEBUG nova.virt.hardware [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1975.230346] env[62824]: DEBUG nova.virt.hardware [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1975.230577] env[62824]: DEBUG nova.virt.hardware [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1975.230775] env[62824]: DEBUG nova.virt.hardware [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1975.230933] env[62824]: DEBUG nova.virt.hardware [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1975.231157] env[62824]: DEBUG nova.virt.hardware [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1975.231402] env[62824]: DEBUG nova.virt.hardware [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1975.231632] env[62824]: DEBUG nova.virt.hardware [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1975.231802] env[62824]: DEBUG nova.virt.hardware [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1975.231978] env[62824]: DEBUG nova.virt.hardware [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1975.233229] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50760317-899e-4646-bdfc-ad3809291283 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1975.237054] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-0a53b27b-71f4-4fee-b778-9e0958eeed63 tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] [instance: 69bcac33-d6ac-4ee7-b674-58ab25389149] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1975.237257] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-0a53b27b-71f4-4fee-b778-9e0958eeed63 tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] [instance: 69bcac33-d6ac-4ee7-b674-58ab25389149] Deleting contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1975.237441] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-0a53b27b-71f4-4fee-b778-9e0958eeed63 tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Deleting the datastore file [datastore2] 69bcac33-d6ac-4ee7-b674-58ab25389149 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1975.238142] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9cbc9c44-9fea-4f76-a69b-a4277f1c8e1a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1975.243732] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba81ca69-1493-4d5f-869d-2844ec8f77c6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1975.248585] env[62824]: DEBUG oslo_vmware.api [None req-0a53b27b-71f4-4fee-b778-9e0958eeed63 tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Waiting for the task: (returnval){ [ 1975.248585] env[62824]: value = "task-2145898" [ 1975.248585] env[62824]: _type = "Task" [ 1975.248585] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1975.266065] env[62824]: DEBUG oslo_vmware.api [None req-0a53b27b-71f4-4fee-b778-9e0958eeed63 tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Task: {'id': task-2145898, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1975.320959] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed4db4f0-e595-4eab-9900-a79cf7eef489 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1975.321340] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a411d618-2a42-4536-8ff8-86b259feb7e2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1975.330257] env[62824]: DEBUG oslo_vmware.api [None req-ed4db4f0-e595-4eab-9900-a79cf7eef489 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for the task: (returnval){ [ 1975.330257] env[62824]: value = "task-2145899" [ 1975.330257] env[62824]: _type = "Task" [ 1975.330257] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1975.339695] env[62824]: DEBUG oslo_vmware.api [None req-ed4db4f0-e595-4eab-9900-a79cf7eef489 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2145899, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1975.459919] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8453c213-135b-48f6-a29c-da7b36504504 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.290s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1975.462495] env[62824]: DEBUG oslo_concurrency.lockutils [None req-995ba297-d41c-4960-b797-2a0283f98d18 tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 37.741s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1975.462744] env[62824]: DEBUG nova.objects.instance [None req-995ba297-d41c-4960-b797-2a0283f98d18 tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Lazy-loading 'resources' on Instance uuid b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1975.486794] env[62824]: INFO nova.scheduler.client.report [None req-8453c213-135b-48f6-a29c-da7b36504504 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Deleted allocations for instance ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5 [ 1975.765385] env[62824]: DEBUG oslo_vmware.api [None req-0a53b27b-71f4-4fee-b778-9e0958eeed63 tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Task: {'id': task-2145898, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.170956} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1975.765686] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-0a53b27b-71f4-4fee-b778-9e0958eeed63 tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1975.765883] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-0a53b27b-71f4-4fee-b778-9e0958eeed63 tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] [instance: 69bcac33-d6ac-4ee7-b674-58ab25389149] Deleted contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1975.766087] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-0a53b27b-71f4-4fee-b778-9e0958eeed63 tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] [instance: 69bcac33-d6ac-4ee7-b674-58ab25389149] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1975.766266] env[62824]: INFO nova.compute.manager [None req-0a53b27b-71f4-4fee-b778-9e0958eeed63 tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] [instance: 69bcac33-d6ac-4ee7-b674-58ab25389149] Took 1.34 seconds to destroy the instance on the hypervisor. [ 1975.766501] env[62824]: DEBUG oslo.service.loopingcall [None req-0a53b27b-71f4-4fee-b778-9e0958eeed63 tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1975.766686] env[62824]: DEBUG nova.compute.manager [-] [instance: 69bcac33-d6ac-4ee7-b674-58ab25389149] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1975.766781] env[62824]: DEBUG nova.network.neutron [-] [instance: 69bcac33-d6ac-4ee7-b674-58ab25389149] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1975.848009] env[62824]: DEBUG oslo_vmware.api [None req-ed4db4f0-e595-4eab-9900-a79cf7eef489 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2145899, 'name': PowerOffVM_Task, 'duration_secs': 0.277585} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1975.849028] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed4db4f0-e595-4eab-9900-a79cf7eef489 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1975.850219] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1462e83b-b33d-4165-bc54-7368d1658fe6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1975.890559] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9a687df-99d5-485f-b103-1be698743899 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1975.919770] env[62824]: DEBUG nova.network.neutron [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Successfully updated port: bd32bec1-cb61-4462-8231-5f4a4ec2ed9d {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1975.930457] env[62824]: DEBUG nova.network.neutron [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] [instance: 5f1237ff-5841-41f1-8d77-df9fabaa62b6] Successfully updated port: 660b16ae-ebd9-43c6-a3dd-3338c7b75f72 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1975.940412] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed4db4f0-e595-4eab-9900-a79cf7eef489 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1975.940412] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-14c1a47c-7bdd-45ee-b374-355565cc4ccf {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1975.950109] env[62824]: DEBUG oslo_vmware.api [None req-ed4db4f0-e595-4eab-9900-a79cf7eef489 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for the task: (returnval){ [ 1975.950109] env[62824]: value = "task-2145900" [ 1975.950109] env[62824]: _type = "Task" [ 1975.950109] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1975.959646] env[62824]: DEBUG oslo_vmware.api [None req-ed4db4f0-e595-4eab-9900-a79cf7eef489 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2145900, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1975.964491] env[62824]: DEBUG nova.compute.manager [req-cf8e5804-330c-4780-808f-0a3418761b20 req-913f5869-a2ed-4f47-ad02-d0fe41d26733 service nova] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Received event network-vif-plugged-bd32bec1-cb61-4462-8231-5f4a4ec2ed9d {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1975.964700] env[62824]: DEBUG oslo_concurrency.lockutils [req-cf8e5804-330c-4780-808f-0a3418761b20 req-913f5869-a2ed-4f47-ad02-d0fe41d26733 service nova] Acquiring lock "bbf1440b-8681-48c8-a178-9a83b925c695-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1975.964966] env[62824]: DEBUG oslo_concurrency.lockutils [req-cf8e5804-330c-4780-808f-0a3418761b20 req-913f5869-a2ed-4f47-ad02-d0fe41d26733 service nova] Lock "bbf1440b-8681-48c8-a178-9a83b925c695-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1975.965155] env[62824]: DEBUG oslo_concurrency.lockutils [req-cf8e5804-330c-4780-808f-0a3418761b20 req-913f5869-a2ed-4f47-ad02-d0fe41d26733 service nova] Lock "bbf1440b-8681-48c8-a178-9a83b925c695-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1975.965323] env[62824]: DEBUG nova.compute.manager [req-cf8e5804-330c-4780-808f-0a3418761b20 req-913f5869-a2ed-4f47-ad02-d0fe41d26733 service nova] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] No waiting events found dispatching network-vif-plugged-bd32bec1-cb61-4462-8231-5f4a4ec2ed9d {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1975.965485] env[62824]: WARNING nova.compute.manager [req-cf8e5804-330c-4780-808f-0a3418761b20 req-913f5869-a2ed-4f47-ad02-d0fe41d26733 service nova] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Received unexpected event network-vif-plugged-bd32bec1-cb61-4462-8231-5f4a4ec2ed9d for instance with vm_state building and task_state spawning. [ 1975.994124] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8453c213-135b-48f6-a29c-da7b36504504 tempest-AttachInterfacesV270Test-593584131 tempest-AttachInterfacesV270Test-593584131-project-member] Lock "ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 42.904s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1976.196311] env[62824]: DEBUG nova.compute.manager [req-b642f7a1-1819-436c-a5f7-9b3e4c5a0f20 req-dc6215fb-446b-4034-99fc-f53b6ea7ef81 service nova] [instance: dc3317e4-6958-4df6-8e39-c89b4a5d861c] Received event network-vif-plugged-ee0328da-cfb9-4f03-94fc-1acb84f8c553 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1976.196526] env[62824]: DEBUG oslo_concurrency.lockutils [req-b642f7a1-1819-436c-a5f7-9b3e4c5a0f20 req-dc6215fb-446b-4034-99fc-f53b6ea7ef81 service nova] Acquiring lock "dc3317e4-6958-4df6-8e39-c89b4a5d861c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1976.196725] env[62824]: DEBUG oslo_concurrency.lockutils [req-b642f7a1-1819-436c-a5f7-9b3e4c5a0f20 req-dc6215fb-446b-4034-99fc-f53b6ea7ef81 service nova] Lock "dc3317e4-6958-4df6-8e39-c89b4a5d861c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1976.196889] env[62824]: DEBUG oslo_concurrency.lockutils [req-b642f7a1-1819-436c-a5f7-9b3e4c5a0f20 req-dc6215fb-446b-4034-99fc-f53b6ea7ef81 service nova] Lock "dc3317e4-6958-4df6-8e39-c89b4a5d861c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1976.197064] env[62824]: DEBUG nova.compute.manager [req-b642f7a1-1819-436c-a5f7-9b3e4c5a0f20 req-dc6215fb-446b-4034-99fc-f53b6ea7ef81 service nova] [instance: dc3317e4-6958-4df6-8e39-c89b4a5d861c] No waiting events found dispatching network-vif-plugged-ee0328da-cfb9-4f03-94fc-1acb84f8c553 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1976.197230] env[62824]: WARNING nova.compute.manager [req-b642f7a1-1819-436c-a5f7-9b3e4c5a0f20 req-dc6215fb-446b-4034-99fc-f53b6ea7ef81 service nova] [instance: dc3317e4-6958-4df6-8e39-c89b4a5d861c] Received unexpected event network-vif-plugged-ee0328da-cfb9-4f03-94fc-1acb84f8c553 for instance with vm_state building and task_state spawning. [ 1976.323120] env[62824]: DEBUG nova.network.neutron [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: dc3317e4-6958-4df6-8e39-c89b4a5d861c] Successfully updated port: ee0328da-cfb9-4f03-94fc-1acb84f8c553 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1976.427232] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Acquiring lock "refresh_cache-bbf1440b-8681-48c8-a178-9a83b925c695" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1976.427387] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Acquired lock "refresh_cache-bbf1440b-8681-48c8-a178-9a83b925c695" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1976.427524] env[62824]: DEBUG nova.network.neutron [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1976.434482] env[62824]: DEBUG oslo_concurrency.lockutils [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Acquiring lock "refresh_cache-5f1237ff-5841-41f1-8d77-df9fabaa62b6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1976.434864] env[62824]: DEBUG oslo_concurrency.lockutils [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Acquired lock "refresh_cache-5f1237ff-5841-41f1-8d77-df9fabaa62b6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1976.434864] env[62824]: DEBUG nova.network.neutron [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] [instance: 5f1237ff-5841-41f1-8d77-df9fabaa62b6] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1976.463715] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed4db4f0-e595-4eab-9900-a79cf7eef489 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] VM already powered off {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1976.463887] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-ed4db4f0-e595-4eab-9900-a79cf7eef489 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1976.464147] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ed4db4f0-e595-4eab-9900-a79cf7eef489 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1976.464292] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ed4db4f0-e595-4eab-9900-a79cf7eef489 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1976.464469] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-ed4db4f0-e595-4eab-9900-a79cf7eef489 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1976.467663] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f9eb205e-12b5-46b9-bd24-d3b191696f3a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1976.479395] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-ed4db4f0-e595-4eab-9900-a79cf7eef489 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1976.479679] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-ed4db4f0-e595-4eab-9900-a79cf7eef489 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1976.480761] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3318db8e-5fd2-4327-bcef-5676096725d7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1976.485257] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19104352-e825-45a4-9b3d-95f4e33b131f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1976.495419] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a5412a5-f903-48f4-8932-f16e1e58c6e2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1976.499322] env[62824]: DEBUG oslo_vmware.api [None req-ed4db4f0-e595-4eab-9900-a79cf7eef489 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for the task: (returnval){ [ 1976.499322] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]529eb56c-b356-eb75-7a7c-bc94ae98d06b" [ 1976.499322] env[62824]: _type = "Task" [ 1976.499322] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1976.535214] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa9c9a51-f33e-4826-8324-46cbcb4e7098 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1976.538208] env[62824]: DEBUG oslo_vmware.api [None req-ed4db4f0-e595-4eab-9900-a79cf7eef489 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]529eb56c-b356-eb75-7a7c-bc94ae98d06b, 'name': SearchDatastore_Task, 'duration_secs': 0.013857} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1976.539546] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0834a708-0744-4e8b-815c-e43710d60cec {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1976.546535] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-645af9cb-146f-4199-a040-910d4c4041a3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1976.553206] env[62824]: DEBUG oslo_vmware.api [None req-ed4db4f0-e595-4eab-9900-a79cf7eef489 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for the task: (returnval){ [ 1976.553206] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52e12372-3968-cf16-b2f0-c6c534770aa9" [ 1976.553206] env[62824]: _type = "Task" [ 1976.553206] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1976.566106] env[62824]: DEBUG nova.compute.provider_tree [None req-995ba297-d41c-4960-b797-2a0283f98d18 tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1976.574112] env[62824]: DEBUG oslo_vmware.api [None req-ed4db4f0-e595-4eab-9900-a79cf7eef489 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52e12372-3968-cf16-b2f0-c6c534770aa9, 'name': SearchDatastore_Task, 'duration_secs': 0.010876} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1976.574437] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ed4db4f0-e595-4eab-9900-a79cf7eef489 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1976.574734] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-ed4db4f0-e595-4eab-9900-a79cf7eef489 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] b3003c4b-ae5a-48df-8c12-a915a76253f4/9e2a7d30-212d-4ab8-9606-c5c6d52629e8-rescue.vmdk. {{(pid=62824) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1976.575628] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a1461c3c-f4e1-48bf-ba95-39863e06bba1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1976.584754] env[62824]: DEBUG oslo_vmware.api [None req-ed4db4f0-e595-4eab-9900-a79cf7eef489 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for the task: (returnval){ [ 1976.584754] env[62824]: value = "task-2145902" [ 1976.584754] env[62824]: _type = "Task" [ 1976.584754] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1976.595544] env[62824]: DEBUG oslo_vmware.api [None req-ed4db4f0-e595-4eab-9900-a79cf7eef489 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2145902, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1976.976872] env[62824]: DEBUG nova.network.neutron [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] [instance: 5f1237ff-5841-41f1-8d77-df9fabaa62b6] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1977.019066] env[62824]: DEBUG nova.network.neutron [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1977.068052] env[62824]: DEBUG nova.network.neutron [-] [instance: 69bcac33-d6ac-4ee7-b674-58ab25389149] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1977.069865] env[62824]: DEBUG nova.scheduler.client.report [None req-995ba297-d41c-4960-b797-2a0283f98d18 tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1977.100114] env[62824]: DEBUG oslo_vmware.api [None req-ed4db4f0-e595-4eab-9900-a79cf7eef489 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2145902, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1977.145665] env[62824]: DEBUG nova.network.neutron [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] [instance: 5f1237ff-5841-41f1-8d77-df9fabaa62b6] Updating instance_info_cache with network_info: [{"id": "660b16ae-ebd9-43c6-a3dd-3338c7b75f72", "address": "fa:16:3e:fb:60:5b", "network": {"id": "1616481c-2d55-4ef3-a6c3-0619a48c8f68", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-209823666-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "87eb5358b5fe4919b0d657d3e0d008f0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d8383707-f093-40a7-a5ba-31b0e07cac45", "external-id": "cl2-zone-18", "segmentation_id": 18, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap660b16ae-eb", "ovs_interfaceid": "660b16ae-ebd9-43c6-a3dd-3338c7b75f72", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1977.379551] env[62824]: DEBUG nova.network.neutron [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Updating instance_info_cache with network_info: [{"id": "bd32bec1-cb61-4462-8231-5f4a4ec2ed9d", "address": "fa:16:3e:40:08:e3", "network": {"id": "b6403dc3-30ef-4c76-b521-380c179f345d", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.84", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "22c5c40c2e004babaa15896c89307303", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1002b79b-224e-41e3-a484-4245a767147a", "external-id": "nsx-vlan-transportzone-353", "segmentation_id": 353, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbd32bec1-cb", "ovs_interfaceid": "bd32bec1-cb61-4462-8231-5f4a4ec2ed9d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1977.575324] env[62824]: DEBUG oslo_concurrency.lockutils [None req-995ba297-d41c-4960-b797-2a0283f98d18 tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.113s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1977.578502] env[62824]: INFO nova.compute.manager [-] [instance: 69bcac33-d6ac-4ee7-b674-58ab25389149] Took 1.81 seconds to deallocate network for instance. [ 1977.579374] env[62824]: DEBUG oslo_concurrency.lockutils [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 38.085s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1977.581095] env[62824]: INFO nova.compute.claims [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1977.599887] env[62824]: DEBUG oslo_vmware.api [None req-ed4db4f0-e595-4eab-9900-a79cf7eef489 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2145902, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1977.611652] env[62824]: INFO nova.scheduler.client.report [None req-995ba297-d41c-4960-b797-2a0283f98d18 tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Deleted allocations for instance b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6 [ 1977.647817] env[62824]: DEBUG oslo_concurrency.lockutils [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Releasing lock "refresh_cache-5f1237ff-5841-41f1-8d77-df9fabaa62b6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1977.648230] env[62824]: DEBUG nova.compute.manager [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] [instance: 5f1237ff-5841-41f1-8d77-df9fabaa62b6] Instance network_info: |[{"id": "660b16ae-ebd9-43c6-a3dd-3338c7b75f72", "address": "fa:16:3e:fb:60:5b", "network": {"id": "1616481c-2d55-4ef3-a6c3-0619a48c8f68", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-209823666-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "87eb5358b5fe4919b0d657d3e0d008f0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d8383707-f093-40a7-a5ba-31b0e07cac45", "external-id": "cl2-zone-18", "segmentation_id": 18, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap660b16ae-eb", "ovs_interfaceid": "660b16ae-ebd9-43c6-a3dd-3338c7b75f72", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1977.648997] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] [instance: 5f1237ff-5841-41f1-8d77-df9fabaa62b6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fb:60:5b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd8383707-f093-40a7-a5ba-31b0e07cac45', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '660b16ae-ebd9-43c6-a3dd-3338c7b75f72', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1977.658541] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Creating folder: Project (87eb5358b5fe4919b0d657d3e0d008f0). Parent ref: group-v438503. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1977.659624] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5fbc7f3f-39eb-4e26-b6f7-f4884fbc2e33 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1977.676267] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Created folder: Project (87eb5358b5fe4919b0d657d3e0d008f0) in parent group-v438503. [ 1977.677047] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Creating folder: Instances. Parent ref: group-v438585. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1977.680774] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-07be9188-f8ab-4ea0-a856-82b6cbe26d01 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1977.694676] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Created folder: Instances in parent group-v438585. [ 1977.695132] env[62824]: DEBUG oslo.service.loopingcall [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1977.695366] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5f1237ff-5841-41f1-8d77-df9fabaa62b6] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1977.695624] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1bd47c07-0f38-4c45-86dc-c565c297d5b7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1977.725914] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1977.725914] env[62824]: value = "task-2145905" [ 1977.725914] env[62824]: _type = "Task" [ 1977.725914] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1977.737179] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145905, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1977.883328] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Releasing lock "refresh_cache-bbf1440b-8681-48c8-a178-9a83b925c695" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1977.883670] env[62824]: DEBUG nova.compute.manager [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Instance network_info: |[{"id": "bd32bec1-cb61-4462-8231-5f4a4ec2ed9d", "address": "fa:16:3e:40:08:e3", "network": {"id": "b6403dc3-30ef-4c76-b521-380c179f345d", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.84", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "22c5c40c2e004babaa15896c89307303", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1002b79b-224e-41e3-a484-4245a767147a", "external-id": "nsx-vlan-transportzone-353", "segmentation_id": 353, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbd32bec1-cb", "ovs_interfaceid": "bd32bec1-cb61-4462-8231-5f4a4ec2ed9d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1977.884368] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:40:08:e3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1002b79b-224e-41e3-a484-4245a767147a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bd32bec1-cb61-4462-8231-5f4a4ec2ed9d', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1977.893180] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Creating folder: Project (1ab99ed209af4474a4c8a0a6e3844b2e). Parent ref: group-v438503. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1977.893519] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e8fefe68-b784-4c11-a011-418e4201f621 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1977.909171] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Created folder: Project (1ab99ed209af4474a4c8a0a6e3844b2e) in parent group-v438503. [ 1977.909380] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Creating folder: Instances. Parent ref: group-v438588. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1977.909664] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5e757242-cb92-4658-87e0-020375109629 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1977.924733] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Created folder: Instances in parent group-v438588. [ 1977.925123] env[62824]: DEBUG oslo.service.loopingcall [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1977.925344] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1977.925791] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b24a3fc7-98e0-4f5f-892e-760c6d1955c4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1977.948067] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1977.948067] env[62824]: value = "task-2145908" [ 1977.948067] env[62824]: _type = "Task" [ 1977.948067] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1977.962373] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145908, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1978.095414] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0a53b27b-71f4-4fee-b778-9e0958eeed63 tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1978.102731] env[62824]: DEBUG oslo_vmware.api [None req-ed4db4f0-e595-4eab-9900-a79cf7eef489 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2145902, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.457644} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1978.102892] env[62824]: INFO nova.virt.vmwareapi.ds_util [None req-ed4db4f0-e595-4eab-9900-a79cf7eef489 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] b3003c4b-ae5a-48df-8c12-a915a76253f4/9e2a7d30-212d-4ab8-9606-c5c6d52629e8-rescue.vmdk. [ 1978.103751] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13edd653-f3f7-4994-b542-e7f06cc75469 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1978.138348] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-ed4db4f0-e595-4eab-9900-a79cf7eef489 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Reconfiguring VM instance instance-00000020 to attach disk [datastore2] b3003c4b-ae5a-48df-8c12-a915a76253f4/9e2a7d30-212d-4ab8-9606-c5c6d52629e8-rescue.vmdk or device None with type thin {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1978.139290] env[62824]: DEBUG oslo_concurrency.lockutils [None req-995ba297-d41c-4960-b797-2a0283f98d18 tempest-ServerExternalEventsTest-473723311 tempest-ServerExternalEventsTest-473723311-project-member] Lock "b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 43.927s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1978.140771] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-da60f6df-271b-42a3-93a3-8812ade271c9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1978.163977] env[62824]: DEBUG oslo_vmware.api [None req-ed4db4f0-e595-4eab-9900-a79cf7eef489 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for the task: (returnval){ [ 1978.163977] env[62824]: value = "task-2145909" [ 1978.163977] env[62824]: _type = "Task" [ 1978.163977] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1978.184361] env[62824]: DEBUG oslo_vmware.api [None req-ed4db4f0-e595-4eab-9900-a79cf7eef489 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2145909, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1978.241125] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145905, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1978.462982] env[62824]: DEBUG nova.compute.manager [req-1b02348e-8fc8-4254-a430-c23898d6e296 req-f1b61601-105e-4a06-b73b-db84a9f11010 service nova] [instance: 5f1237ff-5841-41f1-8d77-df9fabaa62b6] Received event network-vif-plugged-660b16ae-ebd9-43c6-a3dd-3338c7b75f72 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1978.463523] env[62824]: DEBUG oslo_concurrency.lockutils [req-1b02348e-8fc8-4254-a430-c23898d6e296 req-f1b61601-105e-4a06-b73b-db84a9f11010 service nova] Acquiring lock "5f1237ff-5841-41f1-8d77-df9fabaa62b6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1978.463997] env[62824]: DEBUG oslo_concurrency.lockutils [req-1b02348e-8fc8-4254-a430-c23898d6e296 req-f1b61601-105e-4a06-b73b-db84a9f11010 service nova] Lock "5f1237ff-5841-41f1-8d77-df9fabaa62b6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1978.464528] env[62824]: DEBUG oslo_concurrency.lockutils [req-1b02348e-8fc8-4254-a430-c23898d6e296 req-f1b61601-105e-4a06-b73b-db84a9f11010 service nova] Lock "5f1237ff-5841-41f1-8d77-df9fabaa62b6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1978.465100] env[62824]: DEBUG nova.compute.manager [req-1b02348e-8fc8-4254-a430-c23898d6e296 req-f1b61601-105e-4a06-b73b-db84a9f11010 service nova] [instance: 5f1237ff-5841-41f1-8d77-df9fabaa62b6] No waiting events found dispatching network-vif-plugged-660b16ae-ebd9-43c6-a3dd-3338c7b75f72 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1978.465513] env[62824]: WARNING nova.compute.manager [req-1b02348e-8fc8-4254-a430-c23898d6e296 req-f1b61601-105e-4a06-b73b-db84a9f11010 service nova] [instance: 5f1237ff-5841-41f1-8d77-df9fabaa62b6] Received unexpected event network-vif-plugged-660b16ae-ebd9-43c6-a3dd-3338c7b75f72 for instance with vm_state building and task_state spawning. [ 1978.465941] env[62824]: DEBUG nova.compute.manager [req-1b02348e-8fc8-4254-a430-c23898d6e296 req-f1b61601-105e-4a06-b73b-db84a9f11010 service nova] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Received event network-changed-bd32bec1-cb61-4462-8231-5f4a4ec2ed9d {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1978.466382] env[62824]: DEBUG nova.compute.manager [req-1b02348e-8fc8-4254-a430-c23898d6e296 req-f1b61601-105e-4a06-b73b-db84a9f11010 service nova] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Refreshing instance network info cache due to event network-changed-bd32bec1-cb61-4462-8231-5f4a4ec2ed9d. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1978.466857] env[62824]: DEBUG oslo_concurrency.lockutils [req-1b02348e-8fc8-4254-a430-c23898d6e296 req-f1b61601-105e-4a06-b73b-db84a9f11010 service nova] Acquiring lock "refresh_cache-bbf1440b-8681-48c8-a178-9a83b925c695" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1978.467256] env[62824]: DEBUG oslo_concurrency.lockutils [req-1b02348e-8fc8-4254-a430-c23898d6e296 req-f1b61601-105e-4a06-b73b-db84a9f11010 service nova] Acquired lock "refresh_cache-bbf1440b-8681-48c8-a178-9a83b925c695" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1978.468426] env[62824]: DEBUG nova.network.neutron [req-1b02348e-8fc8-4254-a430-c23898d6e296 req-f1b61601-105e-4a06-b73b-db84a9f11010 service nova] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Refreshing network info cache for port bd32bec1-cb61-4462-8231-5f4a4ec2ed9d {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1978.479334] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145908, 'name': CreateVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1978.522062] env[62824]: DEBUG nova.compute.manager [req-760287b4-aa60-4e81-a817-a1f1b8a539da req-cb426890-4a69-483f-8353-5ddf2d972eb1 service nova] [instance: dc3317e4-6958-4df6-8e39-c89b4a5d861c] Received event network-changed-ee0328da-cfb9-4f03-94fc-1acb84f8c553 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1978.522282] env[62824]: DEBUG nova.compute.manager [req-760287b4-aa60-4e81-a817-a1f1b8a539da req-cb426890-4a69-483f-8353-5ddf2d972eb1 service nova] [instance: dc3317e4-6958-4df6-8e39-c89b4a5d861c] Refreshing instance network info cache due to event network-changed-ee0328da-cfb9-4f03-94fc-1acb84f8c553. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1978.522493] env[62824]: DEBUG oslo_concurrency.lockutils [req-760287b4-aa60-4e81-a817-a1f1b8a539da req-cb426890-4a69-483f-8353-5ddf2d972eb1 service nova] Acquiring lock "refresh_cache-dc3317e4-6958-4df6-8e39-c89b4a5d861c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1978.522636] env[62824]: DEBUG oslo_concurrency.lockutils [req-760287b4-aa60-4e81-a817-a1f1b8a539da req-cb426890-4a69-483f-8353-5ddf2d972eb1 service nova] Acquired lock "refresh_cache-dc3317e4-6958-4df6-8e39-c89b4a5d861c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1978.522831] env[62824]: DEBUG nova.network.neutron [req-760287b4-aa60-4e81-a817-a1f1b8a539da req-cb426890-4a69-483f-8353-5ddf2d972eb1 service nova] [instance: dc3317e4-6958-4df6-8e39-c89b4a5d861c] Refreshing network info cache for port ee0328da-cfb9-4f03-94fc-1acb84f8c553 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1978.679309] env[62824]: DEBUG oslo_vmware.api [None req-ed4db4f0-e595-4eab-9900-a79cf7eef489 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2145909, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1978.738255] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145905, 'name': CreateVM_Task, 'duration_secs': 0.644762} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1978.740951] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5f1237ff-5841-41f1-8d77-df9fabaa62b6] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1978.742421] env[62824]: DEBUG oslo_concurrency.lockutils [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1978.742421] env[62824]: DEBUG oslo_concurrency.lockutils [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1978.743375] env[62824]: DEBUG oslo_concurrency.lockutils [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1978.743791] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5e5b3da1-db60-4043-a3b9-bb0092766c54 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1978.750333] env[62824]: DEBUG oslo_vmware.api [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Waiting for the task: (returnval){ [ 1978.750333] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52acb357-1468-37f6-ace1-598fcadd36ab" [ 1978.750333] env[62824]: _type = "Task" [ 1978.750333] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1978.765584] env[62824]: DEBUG oslo_vmware.api [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52acb357-1468-37f6-ace1-598fcadd36ab, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1978.964385] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145908, 'name': CreateVM_Task, 'duration_secs': 0.594706} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1978.964575] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1978.965766] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1979.075516] env[62824]: DEBUG nova.network.neutron [req-760287b4-aa60-4e81-a817-a1f1b8a539da req-cb426890-4a69-483f-8353-5ddf2d972eb1 service nova] [instance: dc3317e4-6958-4df6-8e39-c89b4a5d861c] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1979.182748] env[62824]: DEBUG oslo_vmware.api [None req-ed4db4f0-e595-4eab-9900-a79cf7eef489 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2145909, 'name': ReconfigVM_Task, 'duration_secs': 0.802998} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1979.184777] env[62824]: DEBUG nova.network.neutron [req-760287b4-aa60-4e81-a817-a1f1b8a539da req-cb426890-4a69-483f-8353-5ddf2d972eb1 service nova] [instance: dc3317e4-6958-4df6-8e39-c89b4a5d861c] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1979.187239] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-ed4db4f0-e595-4eab-9900-a79cf7eef489 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Reconfigured VM instance instance-00000020 to attach disk [datastore2] b3003c4b-ae5a-48df-8c12-a915a76253f4/9e2a7d30-212d-4ab8-9606-c5c6d52629e8-rescue.vmdk or device None with type thin {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1979.189315] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f1fc085-106b-4bd7-8d02-e87151749082 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1979.221267] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7ce2238f-752a-404f-b84b-5263fb07dc25 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1979.240355] env[62824]: DEBUG oslo_vmware.api [None req-ed4db4f0-e595-4eab-9900-a79cf7eef489 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for the task: (returnval){ [ 1979.240355] env[62824]: value = "task-2145911" [ 1979.240355] env[62824]: _type = "Task" [ 1979.240355] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1979.253557] env[62824]: DEBUG oslo_vmware.api [None req-ed4db4f0-e595-4eab-9900-a79cf7eef489 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2145911, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1979.267379] env[62824]: DEBUG oslo_vmware.api [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52acb357-1468-37f6-ace1-598fcadd36ab, 'name': SearchDatastore_Task, 'duration_secs': 0.028664} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1979.267788] env[62824]: DEBUG oslo_concurrency.lockutils [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1979.267930] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] [instance: 5f1237ff-5841-41f1-8d77-df9fabaa62b6] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1979.268331] env[62824]: DEBUG oslo_concurrency.lockutils [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1979.268398] env[62824]: DEBUG oslo_concurrency.lockutils [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1979.268528] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1979.268826] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1979.269152] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1979.269391] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-564f7d57-690a-4171-8af3-fa6167af854f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1979.272146] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-61f0562c-c12f-497f-a403-71bb56cfa3a5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1979.274247] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7951a570-db20-49ec-9bd0-6b2a83099846 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1979.282935] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b58e6911-bb44-4b11-9b59-e0da32dbe632 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1979.286110] env[62824]: DEBUG oslo_vmware.api [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Waiting for the task: (returnval){ [ 1979.286110] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5223f4e8-7dc0-6f93-769c-4e7d8edf1075" [ 1979.286110] env[62824]: _type = "Task" [ 1979.286110] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1979.287170] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1979.287342] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1979.291166] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dae055c0-3062-41b9-ba7d-4f3dc05f32aa {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1979.321294] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a1f1114-31a6-4043-9882-b84c59fed62b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1979.328409] env[62824]: DEBUG oslo_vmware.api [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5223f4e8-7dc0-6f93-769c-4e7d8edf1075, 'name': SearchDatastore_Task, 'duration_secs': 0.01376} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1979.328757] env[62824]: DEBUG oslo_vmware.api [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Waiting for the task: (returnval){ [ 1979.328757] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52238262-02c6-1911-75a1-ead9f6ddce1b" [ 1979.328757] env[62824]: _type = "Task" [ 1979.328757] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1979.329416] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1979.329672] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1979.329880] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1979.338599] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0c592de-fd90-4ba2-8c36-caec0c3c16bf {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1979.346174] env[62824]: DEBUG oslo_vmware.api [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52238262-02c6-1911-75a1-ead9f6ddce1b, 'name': SearchDatastore_Task, 'duration_secs': 0.014371} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1979.347486] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-481d7b7c-5156-454a-8e55-df15dbeae576 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1979.357424] env[62824]: DEBUG nova.compute.provider_tree [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1979.363952] env[62824]: DEBUG oslo_vmware.api [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Waiting for the task: (returnval){ [ 1979.363952] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52994daf-f917-eb61-9236-35956204ac84" [ 1979.363952] env[62824]: _type = "Task" [ 1979.363952] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1979.376985] env[62824]: DEBUG oslo_vmware.api [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52994daf-f917-eb61-9236-35956204ac84, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1979.693060] env[62824]: DEBUG oslo_concurrency.lockutils [req-760287b4-aa60-4e81-a817-a1f1b8a539da req-cb426890-4a69-483f-8353-5ddf2d972eb1 service nova] Releasing lock "refresh_cache-dc3317e4-6958-4df6-8e39-c89b4a5d861c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1979.693349] env[62824]: DEBUG nova.compute.manager [req-760287b4-aa60-4e81-a817-a1f1b8a539da req-cb426890-4a69-483f-8353-5ddf2d972eb1 service nova] [instance: 69bcac33-d6ac-4ee7-b674-58ab25389149] Received event network-vif-deleted-08901969-e1cd-4351-9470-7c2ceabab5c4 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1979.752093] env[62824]: DEBUG oslo_vmware.api [None req-ed4db4f0-e595-4eab-9900-a79cf7eef489 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2145911, 'name': ReconfigVM_Task, 'duration_secs': 0.170817} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1979.752432] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed4db4f0-e595-4eab-9900-a79cf7eef489 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1979.752855] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a1618ad3-bd2e-438b-b192-ee6c2e70d3ac {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1979.761596] env[62824]: DEBUG oslo_vmware.api [None req-ed4db4f0-e595-4eab-9900-a79cf7eef489 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for the task: (returnval){ [ 1979.761596] env[62824]: value = "task-2145912" [ 1979.761596] env[62824]: _type = "Task" [ 1979.761596] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1979.767489] env[62824]: DEBUG nova.network.neutron [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: dc3317e4-6958-4df6-8e39-c89b4a5d861c] Successfully updated port: 4c03a7c0-a685-4fe9-9dce-635f6dda43b0 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1979.772152] env[62824]: DEBUG oslo_vmware.api [None req-ed4db4f0-e595-4eab-9900-a79cf7eef489 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2145912, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1979.828888] env[62824]: DEBUG nova.network.neutron [req-1b02348e-8fc8-4254-a430-c23898d6e296 req-f1b61601-105e-4a06-b73b-db84a9f11010 service nova] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Updated VIF entry in instance network info cache for port bd32bec1-cb61-4462-8231-5f4a4ec2ed9d. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1979.829275] env[62824]: DEBUG nova.network.neutron [req-1b02348e-8fc8-4254-a430-c23898d6e296 req-f1b61601-105e-4a06-b73b-db84a9f11010 service nova] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Updating instance_info_cache with network_info: [{"id": "bd32bec1-cb61-4462-8231-5f4a4ec2ed9d", "address": "fa:16:3e:40:08:e3", "network": {"id": "b6403dc3-30ef-4c76-b521-380c179f345d", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.84", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "22c5c40c2e004babaa15896c89307303", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1002b79b-224e-41e3-a484-4245a767147a", "external-id": "nsx-vlan-transportzone-353", "segmentation_id": 353, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbd32bec1-cb", "ovs_interfaceid": "bd32bec1-cb61-4462-8231-5f4a4ec2ed9d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1979.860875] env[62824]: DEBUG nova.scheduler.client.report [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1979.876394] env[62824]: DEBUG oslo_vmware.api [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52994daf-f917-eb61-9236-35956204ac84, 'name': SearchDatastore_Task, 'duration_secs': 0.039386} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1979.877059] env[62824]: DEBUG oslo_concurrency.lockutils [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1979.877400] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 5f1237ff-5841-41f1-8d77-df9fabaa62b6/5f1237ff-5841-41f1-8d77-df9fabaa62b6.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1979.877972] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1979.878172] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1979.878447] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6c19e395-30bf-4a7b-b284-f9620e1f80ea {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1979.882428] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4893939a-fb9f-4fcb-bbef-7da8eaf363db {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1979.893020] env[62824]: DEBUG oslo_vmware.api [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Waiting for the task: (returnval){ [ 1979.893020] env[62824]: value = "task-2145913" [ 1979.893020] env[62824]: _type = "Task" [ 1979.893020] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1979.893849] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1979.894218] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1979.899420] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3d767ef7-0172-4046-b13b-883c4450dc46 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1979.910892] env[62824]: DEBUG oslo_vmware.api [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Task: {'id': task-2145913, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1979.913461] env[62824]: DEBUG oslo_vmware.api [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Waiting for the task: (returnval){ [ 1979.913461] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]529a437c-44b1-5854-e661-0d2aa514e48e" [ 1979.913461] env[62824]: _type = "Task" [ 1979.913461] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1979.927999] env[62824]: DEBUG oslo_vmware.api [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]529a437c-44b1-5854-e661-0d2aa514e48e, 'name': SearchDatastore_Task, 'duration_secs': 0.010723} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1979.927999] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0f2213d4-fadb-4e7b-b7fb-ccc656f652cc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1979.933717] env[62824]: DEBUG oslo_vmware.api [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Waiting for the task: (returnval){ [ 1979.933717] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5276e5f5-3561-acb9-9a5e-4bda5e2fd33c" [ 1979.933717] env[62824]: _type = "Task" [ 1979.933717] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1979.945413] env[62824]: DEBUG oslo_vmware.api [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5276e5f5-3561-acb9-9a5e-4bda5e2fd33c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1980.276954] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Acquiring lock "refresh_cache-dc3317e4-6958-4df6-8e39-c89b4a5d861c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1980.277524] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Acquired lock "refresh_cache-dc3317e4-6958-4df6-8e39-c89b4a5d861c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1980.277759] env[62824]: DEBUG nova.network.neutron [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: dc3317e4-6958-4df6-8e39-c89b4a5d861c] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1980.279307] env[62824]: DEBUG oslo_vmware.api [None req-ed4db4f0-e595-4eab-9900-a79cf7eef489 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2145912, 'name': PowerOnVM_Task, 'duration_secs': 0.44511} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1980.280049] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed4db4f0-e595-4eab-9900-a79cf7eef489 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1980.284766] env[62824]: DEBUG nova.compute.manager [None req-ed4db4f0-e595-4eab-9900-a79cf7eef489 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1980.285770] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08df38b9-9fba-49ed-b44b-ea52532dbe1e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1980.333094] env[62824]: DEBUG oslo_concurrency.lockutils [req-1b02348e-8fc8-4254-a430-c23898d6e296 req-f1b61601-105e-4a06-b73b-db84a9f11010 service nova] Releasing lock "refresh_cache-bbf1440b-8681-48c8-a178-9a83b925c695" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1980.333316] env[62824]: DEBUG nova.compute.manager [req-1b02348e-8fc8-4254-a430-c23898d6e296 req-f1b61601-105e-4a06-b73b-db84a9f11010 service nova] [instance: 5f1237ff-5841-41f1-8d77-df9fabaa62b6] Received event network-changed-660b16ae-ebd9-43c6-a3dd-3338c7b75f72 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1980.333493] env[62824]: DEBUG nova.compute.manager [req-1b02348e-8fc8-4254-a430-c23898d6e296 req-f1b61601-105e-4a06-b73b-db84a9f11010 service nova] [instance: 5f1237ff-5841-41f1-8d77-df9fabaa62b6] Refreshing instance network info cache due to event network-changed-660b16ae-ebd9-43c6-a3dd-3338c7b75f72. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1980.333713] env[62824]: DEBUG oslo_concurrency.lockutils [req-1b02348e-8fc8-4254-a430-c23898d6e296 req-f1b61601-105e-4a06-b73b-db84a9f11010 service nova] Acquiring lock "refresh_cache-5f1237ff-5841-41f1-8d77-df9fabaa62b6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1980.333861] env[62824]: DEBUG oslo_concurrency.lockutils [req-1b02348e-8fc8-4254-a430-c23898d6e296 req-f1b61601-105e-4a06-b73b-db84a9f11010 service nova] Acquired lock "refresh_cache-5f1237ff-5841-41f1-8d77-df9fabaa62b6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1980.334875] env[62824]: DEBUG nova.network.neutron [req-1b02348e-8fc8-4254-a430-c23898d6e296 req-f1b61601-105e-4a06-b73b-db84a9f11010 service nova] [instance: 5f1237ff-5841-41f1-8d77-df9fabaa62b6] Refreshing network info cache for port 660b16ae-ebd9-43c6-a3dd-3338c7b75f72 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1980.371376] env[62824]: DEBUG oslo_concurrency.lockutils [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.792s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1980.371976] env[62824]: DEBUG nova.compute.manager [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1980.375082] env[62824]: DEBUG oslo_concurrency.lockutils [None req-89ab4335-c79f-4e08-ad2d-7263741cbc31 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 30.458s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1980.375082] env[62824]: DEBUG nova.objects.instance [None req-89ab4335-c79f-4e08-ad2d-7263741cbc31 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Lazy-loading 'resources' on Instance uuid 16b466c3-c749-4f96-a82c-32dad31138ec {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1980.405289] env[62824]: DEBUG oslo_vmware.api [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Task: {'id': task-2145913, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1980.447907] env[62824]: DEBUG oslo_vmware.api [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5276e5f5-3561-acb9-9a5e-4bda5e2fd33c, 'name': SearchDatastore_Task, 'duration_secs': 0.011066} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1980.448946] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1980.451455] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] bbf1440b-8681-48c8-a178-9a83b925c695/bbf1440b-8681-48c8-a178-9a83b925c695.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1980.451777] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6e75f45f-e0dd-4d4e-b8f7-fd2c534763b7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1980.464377] env[62824]: DEBUG oslo_vmware.api [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Waiting for the task: (returnval){ [ 1980.464377] env[62824]: value = "task-2145914" [ 1980.464377] env[62824]: _type = "Task" [ 1980.464377] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1980.474544] env[62824]: DEBUG oslo_vmware.api [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2145914, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1980.596640] env[62824]: DEBUG nova.compute.manager [req-7569e534-d9c0-455b-860d-ad14c5402afc req-b74d163f-193a-4d2f-aa12-ff4993b5c5b0 service nova] [instance: dc3317e4-6958-4df6-8e39-c89b4a5d861c] Received event network-vif-plugged-4c03a7c0-a685-4fe9-9dce-635f6dda43b0 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1980.596640] env[62824]: DEBUG oslo_concurrency.lockutils [req-7569e534-d9c0-455b-860d-ad14c5402afc req-b74d163f-193a-4d2f-aa12-ff4993b5c5b0 service nova] Acquiring lock "dc3317e4-6958-4df6-8e39-c89b4a5d861c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1980.597057] env[62824]: DEBUG oslo_concurrency.lockutils [req-7569e534-d9c0-455b-860d-ad14c5402afc req-b74d163f-193a-4d2f-aa12-ff4993b5c5b0 service nova] Lock "dc3317e4-6958-4df6-8e39-c89b4a5d861c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1980.597295] env[62824]: DEBUG oslo_concurrency.lockutils [req-7569e534-d9c0-455b-860d-ad14c5402afc req-b74d163f-193a-4d2f-aa12-ff4993b5c5b0 service nova] Lock "dc3317e4-6958-4df6-8e39-c89b4a5d861c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1980.597506] env[62824]: DEBUG nova.compute.manager [req-7569e534-d9c0-455b-860d-ad14c5402afc req-b74d163f-193a-4d2f-aa12-ff4993b5c5b0 service nova] [instance: dc3317e4-6958-4df6-8e39-c89b4a5d861c] No waiting events found dispatching network-vif-plugged-4c03a7c0-a685-4fe9-9dce-635f6dda43b0 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1980.597683] env[62824]: WARNING nova.compute.manager [req-7569e534-d9c0-455b-860d-ad14c5402afc req-b74d163f-193a-4d2f-aa12-ff4993b5c5b0 service nova] [instance: dc3317e4-6958-4df6-8e39-c89b4a5d861c] Received unexpected event network-vif-plugged-4c03a7c0-a685-4fe9-9dce-635f6dda43b0 for instance with vm_state building and task_state spawning. [ 1980.597856] env[62824]: DEBUG nova.compute.manager [req-7569e534-d9c0-455b-860d-ad14c5402afc req-b74d163f-193a-4d2f-aa12-ff4993b5c5b0 service nova] [instance: dc3317e4-6958-4df6-8e39-c89b4a5d861c] Received event network-changed-4c03a7c0-a685-4fe9-9dce-635f6dda43b0 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1980.598024] env[62824]: DEBUG nova.compute.manager [req-7569e534-d9c0-455b-860d-ad14c5402afc req-b74d163f-193a-4d2f-aa12-ff4993b5c5b0 service nova] [instance: dc3317e4-6958-4df6-8e39-c89b4a5d861c] Refreshing instance network info cache due to event network-changed-4c03a7c0-a685-4fe9-9dce-635f6dda43b0. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1980.598221] env[62824]: DEBUG oslo_concurrency.lockutils [req-7569e534-d9c0-455b-860d-ad14c5402afc req-b74d163f-193a-4d2f-aa12-ff4993b5c5b0 service nova] Acquiring lock "refresh_cache-dc3317e4-6958-4df6-8e39-c89b4a5d861c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1980.848157] env[62824]: DEBUG nova.network.neutron [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: dc3317e4-6958-4df6-8e39-c89b4a5d861c] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1980.879699] env[62824]: DEBUG nova.compute.utils [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1980.888765] env[62824]: DEBUG nova.compute.manager [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1980.889498] env[62824]: DEBUG nova.network.neutron [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1980.917424] env[62824]: DEBUG oslo_vmware.api [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Task: {'id': task-2145913, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.599764} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1980.918777] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 5f1237ff-5841-41f1-8d77-df9fabaa62b6/5f1237ff-5841-41f1-8d77-df9fabaa62b6.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1980.918777] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] [instance: 5f1237ff-5841-41f1-8d77-df9fabaa62b6] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1980.918777] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4e5f0ce7-d1da-4747-9141-24efbdc4805d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1980.932351] env[62824]: DEBUG oslo_vmware.api [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Waiting for the task: (returnval){ [ 1980.932351] env[62824]: value = "task-2145916" [ 1980.932351] env[62824]: _type = "Task" [ 1980.932351] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1980.948702] env[62824]: DEBUG oslo_vmware.api [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Task: {'id': task-2145916, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1980.975571] env[62824]: DEBUG nova.policy [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '574e3bb9d7b647deb801cc2cbc308eaa', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'adf40520c1ef4d97b1ffe2c8cc9649b9', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 1980.992031] env[62824]: DEBUG oslo_vmware.api [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2145914, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.503755} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1980.992244] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] bbf1440b-8681-48c8-a178-9a83b925c695/bbf1440b-8681-48c8-a178-9a83b925c695.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1980.992469] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1980.992807] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-87d19d2c-080b-498f-af71-5bdb3b707410 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1981.024667] env[62824]: DEBUG oslo_vmware.api [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Waiting for the task: (returnval){ [ 1981.024667] env[62824]: value = "task-2145917" [ 1981.024667] env[62824]: _type = "Task" [ 1981.024667] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1981.039521] env[62824]: DEBUG oslo_vmware.api [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2145917, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1981.363862] env[62824]: DEBUG nova.network.neutron [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Successfully created port: b27bda53-5f70-40ff-bb04-9b22f2241093 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1981.388850] env[62824]: DEBUG nova.compute.manager [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1981.447738] env[62824]: DEBUG oslo_vmware.api [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Task: {'id': task-2145916, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.099397} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1981.448028] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] [instance: 5f1237ff-5841-41f1-8d77-df9fabaa62b6] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1981.449019] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8440f497-5de0-48ef-8798-cf9f1061392c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1981.476086] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] [instance: 5f1237ff-5841-41f1-8d77-df9fabaa62b6] Reconfiguring VM instance instance-00000025 to attach disk [datastore1] 5f1237ff-5841-41f1-8d77-df9fabaa62b6/5f1237ff-5841-41f1-8d77-df9fabaa62b6.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1981.478833] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9bf6ff9e-a16e-4ebe-add5-4dfa0929c88d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1981.505278] env[62824]: DEBUG oslo_vmware.api [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Waiting for the task: (returnval){ [ 1981.505278] env[62824]: value = "task-2145918" [ 1981.505278] env[62824]: _type = "Task" [ 1981.505278] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1981.509707] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c0f5040-fd18-4d4e-8d16-4941c8592ac1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1981.520363] env[62824]: DEBUG oslo_vmware.api [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Task: {'id': task-2145918, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1981.521408] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-305a07ae-e6a6-468a-a769-e4fd4b59affa {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1981.556575] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b130311-1095-435d-8e51-3d16cfde49dd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1981.562184] env[62824]: DEBUG oslo_vmware.api [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2145917, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1981.567544] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2aa7311f-3799-4c16-9809-344aef1f1014 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1981.583478] env[62824]: DEBUG nova.compute.provider_tree [None req-89ab4335-c79f-4e08-ad2d-7263741cbc31 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1981.659173] env[62824]: DEBUG nova.network.neutron [req-1b02348e-8fc8-4254-a430-c23898d6e296 req-f1b61601-105e-4a06-b73b-db84a9f11010 service nova] [instance: 5f1237ff-5841-41f1-8d77-df9fabaa62b6] Updated VIF entry in instance network info cache for port 660b16ae-ebd9-43c6-a3dd-3338c7b75f72. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1981.659544] env[62824]: DEBUG nova.network.neutron [req-1b02348e-8fc8-4254-a430-c23898d6e296 req-f1b61601-105e-4a06-b73b-db84a9f11010 service nova] [instance: 5f1237ff-5841-41f1-8d77-df9fabaa62b6] Updating instance_info_cache with network_info: [{"id": "660b16ae-ebd9-43c6-a3dd-3338c7b75f72", "address": "fa:16:3e:fb:60:5b", "network": {"id": "1616481c-2d55-4ef3-a6c3-0619a48c8f68", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-209823666-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "87eb5358b5fe4919b0d657d3e0d008f0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d8383707-f093-40a7-a5ba-31b0e07cac45", "external-id": "cl2-zone-18", "segmentation_id": 18, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap660b16ae-eb", "ovs_interfaceid": "660b16ae-ebd9-43c6-a3dd-3338c7b75f72", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1981.727407] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Acquiring lock "ed7d3e95-f26e-40fd-a370-e17922bbff8e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1981.727407] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Lock "ed7d3e95-f26e-40fd-a370-e17922bbff8e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1982.018717] env[62824]: DEBUG oslo_vmware.api [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Task: {'id': task-2145918, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1982.035993] env[62824]: DEBUG oslo_vmware.api [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2145917, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1982.087850] env[62824]: DEBUG nova.scheduler.client.report [None req-89ab4335-c79f-4e08-ad2d-7263741cbc31 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1982.150103] env[62824]: DEBUG nova.network.neutron [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: dc3317e4-6958-4df6-8e39-c89b4a5d861c] Updating instance_info_cache with network_info: [{"id": "ee0328da-cfb9-4f03-94fc-1acb84f8c553", "address": "fa:16:3e:ac:67:00", "network": {"id": "04834b3b-6a86-4ce1-85fb-4c713962e000", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-538402120", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.136", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4cb963b19c1549ab8a11e5a77e4f202d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3202e2b9-c4a7-4f78-9476-12ed92fabe61", "external-id": "nsx-vlan-transportzone-846", "segmentation_id": 846, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapee0328da-cf", "ovs_interfaceid": "ee0328da-cfb9-4f03-94fc-1acb84f8c553", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "4c03a7c0-a685-4fe9-9dce-635f6dda43b0", "address": "fa:16:3e:3a:30:1d", "network": {"id": "a1dc1a68-2912-4ec5-8bc2-e7294c0563c1", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1056165949", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.118", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "4cb963b19c1549ab8a11e5a77e4f202d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "650f7968-4522-4ba5-8304-1b9949951ed7", "external-id": "nsx-vlan-transportzone-568", "segmentation_id": 568, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c03a7c0-a6", "ovs_interfaceid": "4c03a7c0-a685-4fe9-9dce-635f6dda43b0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1982.162907] env[62824]: DEBUG oslo_concurrency.lockutils [req-1b02348e-8fc8-4254-a430-c23898d6e296 req-f1b61601-105e-4a06-b73b-db84a9f11010 service nova] Releasing lock "refresh_cache-5f1237ff-5841-41f1-8d77-df9fabaa62b6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1982.399296] env[62824]: DEBUG nova.compute.manager [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1982.429569] env[62824]: DEBUG nova.virt.hardware [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1982.429800] env[62824]: DEBUG nova.virt.hardware [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1982.429957] env[62824]: DEBUG nova.virt.hardware [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1982.430158] env[62824]: DEBUG nova.virt.hardware [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1982.430306] env[62824]: DEBUG nova.virt.hardware [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1982.430449] env[62824]: DEBUG nova.virt.hardware [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1982.431787] env[62824]: DEBUG nova.virt.hardware [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1982.432110] env[62824]: DEBUG nova.virt.hardware [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1982.432422] env[62824]: DEBUG nova.virt.hardware [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1982.432662] env[62824]: DEBUG nova.virt.hardware [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1982.432913] env[62824]: DEBUG nova.virt.hardware [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1982.434312] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76cbac37-f523-4597-8c93-ddcb2802307b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1982.443987] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-029e94d4-58c4-451f-81cf-509be155c9f3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1982.520038] env[62824]: DEBUG oslo_vmware.api [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Task: {'id': task-2145918, 'name': ReconfigVM_Task, 'duration_secs': 0.806336} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1982.520391] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] [instance: 5f1237ff-5841-41f1-8d77-df9fabaa62b6] Reconfigured VM instance instance-00000025 to attach disk [datastore1] 5f1237ff-5841-41f1-8d77-df9fabaa62b6/5f1237ff-5841-41f1-8d77-df9fabaa62b6.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1982.521065] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ac01b7d3-5e04-48fc-987e-7a3333747fef {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1982.535774] env[62824]: DEBUG oslo_vmware.api [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Waiting for the task: (returnval){ [ 1982.535774] env[62824]: value = "task-2145919" [ 1982.535774] env[62824]: _type = "Task" [ 1982.535774] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1982.543772] env[62824]: DEBUG oslo_vmware.api [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2145917, 'name': ExtendVirtualDisk_Task, 'duration_secs': 1.093451} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1982.544491] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1982.545684] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73259ec0-f0c8-4be2-aecf-824587b3db76 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1982.551915] env[62824]: DEBUG oslo_vmware.api [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Task: {'id': task-2145919, 'name': Rename_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1982.575223] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Reconfiguring VM instance instance-00000027 to attach disk [datastore1] bbf1440b-8681-48c8-a178-9a83b925c695/bbf1440b-8681-48c8-a178-9a83b925c695.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1982.575565] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-241ab847-c9be-4504-b64e-78dec5314feb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1982.593640] env[62824]: DEBUG oslo_concurrency.lockutils [None req-89ab4335-c79f-4e08-ad2d-7263741cbc31 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.219s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1982.596133] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 30.317s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1982.596432] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1982.596560] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62824) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1982.596737] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.964s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1982.598339] env[62824]: INFO nova.compute.claims [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1982.606329] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a274147-7d38-476c-9a78-dfe570891c8d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1982.611452] env[62824]: DEBUG oslo_vmware.api [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Waiting for the task: (returnval){ [ 1982.611452] env[62824]: value = "task-2145920" [ 1982.611452] env[62824]: _type = "Task" [ 1982.611452] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1982.620328] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72aee6b1-306b-41ee-9931-ba6d136c540b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1982.628526] env[62824]: DEBUG oslo_vmware.api [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2145920, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1982.641617] env[62824]: INFO nova.scheduler.client.report [None req-89ab4335-c79f-4e08-ad2d-7263741cbc31 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Deleted allocations for instance 16b466c3-c749-4f96-a82c-32dad31138ec [ 1982.647111] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a47ad24c-dd4e-4272-8110-a051241c1f19 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1982.653726] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Releasing lock "refresh_cache-dc3317e4-6958-4df6-8e39-c89b4a5d861c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1982.653837] env[62824]: DEBUG nova.compute.manager [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: dc3317e4-6958-4df6-8e39-c89b4a5d861c] Instance network_info: |[{"id": "ee0328da-cfb9-4f03-94fc-1acb84f8c553", "address": "fa:16:3e:ac:67:00", "network": {"id": "04834b3b-6a86-4ce1-85fb-4c713962e000", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-538402120", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.136", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4cb963b19c1549ab8a11e5a77e4f202d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3202e2b9-c4a7-4f78-9476-12ed92fabe61", "external-id": "nsx-vlan-transportzone-846", "segmentation_id": 846, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapee0328da-cf", "ovs_interfaceid": "ee0328da-cfb9-4f03-94fc-1acb84f8c553", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "4c03a7c0-a685-4fe9-9dce-635f6dda43b0", "address": "fa:16:3e:3a:30:1d", "network": {"id": "a1dc1a68-2912-4ec5-8bc2-e7294c0563c1", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1056165949", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.118", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "4cb963b19c1549ab8a11e5a77e4f202d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "650f7968-4522-4ba5-8304-1b9949951ed7", "external-id": "nsx-vlan-transportzone-568", "segmentation_id": 568, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c03a7c0-a6", "ovs_interfaceid": "4c03a7c0-a685-4fe9-9dce-635f6dda43b0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1982.654572] env[62824]: DEBUG oslo_concurrency.lockutils [req-7569e534-d9c0-455b-860d-ad14c5402afc req-b74d163f-193a-4d2f-aa12-ff4993b5c5b0 service nova] Acquired lock "refresh_cache-dc3317e4-6958-4df6-8e39-c89b4a5d861c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1982.654783] env[62824]: DEBUG nova.network.neutron [req-7569e534-d9c0-455b-860d-ad14c5402afc req-b74d163f-193a-4d2f-aa12-ff4993b5c5b0 service nova] [instance: dc3317e4-6958-4df6-8e39-c89b4a5d861c] Refreshing network info cache for port 4c03a7c0-a685-4fe9-9dce-635f6dda43b0 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1982.656381] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: dc3317e4-6958-4df6-8e39-c89b4a5d861c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ac:67:00', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3202e2b9-c4a7-4f78-9476-12ed92fabe61', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ee0328da-cfb9-4f03-94fc-1acb84f8c553', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:3a:30:1d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '650f7968-4522-4ba5-8304-1b9949951ed7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4c03a7c0-a685-4fe9-9dce-635f6dda43b0', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1982.667466] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Creating folder: Project (4cb963b19c1549ab8a11e5a77e4f202d). Parent ref: group-v438503. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1982.672503] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5c4f7479-65b6-4931-91ee-28b5e45709bb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1982.676292] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20f66355-1f9f-4a49-a14b-ed20f00d1e9b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1982.710917] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180167MB free_disk=174GB free_vcpus=48 pci_devices=None {{(pid=62824) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1982.711137] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1982.713169] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Created folder: Project (4cb963b19c1549ab8a11e5a77e4f202d) in parent group-v438503. [ 1982.713515] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Creating folder: Instances. Parent ref: group-v438592. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1982.714381] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-08e49daf-52d4-4d4b-bb6b-31d91e5b65e3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1982.728653] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Created folder: Instances in parent group-v438592. [ 1982.728947] env[62824]: DEBUG oslo.service.loopingcall [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1982.729301] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dc3317e4-6958-4df6-8e39-c89b4a5d861c] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1982.729717] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-11b80329-a219-433c-a029-f61b7cbb3e9d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1982.760832] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1982.760832] env[62824]: value = "task-2145923" [ 1982.760832] env[62824]: _type = "Task" [ 1982.760832] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1982.773276] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145923, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1982.933238] env[62824]: DEBUG nova.network.neutron [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Successfully updated port: b27bda53-5f70-40ff-bb04-9b22f2241093 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1982.968334] env[62824]: DEBUG nova.compute.manager [req-7af9cd3a-a8df-4124-9bf4-07bbc2e603f9 req-6711f637-f39e-4188-862a-3aa77869a5e3 service nova] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Received event network-vif-plugged-b27bda53-5f70-40ff-bb04-9b22f2241093 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1982.968334] env[62824]: DEBUG oslo_concurrency.lockutils [req-7af9cd3a-a8df-4124-9bf4-07bbc2e603f9 req-6711f637-f39e-4188-862a-3aa77869a5e3 service nova] Acquiring lock "7febb8b4-f984-4d79-a888-b2829f2a9df6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1982.968334] env[62824]: DEBUG oslo_concurrency.lockutils [req-7af9cd3a-a8df-4124-9bf4-07bbc2e603f9 req-6711f637-f39e-4188-862a-3aa77869a5e3 service nova] Lock "7febb8b4-f984-4d79-a888-b2829f2a9df6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1982.969036] env[62824]: DEBUG oslo_concurrency.lockutils [req-7af9cd3a-a8df-4124-9bf4-07bbc2e603f9 req-6711f637-f39e-4188-862a-3aa77869a5e3 service nova] Lock "7febb8b4-f984-4d79-a888-b2829f2a9df6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1982.969036] env[62824]: DEBUG nova.compute.manager [req-7af9cd3a-a8df-4124-9bf4-07bbc2e603f9 req-6711f637-f39e-4188-862a-3aa77869a5e3 service nova] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] No waiting events found dispatching network-vif-plugged-b27bda53-5f70-40ff-bb04-9b22f2241093 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1982.969036] env[62824]: WARNING nova.compute.manager [req-7af9cd3a-a8df-4124-9bf4-07bbc2e603f9 req-6711f637-f39e-4188-862a-3aa77869a5e3 service nova] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Received unexpected event network-vif-plugged-b27bda53-5f70-40ff-bb04-9b22f2241093 for instance with vm_state building and task_state spawning. [ 1983.046933] env[62824]: DEBUG oslo_vmware.api [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Task: {'id': task-2145919, 'name': Rename_Task, 'duration_secs': 0.164178} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1983.047244] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] [instance: 5f1237ff-5841-41f1-8d77-df9fabaa62b6] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1983.047502] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b9e551a3-7fb8-4051-8a50-4cfa882203c1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1983.055356] env[62824]: DEBUG oslo_vmware.api [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Waiting for the task: (returnval){ [ 1983.055356] env[62824]: value = "task-2145924" [ 1983.055356] env[62824]: _type = "Task" [ 1983.055356] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1983.063843] env[62824]: DEBUG oslo_vmware.api [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Task: {'id': task-2145924, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1983.123354] env[62824]: DEBUG oslo_vmware.api [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2145920, 'name': ReconfigVM_Task, 'duration_secs': 0.442672} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1983.124337] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Reconfigured VM instance instance-00000027 to attach disk [datastore1] bbf1440b-8681-48c8-a178-9a83b925c695/bbf1440b-8681-48c8-a178-9a83b925c695.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1983.125037] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3ab897b8-d60b-45ba-8180-a0abb3ba9d4e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1983.131869] env[62824]: DEBUG oslo_vmware.api [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Waiting for the task: (returnval){ [ 1983.131869] env[62824]: value = "task-2145925" [ 1983.131869] env[62824]: _type = "Task" [ 1983.131869] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1983.141701] env[62824]: DEBUG oslo_vmware.api [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2145925, 'name': Rename_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1983.156059] env[62824]: DEBUG oslo_concurrency.lockutils [None req-89ab4335-c79f-4e08-ad2d-7263741cbc31 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Lock "16b466c3-c749-4f96-a82c-32dad31138ec" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 36.467s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1983.254974] env[62824]: INFO nova.compute.manager [None req-f73e7f3e-71a9-4684-9e15-596a02702690 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Unrescuing [ 1983.255262] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f73e7f3e-71a9-4684-9e15-596a02702690 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Acquiring lock "refresh_cache-b3003c4b-ae5a-48df-8c12-a915a76253f4" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1983.255414] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f73e7f3e-71a9-4684-9e15-596a02702690 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Acquired lock "refresh_cache-b3003c4b-ae5a-48df-8c12-a915a76253f4" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1983.255573] env[62824]: DEBUG nova.network.neutron [None req-f73e7f3e-71a9-4684-9e15-596a02702690 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1983.272149] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145923, 'name': CreateVM_Task, 'duration_secs': 0.442559} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1983.275266] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dc3317e4-6958-4df6-8e39-c89b4a5d861c] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1983.275957] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1983.276130] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1983.276445] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1983.277216] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a4f9e61f-dd68-4194-8bc9-9476000f5d81 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1983.282499] env[62824]: DEBUG oslo_vmware.api [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Waiting for the task: (returnval){ [ 1983.282499] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5263dfc7-4b7a-5cae-3979-f4b0680a4f05" [ 1983.282499] env[62824]: _type = "Task" [ 1983.282499] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1983.292626] env[62824]: DEBUG oslo_vmware.api [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5263dfc7-4b7a-5cae-3979-f4b0680a4f05, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1983.389397] env[62824]: DEBUG nova.network.neutron [req-7569e534-d9c0-455b-860d-ad14c5402afc req-b74d163f-193a-4d2f-aa12-ff4993b5c5b0 service nova] [instance: dc3317e4-6958-4df6-8e39-c89b4a5d861c] Updated VIF entry in instance network info cache for port 4c03a7c0-a685-4fe9-9dce-635f6dda43b0. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1983.389844] env[62824]: DEBUG nova.network.neutron [req-7569e534-d9c0-455b-860d-ad14c5402afc req-b74d163f-193a-4d2f-aa12-ff4993b5c5b0 service nova] [instance: dc3317e4-6958-4df6-8e39-c89b4a5d861c] Updating instance_info_cache with network_info: [{"id": "ee0328da-cfb9-4f03-94fc-1acb84f8c553", "address": "fa:16:3e:ac:67:00", "network": {"id": "04834b3b-6a86-4ce1-85fb-4c713962e000", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-538402120", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.136", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4cb963b19c1549ab8a11e5a77e4f202d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3202e2b9-c4a7-4f78-9476-12ed92fabe61", "external-id": "nsx-vlan-transportzone-846", "segmentation_id": 846, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapee0328da-cf", "ovs_interfaceid": "ee0328da-cfb9-4f03-94fc-1acb84f8c553", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "4c03a7c0-a685-4fe9-9dce-635f6dda43b0", "address": "fa:16:3e:3a:30:1d", "network": {"id": "a1dc1a68-2912-4ec5-8bc2-e7294c0563c1", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1056165949", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.118", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "4cb963b19c1549ab8a11e5a77e4f202d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "650f7968-4522-4ba5-8304-1b9949951ed7", "external-id": "nsx-vlan-transportzone-568", "segmentation_id": 568, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c03a7c0-a6", "ovs_interfaceid": "4c03a7c0-a685-4fe9-9dce-635f6dda43b0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1983.436662] env[62824]: DEBUG oslo_concurrency.lockutils [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Acquiring lock "refresh_cache-7febb8b4-f984-4d79-a888-b2829f2a9df6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1983.436898] env[62824]: DEBUG oslo_concurrency.lockutils [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Acquired lock "refresh_cache-7febb8b4-f984-4d79-a888-b2829f2a9df6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1983.437068] env[62824]: DEBUG nova.network.neutron [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1983.569276] env[62824]: DEBUG oslo_vmware.api [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Task: {'id': task-2145924, 'name': PowerOnVM_Task, 'duration_secs': 0.465303} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1983.569548] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] [instance: 5f1237ff-5841-41f1-8d77-df9fabaa62b6] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1983.569740] env[62824]: INFO nova.compute.manager [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] [instance: 5f1237ff-5841-41f1-8d77-df9fabaa62b6] Took 13.52 seconds to spawn the instance on the hypervisor. [ 1983.569911] env[62824]: DEBUG nova.compute.manager [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] [instance: 5f1237ff-5841-41f1-8d77-df9fabaa62b6] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1983.570889] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84a027ce-3028-40a1-a637-bab3d5b77116 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1983.644196] env[62824]: DEBUG oslo_vmware.api [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2145925, 'name': Rename_Task, 'duration_secs': 0.211732} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1983.645051] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1983.645051] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b722cb88-c122-498b-8c84-e614c9995548 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1983.654927] env[62824]: DEBUG oslo_vmware.api [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Waiting for the task: (returnval){ [ 1983.654927] env[62824]: value = "task-2145926" [ 1983.654927] env[62824]: _type = "Task" [ 1983.654927] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1983.669185] env[62824]: DEBUG oslo_vmware.api [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2145926, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1983.794526] env[62824]: DEBUG oslo_vmware.api [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5263dfc7-4b7a-5cae-3979-f4b0680a4f05, 'name': SearchDatastore_Task, 'duration_secs': 0.015029} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1983.795485] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1983.795485] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: dc3317e4-6958-4df6-8e39-c89b4a5d861c] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1983.795485] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1983.795886] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1983.796743] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1983.801618] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-73c231ac-22aa-40b8-96cd-24f71f092ccf {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1983.814518] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1983.814760] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1983.818066] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-91b106f5-fdef-47eb-b0fa-dfcdfa9d5258 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1983.826565] env[62824]: DEBUG oslo_vmware.api [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Waiting for the task: (returnval){ [ 1983.826565] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]520848e2-8b89-d62f-7b17-0043d0355f0d" [ 1983.826565] env[62824]: _type = "Task" [ 1983.826565] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1983.840142] env[62824]: DEBUG oslo_vmware.api [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]520848e2-8b89-d62f-7b17-0043d0355f0d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1983.894559] env[62824]: DEBUG oslo_concurrency.lockutils [req-7569e534-d9c0-455b-860d-ad14c5402afc req-b74d163f-193a-4d2f-aa12-ff4993b5c5b0 service nova] Releasing lock "refresh_cache-dc3317e4-6958-4df6-8e39-c89b4a5d861c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1983.992796] env[62824]: DEBUG nova.network.neutron [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1984.023751] env[62824]: DEBUG nova.network.neutron [None req-f73e7f3e-71a9-4684-9e15-596a02702690 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Updating instance_info_cache with network_info: [{"id": "4e0d51df-4419-4059-9dab-be44957e4f77", "address": "fa:16:3e:15:1a:78", "network": {"id": "a410f4f3-d072-4034-b797-8210e1a79093", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1482574391-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e10208040df34eb5adfe8dcbc76043d5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f4399275-8e92-4448-be9e-d4984e93e89c", "external-id": "nsx-vlan-transportzone-192", "segmentation_id": 192, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e0d51df-44", "ovs_interfaceid": "4e0d51df-4419-4059-9dab-be44957e4f77", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1984.103094] env[62824]: INFO nova.compute.manager [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] [instance: 5f1237ff-5841-41f1-8d77-df9fabaa62b6] Took 53.15 seconds to build instance. [ 1984.167233] env[62824]: DEBUG oslo_vmware.api [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2145926, 'name': PowerOnVM_Task} progress is 87%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1984.190708] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8180ab15-8ed4-41cd-9f30-0facb703a1f1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1984.199505] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86127e28-3cb5-451d-9872-dbfc576491c5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1984.245108] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aacee6be-7a66-4024-9a03-830b1f8cef5c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1984.255062] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-922654f5-9fd3-4319-ab93-5f65f0e22ab1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1984.273469] env[62824]: DEBUG nova.compute.provider_tree [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1984.339180] env[62824]: DEBUG oslo_vmware.api [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]520848e2-8b89-d62f-7b17-0043d0355f0d, 'name': SearchDatastore_Task, 'duration_secs': 0.023434} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1984.340425] env[62824]: DEBUG nova.network.neutron [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Updating instance_info_cache with network_info: [{"id": "b27bda53-5f70-40ff-bb04-9b22f2241093", "address": "fa:16:3e:4d:09:75", "network": {"id": "3d2f0f85-4d51-4459-ad99-443b0821fd9d", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1138734647-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "adf40520c1ef4d97b1ffe2c8cc9649b9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "10b81051-1eb1-406b-888c-4548c470c77e", "external-id": "nsx-vlan-transportzone-207", "segmentation_id": 207, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb27bda53-5f", "ovs_interfaceid": "b27bda53-5f70-40ff-bb04-9b22f2241093", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1984.342615] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d4e7b216-6776-4b63-8102-f843c875ac1b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1984.350176] env[62824]: DEBUG oslo_vmware.api [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Waiting for the task: (returnval){ [ 1984.350176] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52cb9f3a-4ecc-b256-21b2-e4f3a418c85c" [ 1984.350176] env[62824]: _type = "Task" [ 1984.350176] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1984.362239] env[62824]: DEBUG oslo_vmware.api [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52cb9f3a-4ecc-b256-21b2-e4f3a418c85c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1984.529129] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f73e7f3e-71a9-4684-9e15-596a02702690 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Releasing lock "refresh_cache-b3003c4b-ae5a-48df-8c12-a915a76253f4" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1984.529129] env[62824]: DEBUG nova.objects.instance [None req-f73e7f3e-71a9-4684-9e15-596a02702690 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Lazy-loading 'flavor' on Instance uuid b3003c4b-ae5a-48df-8c12-a915a76253f4 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1984.611363] env[62824]: DEBUG oslo_concurrency.lockutils [None req-816f632c-5cc5-41ac-8ada-89bedd60505f tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Lock "5f1237ff-5841-41f1-8d77-df9fabaa62b6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 106.598s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1984.669246] env[62824]: DEBUG oslo_vmware.api [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2145926, 'name': PowerOnVM_Task, 'duration_secs': 0.830301} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1984.669567] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1984.669708] env[62824]: INFO nova.compute.manager [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Took 9.47 seconds to spawn the instance on the hypervisor. [ 1984.669874] env[62824]: DEBUG nova.compute.manager [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1984.670654] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc168d21-0d17-4dc0-8736-e5b985df83f2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1984.778945] env[62824]: DEBUG nova.scheduler.client.report [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1984.843542] env[62824]: DEBUG oslo_concurrency.lockutils [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Releasing lock "refresh_cache-7febb8b4-f984-4d79-a888-b2829f2a9df6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1984.843932] env[62824]: DEBUG nova.compute.manager [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Instance network_info: |[{"id": "b27bda53-5f70-40ff-bb04-9b22f2241093", "address": "fa:16:3e:4d:09:75", "network": {"id": "3d2f0f85-4d51-4459-ad99-443b0821fd9d", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1138734647-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "adf40520c1ef4d97b1ffe2c8cc9649b9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "10b81051-1eb1-406b-888c-4548c470c77e", "external-id": "nsx-vlan-transportzone-207", "segmentation_id": 207, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb27bda53-5f", "ovs_interfaceid": "b27bda53-5f70-40ff-bb04-9b22f2241093", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1984.844439] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4d:09:75', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '10b81051-1eb1-406b-888c-4548c470c77e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b27bda53-5f70-40ff-bb04-9b22f2241093', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1984.853453] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Creating folder: Project (adf40520c1ef4d97b1ffe2c8cc9649b9). Parent ref: group-v438503. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1984.853854] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a79f2e6a-860c-428a-8214-dea4abc04d9b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1984.867702] env[62824]: DEBUG oslo_vmware.api [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52cb9f3a-4ecc-b256-21b2-e4f3a418c85c, 'name': SearchDatastore_Task, 'duration_secs': 0.013533} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1984.867968] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1984.868295] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] dc3317e4-6958-4df6-8e39-c89b4a5d861c/dc3317e4-6958-4df6-8e39-c89b4a5d861c.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1984.868574] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a1ddc775-96a0-45f5-9960-097f1a96e3ec {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1984.871864] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Created folder: Project (adf40520c1ef4d97b1ffe2c8cc9649b9) in parent group-v438503. [ 1984.872053] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Creating folder: Instances. Parent ref: group-v438595. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1984.872612] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-122e8760-ebec-45fd-9538-2f8b9461a02d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1984.878035] env[62824]: DEBUG oslo_vmware.api [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Waiting for the task: (returnval){ [ 1984.878035] env[62824]: value = "task-2145929" [ 1984.878035] env[62824]: _type = "Task" [ 1984.878035] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1984.885386] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Created folder: Instances in parent group-v438595. [ 1984.887538] env[62824]: DEBUG oslo.service.loopingcall [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1984.887538] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1984.887706] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5f61bf1f-72c1-43cb-80c3-682cebf54c27 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1984.912408] env[62824]: DEBUG oslo_vmware.api [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Task: {'id': task-2145929, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1984.919391] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1984.919391] env[62824]: value = "task-2145931" [ 1984.919391] env[62824]: _type = "Task" [ 1984.919391] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1984.933834] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145931, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1985.034141] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77124613-e06f-4346-97ae-0c6d2bab6bef {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1985.058017] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-f73e7f3e-71a9-4684-9e15-596a02702690 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1985.058271] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e42d650c-b6d9-4c65-933e-abd980159bee {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1985.069454] env[62824]: DEBUG oslo_vmware.api [None req-f73e7f3e-71a9-4684-9e15-596a02702690 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for the task: (returnval){ [ 1985.069454] env[62824]: value = "task-2145932" [ 1985.069454] env[62824]: _type = "Task" [ 1985.069454] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1985.080765] env[62824]: DEBUG oslo_vmware.api [None req-f73e7f3e-71a9-4684-9e15-596a02702690 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2145932, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1985.113990] env[62824]: DEBUG nova.compute.manager [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1985.122256] env[62824]: DEBUG oslo_concurrency.lockutils [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Acquiring lock "b0456b8a-348a-4503-a92c-58e5ab455d1c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1985.124405] env[62824]: DEBUG oslo_concurrency.lockutils [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Lock "b0456b8a-348a-4503-a92c-58e5ab455d1c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1985.126726] env[62824]: DEBUG nova.compute.manager [req-3617e7b3-c93f-4275-a221-adca0fc88221 req-8fab501a-06c0-47a2-8b2f-fc049116c618 service nova] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Received event network-changed-b27bda53-5f70-40ff-bb04-9b22f2241093 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1985.126944] env[62824]: DEBUG nova.compute.manager [req-3617e7b3-c93f-4275-a221-adca0fc88221 req-8fab501a-06c0-47a2-8b2f-fc049116c618 service nova] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Refreshing instance network info cache due to event network-changed-b27bda53-5f70-40ff-bb04-9b22f2241093. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1985.127219] env[62824]: DEBUG oslo_concurrency.lockutils [req-3617e7b3-c93f-4275-a221-adca0fc88221 req-8fab501a-06c0-47a2-8b2f-fc049116c618 service nova] Acquiring lock "refresh_cache-7febb8b4-f984-4d79-a888-b2829f2a9df6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1985.127363] env[62824]: DEBUG oslo_concurrency.lockutils [req-3617e7b3-c93f-4275-a221-adca0fc88221 req-8fab501a-06c0-47a2-8b2f-fc049116c618 service nova] Acquired lock "refresh_cache-7febb8b4-f984-4d79-a888-b2829f2a9df6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1985.127557] env[62824]: DEBUG nova.network.neutron [req-3617e7b3-c93f-4275-a221-adca0fc88221 req-8fab501a-06c0-47a2-8b2f-fc049116c618 service nova] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Refreshing network info cache for port b27bda53-5f70-40ff-bb04-9b22f2241093 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1985.152131] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7a7b2101-f65d-4821-85ef-6e89be960d5e tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Acquiring lock "5f1237ff-5841-41f1-8d77-df9fabaa62b6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1985.152302] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7a7b2101-f65d-4821-85ef-6e89be960d5e tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Lock "5f1237ff-5841-41f1-8d77-df9fabaa62b6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1985.152632] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7a7b2101-f65d-4821-85ef-6e89be960d5e tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Acquiring lock "5f1237ff-5841-41f1-8d77-df9fabaa62b6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1985.152793] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7a7b2101-f65d-4821-85ef-6e89be960d5e tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Lock "5f1237ff-5841-41f1-8d77-df9fabaa62b6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1985.152989] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7a7b2101-f65d-4821-85ef-6e89be960d5e tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Lock "5f1237ff-5841-41f1-8d77-df9fabaa62b6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1985.155248] env[62824]: INFO nova.compute.manager [None req-7a7b2101-f65d-4821-85ef-6e89be960d5e tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] [instance: 5f1237ff-5841-41f1-8d77-df9fabaa62b6] Terminating instance [ 1985.190280] env[62824]: INFO nova.compute.manager [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Took 49.92 seconds to build instance. [ 1985.283356] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.686s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1985.283989] env[62824]: DEBUG nova.compute.manager [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1985.287833] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.318s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1985.288629] env[62824]: INFO nova.compute.claims [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 8bb90a4d-93c9-4f54-b15e-48fe966e6c97] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1985.389745] env[62824]: DEBUG oslo_vmware.api [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Task: {'id': task-2145929, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.508391} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1985.391067] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] dc3317e4-6958-4df6-8e39-c89b4a5d861c/dc3317e4-6958-4df6-8e39-c89b4a5d861c.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1985.391067] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: dc3317e4-6958-4df6-8e39-c89b4a5d861c] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1985.391067] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1c50cb97-e913-47d0-bba3-87cf6e5cb6b8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1985.399071] env[62824]: DEBUG oslo_vmware.api [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Waiting for the task: (returnval){ [ 1985.399071] env[62824]: value = "task-2145933" [ 1985.399071] env[62824]: _type = "Task" [ 1985.399071] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1985.409409] env[62824]: DEBUG oslo_vmware.api [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Task: {'id': task-2145933, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1985.432121] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145931, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1985.580469] env[62824]: DEBUG oslo_vmware.api [None req-f73e7f3e-71a9-4684-9e15-596a02702690 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2145932, 'name': PowerOffVM_Task, 'duration_secs': 0.282204} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1985.580469] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-f73e7f3e-71a9-4684-9e15-596a02702690 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1985.585170] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-f73e7f3e-71a9-4684-9e15-596a02702690 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Reconfiguring VM instance instance-00000020 to detach disk 2001 {{(pid=62824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1985.585455] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-16851a4f-4549-4c12-9db3-63d3c20b3a95 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1985.604761] env[62824]: DEBUG oslo_vmware.api [None req-f73e7f3e-71a9-4684-9e15-596a02702690 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for the task: (returnval){ [ 1985.604761] env[62824]: value = "task-2145934" [ 1985.604761] env[62824]: _type = "Task" [ 1985.604761] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1985.615336] env[62824]: DEBUG oslo_vmware.api [None req-f73e7f3e-71a9-4684-9e15-596a02702690 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2145934, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1985.642082] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1985.659574] env[62824]: DEBUG nova.compute.manager [None req-7a7b2101-f65d-4821-85ef-6e89be960d5e tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] [instance: 5f1237ff-5841-41f1-8d77-df9fabaa62b6] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1985.660012] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-7a7b2101-f65d-4821-85ef-6e89be960d5e tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] [instance: 5f1237ff-5841-41f1-8d77-df9fabaa62b6] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1985.660938] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e126493b-997c-4e0c-bab0-f99dbb9dab80 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1985.668874] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a7b2101-f65d-4821-85ef-6e89be960d5e tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] [instance: 5f1237ff-5841-41f1-8d77-df9fabaa62b6] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1985.669122] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-04fe9cdd-8ddf-4a79-bbbb-31cf4decebdf {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1985.675329] env[62824]: DEBUG oslo_vmware.api [None req-7a7b2101-f65d-4821-85ef-6e89be960d5e tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Waiting for the task: (returnval){ [ 1985.675329] env[62824]: value = "task-2145935" [ 1985.675329] env[62824]: _type = "Task" [ 1985.675329] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1985.683344] env[62824]: DEBUG oslo_vmware.api [None req-7a7b2101-f65d-4821-85ef-6e89be960d5e tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Task: {'id': task-2145935, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1985.691984] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fd710340-2cbc-4d7b-b47c-d8826285f2ca tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Lock "bbf1440b-8681-48c8-a178-9a83b925c695" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 97.330s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1985.793193] env[62824]: DEBUG nova.compute.utils [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1985.799979] env[62824]: DEBUG nova.compute.manager [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1985.800186] env[62824]: DEBUG nova.network.neutron [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1985.874978] env[62824]: DEBUG nova.policy [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '31bf68236cfe4cfeb562a22258f003d9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1ab99ed209af4474a4c8a0a6e3844b2e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 1985.912259] env[62824]: DEBUG oslo_vmware.api [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Task: {'id': task-2145933, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074371} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1985.912259] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: dc3317e4-6958-4df6-8e39-c89b4a5d861c] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1985.912259] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5c04972-d965-4d29-a5f2-0240305f79e4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1985.938176] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: dc3317e4-6958-4df6-8e39-c89b4a5d861c] Reconfiguring VM instance instance-00000026 to attach disk [datastore1] dc3317e4-6958-4df6-8e39-c89b4a5d861c/dc3317e4-6958-4df6-8e39-c89b4a5d861c.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1985.941653] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8f645ae5-e673-4802-a1bf-29142b81e0a7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1985.966634] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145931, 'name': CreateVM_Task, 'duration_secs': 0.627586} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1985.967809] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1985.968156] env[62824]: DEBUG oslo_vmware.api [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Waiting for the task: (returnval){ [ 1985.968156] env[62824]: value = "task-2145936" [ 1985.968156] env[62824]: _type = "Task" [ 1985.968156] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1985.968764] env[62824]: DEBUG oslo_concurrency.lockutils [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1985.968925] env[62824]: DEBUG oslo_concurrency.lockutils [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1985.969591] env[62824]: DEBUG oslo_concurrency.lockutils [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1985.970920] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d817bdd9-3d99-48a0-943f-8737d9dad095 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1985.980423] env[62824]: DEBUG oslo_vmware.api [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Waiting for the task: (returnval){ [ 1985.980423] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]524e7d59-9cb5-6b32-aa8a-7e78df4f6ea2" [ 1985.980423] env[62824]: _type = "Task" [ 1985.980423] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1985.984053] env[62824]: DEBUG oslo_vmware.api [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Task: {'id': task-2145936, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1985.996607] env[62824]: DEBUG oslo_vmware.api [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]524e7d59-9cb5-6b32-aa8a-7e78df4f6ea2, 'name': SearchDatastore_Task, 'duration_secs': 0.011183} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1985.997207] env[62824]: DEBUG oslo_concurrency.lockutils [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1985.997610] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1985.998039] env[62824]: DEBUG oslo_concurrency.lockutils [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1985.998210] env[62824]: DEBUG oslo_concurrency.lockutils [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1985.998422] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1985.998727] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8f9af5b0-ec9f-4d06-a781-fab213423cb4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1986.008941] env[62824]: DEBUG nova.network.neutron [req-3617e7b3-c93f-4275-a221-adca0fc88221 req-8fab501a-06c0-47a2-8b2f-fc049116c618 service nova] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Updated VIF entry in instance network info cache for port b27bda53-5f70-40ff-bb04-9b22f2241093. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1986.009340] env[62824]: DEBUG nova.network.neutron [req-3617e7b3-c93f-4275-a221-adca0fc88221 req-8fab501a-06c0-47a2-8b2f-fc049116c618 service nova] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Updating instance_info_cache with network_info: [{"id": "b27bda53-5f70-40ff-bb04-9b22f2241093", "address": "fa:16:3e:4d:09:75", "network": {"id": "3d2f0f85-4d51-4459-ad99-443b0821fd9d", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1138734647-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "adf40520c1ef4d97b1ffe2c8cc9649b9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "10b81051-1eb1-406b-888c-4548c470c77e", "external-id": "nsx-vlan-transportzone-207", "segmentation_id": 207, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb27bda53-5f", "ovs_interfaceid": "b27bda53-5f70-40ff-bb04-9b22f2241093", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1986.011648] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1986.011830] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1986.012837] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bc5e5ea9-4450-441f-ba81-c4375b0467a6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1986.020938] env[62824]: DEBUG oslo_vmware.api [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Waiting for the task: (returnval){ [ 1986.020938] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]521e64f9-9b40-9aaf-38ef-6c506a6411b5" [ 1986.020938] env[62824]: _type = "Task" [ 1986.020938] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1986.030939] env[62824]: DEBUG oslo_vmware.api [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]521e64f9-9b40-9aaf-38ef-6c506a6411b5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1986.117159] env[62824]: DEBUG oslo_vmware.api [None req-f73e7f3e-71a9-4684-9e15-596a02702690 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2145934, 'name': ReconfigVM_Task, 'duration_secs': 0.500245} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1986.117501] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-f73e7f3e-71a9-4684-9e15-596a02702690 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Reconfigured VM instance instance-00000020 to detach disk 2001 {{(pid=62824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1986.117723] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-f73e7f3e-71a9-4684-9e15-596a02702690 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1986.118034] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-93f5627d-47de-47a6-9723-3c49d146fdc5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1986.127177] env[62824]: DEBUG oslo_vmware.api [None req-f73e7f3e-71a9-4684-9e15-596a02702690 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for the task: (returnval){ [ 1986.127177] env[62824]: value = "task-2145937" [ 1986.127177] env[62824]: _type = "Task" [ 1986.127177] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1986.139898] env[62824]: DEBUG oslo_vmware.api [None req-f73e7f3e-71a9-4684-9e15-596a02702690 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2145937, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1986.186965] env[62824]: DEBUG oslo_vmware.api [None req-7a7b2101-f65d-4821-85ef-6e89be960d5e tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Task: {'id': task-2145935, 'name': PowerOffVM_Task, 'duration_secs': 0.188584} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1986.187294] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a7b2101-f65d-4821-85ef-6e89be960d5e tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] [instance: 5f1237ff-5841-41f1-8d77-df9fabaa62b6] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1986.187469] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-7a7b2101-f65d-4821-85ef-6e89be960d5e tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] [instance: 5f1237ff-5841-41f1-8d77-df9fabaa62b6] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1986.187730] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ae2fb231-8181-4783-8870-49638a1c2b98 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1986.195174] env[62824]: DEBUG nova.compute.manager [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02405b17-7022-4e2c-a357-415de88f63d0] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1986.208472] env[62824]: DEBUG nova.network.neutron [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2] Successfully created port: b90aaedc-aa4e-4f70-9365-1ee7a5bd49ee {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1986.307747] env[62824]: DEBUG nova.compute.manager [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1986.310268] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-7a7b2101-f65d-4821-85ef-6e89be960d5e tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] [instance: 5f1237ff-5841-41f1-8d77-df9fabaa62b6] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1986.310416] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-7a7b2101-f65d-4821-85ef-6e89be960d5e tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] [instance: 5f1237ff-5841-41f1-8d77-df9fabaa62b6] Deleting contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1986.314021] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-7a7b2101-f65d-4821-85ef-6e89be960d5e tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Deleting the datastore file [datastore1] 5f1237ff-5841-41f1-8d77-df9fabaa62b6 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1986.314021] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5e2af72b-6ab4-471c-ba7e-87ea1373c69d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1986.319285] env[62824]: DEBUG oslo_vmware.api [None req-7a7b2101-f65d-4821-85ef-6e89be960d5e tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Waiting for the task: (returnval){ [ 1986.319285] env[62824]: value = "task-2145939" [ 1986.319285] env[62824]: _type = "Task" [ 1986.319285] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1986.329102] env[62824]: DEBUG oslo_vmware.api [None req-7a7b2101-f65d-4821-85ef-6e89be960d5e tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Task: {'id': task-2145939, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1986.481768] env[62824]: DEBUG oslo_vmware.api [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Task: {'id': task-2145936, 'name': ReconfigVM_Task, 'duration_secs': 0.311664} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1986.481768] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: dc3317e4-6958-4df6-8e39-c89b4a5d861c] Reconfigured VM instance instance-00000026 to attach disk [datastore1] dc3317e4-6958-4df6-8e39-c89b4a5d861c/dc3317e4-6958-4df6-8e39-c89b4a5d861c.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1986.482423] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4d187b6e-cd85-4656-8e28-f88719702780 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1986.490040] env[62824]: DEBUG oslo_vmware.api [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Waiting for the task: (returnval){ [ 1986.490040] env[62824]: value = "task-2145940" [ 1986.490040] env[62824]: _type = "Task" [ 1986.490040] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1986.501853] env[62824]: DEBUG oslo_vmware.api [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Task: {'id': task-2145940, 'name': Rename_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1986.513719] env[62824]: DEBUG oslo_concurrency.lockutils [req-3617e7b3-c93f-4275-a221-adca0fc88221 req-8fab501a-06c0-47a2-8b2f-fc049116c618 service nova] Releasing lock "refresh_cache-7febb8b4-f984-4d79-a888-b2829f2a9df6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1986.530640] env[62824]: DEBUG oslo_vmware.api [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]521e64f9-9b40-9aaf-38ef-6c506a6411b5, 'name': SearchDatastore_Task, 'duration_secs': 0.010183} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1986.533936] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8b773a30-c3eb-496d-be7d-19b2d05c8ab3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1986.539662] env[62824]: DEBUG oslo_vmware.api [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Waiting for the task: (returnval){ [ 1986.539662] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52554de3-93d9-6778-7728-5fee942cef8d" [ 1986.539662] env[62824]: _type = "Task" [ 1986.539662] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1986.547840] env[62824]: DEBUG oslo_vmware.api [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52554de3-93d9-6778-7728-5fee942cef8d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1986.641852] env[62824]: DEBUG oslo_vmware.api [None req-f73e7f3e-71a9-4684-9e15-596a02702690 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2145937, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1986.715872] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1986.796552] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5178e5a-18ed-4dbb-a13d-e66e99a3f20e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1986.804714] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8bc1e1b-8e63-4690-8f90-d468e4e5ee95 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1986.845453] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d5acaf4-7409-4e7f-bc11-b03bcdbba0cd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1986.857145] env[62824]: DEBUG oslo_vmware.api [None req-7a7b2101-f65d-4821-85ef-6e89be960d5e tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Task: {'id': task-2145939, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.161527} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1986.857975] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac290931-efc0-4128-9b58-625351948e8f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1986.862133] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-7a7b2101-f65d-4821-85ef-6e89be960d5e tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1986.862322] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-7a7b2101-f65d-4821-85ef-6e89be960d5e tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] [instance: 5f1237ff-5841-41f1-8d77-df9fabaa62b6] Deleted contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1986.862511] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-7a7b2101-f65d-4821-85ef-6e89be960d5e tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] [instance: 5f1237ff-5841-41f1-8d77-df9fabaa62b6] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1986.862808] env[62824]: INFO nova.compute.manager [None req-7a7b2101-f65d-4821-85ef-6e89be960d5e tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] [instance: 5f1237ff-5841-41f1-8d77-df9fabaa62b6] Took 1.20 seconds to destroy the instance on the hypervisor. [ 1986.862976] env[62824]: DEBUG oslo.service.loopingcall [None req-7a7b2101-f65d-4821-85ef-6e89be960d5e tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1986.863540] env[62824]: DEBUG nova.compute.manager [-] [instance: 5f1237ff-5841-41f1-8d77-df9fabaa62b6] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1986.863666] env[62824]: DEBUG nova.network.neutron [-] [instance: 5f1237ff-5841-41f1-8d77-df9fabaa62b6] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1986.878059] env[62824]: DEBUG nova.compute.provider_tree [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1987.001526] env[62824]: DEBUG oslo_vmware.api [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Task: {'id': task-2145940, 'name': Rename_Task, 'duration_secs': 0.19145} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1987.001918] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: dc3317e4-6958-4df6-8e39-c89b4a5d861c] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1987.002226] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f7d8481f-0495-4542-878c-2adcc1fc9eff {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1987.021296] env[62824]: DEBUG oslo_vmware.api [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Waiting for the task: (returnval){ [ 1987.021296] env[62824]: value = "task-2145942" [ 1987.021296] env[62824]: _type = "Task" [ 1987.021296] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1987.031389] env[62824]: DEBUG oslo_vmware.api [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Task: {'id': task-2145942, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1987.051714] env[62824]: DEBUG oslo_vmware.api [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52554de3-93d9-6778-7728-5fee942cef8d, 'name': SearchDatastore_Task, 'duration_secs': 0.010631} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1987.052058] env[62824]: DEBUG oslo_concurrency.lockutils [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1987.052361] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 7febb8b4-f984-4d79-a888-b2829f2a9df6/7febb8b4-f984-4d79-a888-b2829f2a9df6.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1987.052676] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0b428796-156c-44a6-a9bf-7a73352fc163 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1987.062269] env[62824]: DEBUG oslo_vmware.api [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Waiting for the task: (returnval){ [ 1987.062269] env[62824]: value = "task-2145943" [ 1987.062269] env[62824]: _type = "Task" [ 1987.062269] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1987.073130] env[62824]: DEBUG oslo_vmware.api [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Task: {'id': task-2145943, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1987.140159] env[62824]: DEBUG oslo_vmware.api [None req-f73e7f3e-71a9-4684-9e15-596a02702690 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2145937, 'name': PowerOnVM_Task, 'duration_secs': 0.603203} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1987.140479] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-f73e7f3e-71a9-4684-9e15-596a02702690 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1987.140737] env[62824]: DEBUG nova.compute.manager [None req-f73e7f3e-71a9-4684-9e15-596a02702690 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1987.141671] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-397b7224-8754-4391-9715-1981f1c347f6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1987.343853] env[62824]: DEBUG nova.compute.manager [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1987.377898] env[62824]: DEBUG nova.virt.hardware [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1987.378279] env[62824]: DEBUG nova.virt.hardware [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1987.378459] env[62824]: DEBUG nova.virt.hardware [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1987.378690] env[62824]: DEBUG nova.virt.hardware [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1987.378848] env[62824]: DEBUG nova.virt.hardware [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1987.378983] env[62824]: DEBUG nova.virt.hardware [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1987.379213] env[62824]: DEBUG nova.virt.hardware [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1987.379376] env[62824]: DEBUG nova.virt.hardware [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1987.379610] env[62824]: DEBUG nova.virt.hardware [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1987.379814] env[62824]: DEBUG nova.virt.hardware [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1987.380180] env[62824]: DEBUG nova.virt.hardware [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1987.381286] env[62824]: DEBUG nova.scheduler.client.report [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1987.385384] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ac5cd1c-43ee-462e-a844-c8b9c344b593 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1987.397429] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cca524a-7920-4ca2-a7f1-b5da042759bb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1987.532565] env[62824]: DEBUG oslo_vmware.api [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Task: {'id': task-2145942, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1987.575145] env[62824]: DEBUG oslo_vmware.api [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Task: {'id': task-2145943, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.481002} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1987.575433] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 7febb8b4-f984-4d79-a888-b2829f2a9df6/7febb8b4-f984-4d79-a888-b2829f2a9df6.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1987.575649] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1987.575904] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a30ce2f8-adec-482e-837d-7bea55e10dbb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1987.585367] env[62824]: DEBUG oslo_vmware.api [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Waiting for the task: (returnval){ [ 1987.585367] env[62824]: value = "task-2145944" [ 1987.585367] env[62824]: _type = "Task" [ 1987.585367] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1987.595576] env[62824]: DEBUG oslo_vmware.api [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Task: {'id': task-2145944, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1987.688942] env[62824]: DEBUG nova.compute.manager [req-ae8162e8-43ce-44d1-840e-2c7148e6f662 req-f65015f3-0990-4f6a-8c9e-99dd8de4afb7 service nova] [instance: 5f1237ff-5841-41f1-8d77-df9fabaa62b6] Received event network-vif-deleted-660b16ae-ebd9-43c6-a3dd-3338c7b75f72 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1987.689260] env[62824]: INFO nova.compute.manager [req-ae8162e8-43ce-44d1-840e-2c7148e6f662 req-f65015f3-0990-4f6a-8c9e-99dd8de4afb7 service nova] [instance: 5f1237ff-5841-41f1-8d77-df9fabaa62b6] Neutron deleted interface 660b16ae-ebd9-43c6-a3dd-3338c7b75f72; detaching it from the instance and deleting it from the info cache [ 1987.689497] env[62824]: DEBUG nova.network.neutron [req-ae8162e8-43ce-44d1-840e-2c7148e6f662 req-f65015f3-0990-4f6a-8c9e-99dd8de4afb7 service nova] [instance: 5f1237ff-5841-41f1-8d77-df9fabaa62b6] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1987.766485] env[62824]: DEBUG nova.network.neutron [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2] Successfully updated port: b90aaedc-aa4e-4f70-9365-1ee7a5bd49ee {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1987.837126] env[62824]: DEBUG nova.compute.manager [req-39a0e6c9-bb6a-415b-8244-675d2251a701 req-ca54a6bb-90db-4416-81f6-800477ff4497 service nova] [instance: 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2] Received event network-vif-plugged-b90aaedc-aa4e-4f70-9365-1ee7a5bd49ee {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1987.837387] env[62824]: DEBUG oslo_concurrency.lockutils [req-39a0e6c9-bb6a-415b-8244-675d2251a701 req-ca54a6bb-90db-4416-81f6-800477ff4497 service nova] Acquiring lock "387b7f14-f3c1-43ed-8c4e-6de3ce9822e2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1987.837611] env[62824]: DEBUG oslo_concurrency.lockutils [req-39a0e6c9-bb6a-415b-8244-675d2251a701 req-ca54a6bb-90db-4416-81f6-800477ff4497 service nova] Lock "387b7f14-f3c1-43ed-8c4e-6de3ce9822e2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1987.837781] env[62824]: DEBUG oslo_concurrency.lockutils [req-39a0e6c9-bb6a-415b-8244-675d2251a701 req-ca54a6bb-90db-4416-81f6-800477ff4497 service nova] Lock "387b7f14-f3c1-43ed-8c4e-6de3ce9822e2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1987.837952] env[62824]: DEBUG nova.compute.manager [req-39a0e6c9-bb6a-415b-8244-675d2251a701 req-ca54a6bb-90db-4416-81f6-800477ff4497 service nova] [instance: 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2] No waiting events found dispatching network-vif-plugged-b90aaedc-aa4e-4f70-9365-1ee7a5bd49ee {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1987.838131] env[62824]: WARNING nova.compute.manager [req-39a0e6c9-bb6a-415b-8244-675d2251a701 req-ca54a6bb-90db-4416-81f6-800477ff4497 service nova] [instance: 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2] Received unexpected event network-vif-plugged-b90aaedc-aa4e-4f70-9365-1ee7a5bd49ee for instance with vm_state building and task_state spawning. [ 1987.889451] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.602s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1987.890374] env[62824]: DEBUG nova.compute.manager [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 8bb90a4d-93c9-4f54-b15e-48fe966e6c97] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1987.895022] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.856s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1987.895022] env[62824]: INFO nova.compute.claims [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 2afecec5-763d-4616-b690-41d3101cfc47] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1988.032978] env[62824]: DEBUG oslo_vmware.api [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Task: {'id': task-2145942, 'name': PowerOnVM_Task, 'duration_secs': 0.667681} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1988.036452] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: dc3317e4-6958-4df6-8e39-c89b4a5d861c] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1988.036661] env[62824]: INFO nova.compute.manager [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: dc3317e4-6958-4df6-8e39-c89b4a5d861c] Took 15.62 seconds to spawn the instance on the hypervisor. [ 1988.036865] env[62824]: DEBUG nova.compute.manager [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: dc3317e4-6958-4df6-8e39-c89b4a5d861c] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1988.037207] env[62824]: DEBUG nova.network.neutron [-] [instance: 5f1237ff-5841-41f1-8d77-df9fabaa62b6] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1988.039961] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2e537e1-85df-4b99-ad99-1f547768059a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1988.099177] env[62824]: DEBUG oslo_vmware.api [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Task: {'id': task-2145944, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.207327} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1988.103018] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1988.103018] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df86fd7b-36d1-401c-89dc-c2a4db884957 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1988.125476] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Reconfiguring VM instance instance-00000028 to attach disk [datastore1] 7febb8b4-f984-4d79-a888-b2829f2a9df6/7febb8b4-f984-4d79-a888-b2829f2a9df6.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1988.126089] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ccd7a896-a3f9-48aa-9469-9604e97daaf4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1988.147767] env[62824]: DEBUG oslo_vmware.api [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Waiting for the task: (returnval){ [ 1988.147767] env[62824]: value = "task-2145945" [ 1988.147767] env[62824]: _type = "Task" [ 1988.147767] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1988.156724] env[62824]: DEBUG oslo_vmware.api [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Task: {'id': task-2145945, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1988.193510] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-deeaa1cb-2d5f-4689-b303-7c67fbe398ff {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1988.204511] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecd46b1c-fb3e-4b0c-b2c7-b9d325af3f71 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1988.239963] env[62824]: DEBUG nova.compute.manager [req-ae8162e8-43ce-44d1-840e-2c7148e6f662 req-f65015f3-0990-4f6a-8c9e-99dd8de4afb7 service nova] [instance: 5f1237ff-5841-41f1-8d77-df9fabaa62b6] Detach interface failed, port_id=660b16ae-ebd9-43c6-a3dd-3338c7b75f72, reason: Instance 5f1237ff-5841-41f1-8d77-df9fabaa62b6 could not be found. {{(pid=62824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11482}} [ 1988.272155] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Acquiring lock "refresh_cache-387b7f14-f3c1-43ed-8c4e-6de3ce9822e2" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1988.272352] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Acquired lock "refresh_cache-387b7f14-f3c1-43ed-8c4e-6de3ce9822e2" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1988.272441] env[62824]: DEBUG nova.network.neutron [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1988.399779] env[62824]: DEBUG nova.compute.utils [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1988.401226] env[62824]: DEBUG nova.compute.manager [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 8bb90a4d-93c9-4f54-b15e-48fe966e6c97] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1988.401499] env[62824]: DEBUG nova.network.neutron [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 8bb90a4d-93c9-4f54-b15e-48fe966e6c97] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1988.448941] env[62824]: DEBUG nova.policy [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd8c161eab02941fabfcc430c451c682a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '31bcd298e89d49da977dda4cefd20432', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 1988.543034] env[62824]: INFO nova.compute.manager [-] [instance: 5f1237ff-5841-41f1-8d77-df9fabaa62b6] Took 1.68 seconds to deallocate network for instance. [ 1988.556460] env[62824]: INFO nova.compute.manager [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: dc3317e4-6958-4df6-8e39-c89b4a5d861c] Took 55.07 seconds to build instance. [ 1988.658897] env[62824]: DEBUG oslo_vmware.api [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Task: {'id': task-2145945, 'name': ReconfigVM_Task, 'duration_secs': 0.303046} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1988.659226] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Reconfigured VM instance instance-00000028 to attach disk [datastore1] 7febb8b4-f984-4d79-a888-b2829f2a9df6/7febb8b4-f984-4d79-a888-b2829f2a9df6.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1988.659861] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-233e03e2-d56b-49b8-b6f3-5ad1da2787e6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1988.669189] env[62824]: DEBUG oslo_vmware.api [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Waiting for the task: (returnval){ [ 1988.669189] env[62824]: value = "task-2145946" [ 1988.669189] env[62824]: _type = "Task" [ 1988.669189] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1988.679463] env[62824]: DEBUG oslo_vmware.api [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Task: {'id': task-2145946, 'name': Rename_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1988.706404] env[62824]: DEBUG nova.network.neutron [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 8bb90a4d-93c9-4f54-b15e-48fe966e6c97] Successfully created port: ec4e83ca-0312-4fc5-8c4f-475c5235bd4e {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1988.824074] env[62824]: DEBUG nova.network.neutron [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1988.907477] env[62824]: DEBUG nova.compute.manager [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 8bb90a4d-93c9-4f54-b15e-48fe966e6c97] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1989.049642] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7a7b2101-f65d-4821-85ef-6e89be960d5e tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1989.055045] env[62824]: DEBUG nova.network.neutron [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2] Updating instance_info_cache with network_info: [{"id": "b90aaedc-aa4e-4f70-9365-1ee7a5bd49ee", "address": "fa:16:3e:be:3f:38", "network": {"id": "b6403dc3-30ef-4c76-b521-380c179f345d", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.91", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "22c5c40c2e004babaa15896c89307303", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1002b79b-224e-41e3-a484-4245a767147a", "external-id": "nsx-vlan-transportzone-353", "segmentation_id": 353, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb90aaedc-aa", "ovs_interfaceid": "b90aaedc-aa4e-4f70-9365-1ee7a5bd49ee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1989.057856] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fcfcd374-cb5c-413d-86e5-4a73d7dcd4a6 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Lock "dc3317e4-6958-4df6-8e39-c89b4a5d861c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 104.676s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1989.179844] env[62824]: DEBUG oslo_vmware.api [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Task: {'id': task-2145946, 'name': Rename_Task, 'duration_secs': 0.153995} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1989.180152] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1989.180391] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b43de81a-87e6-4d93-84b7-1ae4a6c42898 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1989.189909] env[62824]: DEBUG oslo_vmware.api [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Waiting for the task: (returnval){ [ 1989.189909] env[62824]: value = "task-2145948" [ 1989.189909] env[62824]: _type = "Task" [ 1989.189909] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1989.199225] env[62824]: DEBUG oslo_vmware.api [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Task: {'id': task-2145948, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1989.433030] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a45c6037-0c17-465b-92d8-5ea713401202 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1989.440018] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-198d9da4-ad63-4447-8d71-18a51330f710 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1989.477108] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac08afd3-be64-4abc-a295-5162df0afa80 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1989.485436] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc50ec21-451c-4a63-9e03-f2897d6fb751 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1989.502009] env[62824]: DEBUG nova.compute.provider_tree [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1989.557416] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Releasing lock "refresh_cache-387b7f14-f3c1-43ed-8c4e-6de3ce9822e2" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1989.557751] env[62824]: DEBUG nova.compute.manager [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2] Instance network_info: |[{"id": "b90aaedc-aa4e-4f70-9365-1ee7a5bd49ee", "address": "fa:16:3e:be:3f:38", "network": {"id": "b6403dc3-30ef-4c76-b521-380c179f345d", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.91", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "22c5c40c2e004babaa15896c89307303", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1002b79b-224e-41e3-a484-4245a767147a", "external-id": "nsx-vlan-transportzone-353", "segmentation_id": 353, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb90aaedc-aa", "ovs_interfaceid": "b90aaedc-aa4e-4f70-9365-1ee7a5bd49ee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1989.558409] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:be:3f:38', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1002b79b-224e-41e3-a484-4245a767147a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b90aaedc-aa4e-4f70-9365-1ee7a5bd49ee', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1989.567146] env[62824]: DEBUG oslo.service.loopingcall [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1989.567494] env[62824]: DEBUG nova.compute.manager [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: c65ccd20-b682-420a-9c1a-47a45959197e] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1989.570025] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1989.570245] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-21d2763c-267b-4fd7-ad86-6633e0710eb0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1989.593596] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1989.593596] env[62824]: value = "task-2145949" [ 1989.593596] env[62824]: _type = "Task" [ 1989.593596] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1989.609105] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145949, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1989.701585] env[62824]: DEBUG oslo_vmware.api [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Task: {'id': task-2145948, 'name': PowerOnVM_Task, 'duration_secs': 0.459029} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1989.702082] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1989.702377] env[62824]: INFO nova.compute.manager [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Took 7.30 seconds to spawn the instance on the hypervisor. [ 1989.702601] env[62824]: DEBUG nova.compute.manager [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1989.703564] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97cda24f-9319-42e4-a06a-a751802c627d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1989.920873] env[62824]: DEBUG nova.compute.manager [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 8bb90a4d-93c9-4f54-b15e-48fe966e6c97] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1989.953916] env[62824]: DEBUG nova.virt.hardware [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1989.954056] env[62824]: DEBUG nova.virt.hardware [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1989.954229] env[62824]: DEBUG nova.virt.hardware [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1989.954415] env[62824]: DEBUG nova.virt.hardware [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1989.954560] env[62824]: DEBUG nova.virt.hardware [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1989.954709] env[62824]: DEBUG nova.virt.hardware [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1989.954939] env[62824]: DEBUG nova.virt.hardware [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1989.955459] env[62824]: DEBUG nova.virt.hardware [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1989.955672] env[62824]: DEBUG nova.virt.hardware [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1989.955848] env[62824]: DEBUG nova.virt.hardware [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1989.956067] env[62824]: DEBUG nova.virt.hardware [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1989.957152] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02541380-21c2-4061-856e-d3bf99622c01 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1989.962709] env[62824]: DEBUG nova.compute.manager [req-fa966d88-b48e-404d-8cd4-306563db25d5 req-ef814924-de49-432c-889e-1b8de336d48d service nova] [instance: 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2] Received event network-changed-b90aaedc-aa4e-4f70-9365-1ee7a5bd49ee {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1989.962709] env[62824]: DEBUG nova.compute.manager [req-fa966d88-b48e-404d-8cd4-306563db25d5 req-ef814924-de49-432c-889e-1b8de336d48d service nova] [instance: 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2] Refreshing instance network info cache due to event network-changed-b90aaedc-aa4e-4f70-9365-1ee7a5bd49ee. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1989.963378] env[62824]: DEBUG oslo_concurrency.lockutils [req-fa966d88-b48e-404d-8cd4-306563db25d5 req-ef814924-de49-432c-889e-1b8de336d48d service nova] Acquiring lock "refresh_cache-387b7f14-f3c1-43ed-8c4e-6de3ce9822e2" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1989.963378] env[62824]: DEBUG oslo_concurrency.lockutils [req-fa966d88-b48e-404d-8cd4-306563db25d5 req-ef814924-de49-432c-889e-1b8de336d48d service nova] Acquired lock "refresh_cache-387b7f14-f3c1-43ed-8c4e-6de3ce9822e2" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1989.963719] env[62824]: DEBUG nova.network.neutron [req-fa966d88-b48e-404d-8cd4-306563db25d5 req-ef814924-de49-432c-889e-1b8de336d48d service nova] [instance: 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2] Refreshing network info cache for port b90aaedc-aa4e-4f70-9365-1ee7a5bd49ee {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1989.971122] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d35c359-e848-46de-9c5e-f6199239316e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1990.004307] env[62824]: DEBUG nova.scheduler.client.report [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1990.105842] env[62824]: DEBUG oslo_concurrency.lockutils [None req-870c96bf-3657-45b9-96d5-c5182b858ac4 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Acquiring lock "dc3317e4-6958-4df6-8e39-c89b4a5d861c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1990.106167] env[62824]: DEBUG oslo_concurrency.lockutils [None req-870c96bf-3657-45b9-96d5-c5182b858ac4 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Lock "dc3317e4-6958-4df6-8e39-c89b4a5d861c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1990.106381] env[62824]: DEBUG oslo_concurrency.lockutils [None req-870c96bf-3657-45b9-96d5-c5182b858ac4 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Acquiring lock "dc3317e4-6958-4df6-8e39-c89b4a5d861c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1990.106564] env[62824]: DEBUG oslo_concurrency.lockutils [None req-870c96bf-3657-45b9-96d5-c5182b858ac4 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Lock "dc3317e4-6958-4df6-8e39-c89b4a5d861c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1990.106729] env[62824]: DEBUG oslo_concurrency.lockutils [None req-870c96bf-3657-45b9-96d5-c5182b858ac4 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Lock "dc3317e4-6958-4df6-8e39-c89b4a5d861c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1990.108155] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145949, 'name': CreateVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1990.109373] env[62824]: DEBUG oslo_concurrency.lockutils [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1990.109783] env[62824]: INFO nova.compute.manager [None req-870c96bf-3657-45b9-96d5-c5182b858ac4 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: dc3317e4-6958-4df6-8e39-c89b4a5d861c] Terminating instance [ 1990.223717] env[62824]: INFO nova.compute.manager [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Took 50.75 seconds to build instance. [ 1990.260161] env[62824]: DEBUG nova.network.neutron [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 8bb90a4d-93c9-4f54-b15e-48fe966e6c97] Successfully updated port: ec4e83ca-0312-4fc5-8c4f-475c5235bd4e {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1990.509771] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.617s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1990.512084] env[62824]: DEBUG nova.compute.manager [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 2afecec5-763d-4616-b690-41d3101cfc47] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1990.512944] env[62824]: DEBUG oslo_concurrency.lockutils [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.837s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1990.514262] env[62824]: INFO nova.compute.claims [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 41f068d3-2c8d-46f0-8d84-78531d28b0dd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1990.607998] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145949, 'name': CreateVM_Task, 'duration_secs': 0.61081} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1990.608235] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1990.609176] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1990.609416] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1990.609874] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1990.613737] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0be3b9ee-b7e8-4ac2-8e95-091dedeb6fef {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1990.616581] env[62824]: DEBUG nova.compute.manager [None req-870c96bf-3657-45b9-96d5-c5182b858ac4 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: dc3317e4-6958-4df6-8e39-c89b4a5d861c] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1990.616960] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-870c96bf-3657-45b9-96d5-c5182b858ac4 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: dc3317e4-6958-4df6-8e39-c89b4a5d861c] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1990.618529] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b54f294-d8db-4a72-8b5c-7209adea798a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1990.628631] env[62824]: DEBUG oslo_vmware.api [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Waiting for the task: (returnval){ [ 1990.628631] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5261a8d1-e8f6-2caf-d0d3-11388dcd9dd5" [ 1990.628631] env[62824]: _type = "Task" [ 1990.628631] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1990.631775] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-870c96bf-3657-45b9-96d5-c5182b858ac4 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: dc3317e4-6958-4df6-8e39-c89b4a5d861c] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1990.635713] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-72059942-ffe4-4add-abfa-fd54544badc7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1990.644174] env[62824]: DEBUG oslo_vmware.api [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5261a8d1-e8f6-2caf-d0d3-11388dcd9dd5, 'name': SearchDatastore_Task, 'duration_secs': 0.013839} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1990.644476] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1990.644709] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1990.644957] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1990.645655] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1990.645655] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1990.646127] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9f4a05e9-5b61-45d8-80b5-849ba85d90f6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1990.648782] env[62824]: DEBUG oslo_vmware.api [None req-870c96bf-3657-45b9-96d5-c5182b858ac4 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Waiting for the task: (returnval){ [ 1990.648782] env[62824]: value = "task-2145950" [ 1990.648782] env[62824]: _type = "Task" [ 1990.648782] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1990.660130] env[62824]: DEBUG oslo_vmware.api [None req-870c96bf-3657-45b9-96d5-c5182b858ac4 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Task: {'id': task-2145950, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1990.660749] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1990.660922] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1990.662028] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-afc34143-b848-4a78-ba8a-5ee849781b6e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1990.669398] env[62824]: DEBUG oslo_vmware.api [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Waiting for the task: (returnval){ [ 1990.669398] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]529e7c35-dced-3899-d4f7-45b88ea277f2" [ 1990.669398] env[62824]: _type = "Task" [ 1990.669398] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1990.679045] env[62824]: DEBUG oslo_vmware.api [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]529e7c35-dced-3899-d4f7-45b88ea277f2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1990.724948] env[62824]: DEBUG oslo_concurrency.lockutils [None req-066874d1-79cd-4166-beff-eae64e128661 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Lock "7febb8b4-f984-4d79-a888-b2829f2a9df6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 101.924s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1990.764328] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Acquiring lock "refresh_cache-8bb90a4d-93c9-4f54-b15e-48fe966e6c97" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1990.764715] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Acquired lock "refresh_cache-8bb90a4d-93c9-4f54-b15e-48fe966e6c97" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1990.765436] env[62824]: DEBUG nova.network.neutron [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 8bb90a4d-93c9-4f54-b15e-48fe966e6c97] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1990.768989] env[62824]: DEBUG nova.network.neutron [req-fa966d88-b48e-404d-8cd4-306563db25d5 req-ef814924-de49-432c-889e-1b8de336d48d service nova] [instance: 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2] Updated VIF entry in instance network info cache for port b90aaedc-aa4e-4f70-9365-1ee7a5bd49ee. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1990.769593] env[62824]: DEBUG nova.network.neutron [req-fa966d88-b48e-404d-8cd4-306563db25d5 req-ef814924-de49-432c-889e-1b8de336d48d service nova] [instance: 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2] Updating instance_info_cache with network_info: [{"id": "b90aaedc-aa4e-4f70-9365-1ee7a5bd49ee", "address": "fa:16:3e:be:3f:38", "network": {"id": "b6403dc3-30ef-4c76-b521-380c179f345d", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.91", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "22c5c40c2e004babaa15896c89307303", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1002b79b-224e-41e3-a484-4245a767147a", "external-id": "nsx-vlan-transportzone-353", "segmentation_id": 353, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb90aaedc-aa", "ovs_interfaceid": "b90aaedc-aa4e-4f70-9365-1ee7a5bd49ee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1991.019160] env[62824]: DEBUG nova.compute.utils [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1991.022473] env[62824]: DEBUG nova.compute.manager [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 2afecec5-763d-4616-b690-41d3101cfc47] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1991.022638] env[62824]: DEBUG nova.network.neutron [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 2afecec5-763d-4616-b690-41d3101cfc47] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1991.096745] env[62824]: DEBUG nova.policy [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd8c161eab02941fabfcc430c451c682a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '31bcd298e89d49da977dda4cefd20432', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 1991.160825] env[62824]: DEBUG oslo_vmware.api [None req-870c96bf-3657-45b9-96d5-c5182b858ac4 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Task: {'id': task-2145950, 'name': PowerOffVM_Task, 'duration_secs': 0.206628} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1991.161201] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-870c96bf-3657-45b9-96d5-c5182b858ac4 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: dc3317e4-6958-4df6-8e39-c89b4a5d861c] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1991.161406] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-870c96bf-3657-45b9-96d5-c5182b858ac4 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: dc3317e4-6958-4df6-8e39-c89b4a5d861c] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1991.161668] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-90692d69-e61b-4ba8-b6c3-20a45cb5dbe2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1991.181203] env[62824]: DEBUG oslo_vmware.api [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]529e7c35-dced-3899-d4f7-45b88ea277f2, 'name': SearchDatastore_Task, 'duration_secs': 0.019269} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1991.182054] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-78b45d99-6f88-4488-bca9-ca6eae611f58 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1991.192055] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Acquiring lock "bbfcb2e3-9326-4548-b15b-e054cbfd192e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1991.192055] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Lock "bbfcb2e3-9326-4548-b15b-e054cbfd192e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1991.195731] env[62824]: DEBUG oslo_vmware.api [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Waiting for the task: (returnval){ [ 1991.195731] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52c712f4-0bbb-530f-04aa-997f19b156f1" [ 1991.195731] env[62824]: _type = "Task" [ 1991.195731] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1991.205181] env[62824]: DEBUG oslo_vmware.api [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52c712f4-0bbb-530f-04aa-997f19b156f1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1991.228571] env[62824]: DEBUG nova.compute.manager [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] [instance: 4423d172-acdd-4c69-b3b8-ff166e1b8548] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1991.278335] env[62824]: DEBUG oslo_concurrency.lockutils [req-fa966d88-b48e-404d-8cd4-306563db25d5 req-ef814924-de49-432c-889e-1b8de336d48d service nova] Releasing lock "refresh_cache-387b7f14-f3c1-43ed-8c4e-6de3ce9822e2" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1991.317526] env[62824]: DEBUG nova.network.neutron [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 8bb90a4d-93c9-4f54-b15e-48fe966e6c97] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1991.490450] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-870c96bf-3657-45b9-96d5-c5182b858ac4 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: dc3317e4-6958-4df6-8e39-c89b4a5d861c] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1991.493402] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-870c96bf-3657-45b9-96d5-c5182b858ac4 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: dc3317e4-6958-4df6-8e39-c89b4a5d861c] Deleting contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1991.493402] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-870c96bf-3657-45b9-96d5-c5182b858ac4 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Deleting the datastore file [datastore1] dc3317e4-6958-4df6-8e39-c89b4a5d861c {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1991.493402] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ff0d0574-96ab-4d1e-83da-3b6c30d01c59 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1991.503020] env[62824]: DEBUG oslo_vmware.api [None req-870c96bf-3657-45b9-96d5-c5182b858ac4 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Waiting for the task: (returnval){ [ 1991.503020] env[62824]: value = "task-2145952" [ 1991.503020] env[62824]: _type = "Task" [ 1991.503020] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1991.511367] env[62824]: DEBUG oslo_vmware.api [None req-870c96bf-3657-45b9-96d5-c5182b858ac4 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Task: {'id': task-2145952, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1991.515299] env[62824]: DEBUG nova.network.neutron [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 8bb90a4d-93c9-4f54-b15e-48fe966e6c97] Updating instance_info_cache with network_info: [{"id": "ec4e83ca-0312-4fc5-8c4f-475c5235bd4e", "address": "fa:16:3e:27:78:10", "network": {"id": "f099da67-9797-4509-a5fe-7c1d5d8560d5", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-13908051-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "31bcd298e89d49da977dda4cefd20432", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "023d6500-887e-4dc4-bec5-06b40450d9c0", "external-id": "nsx-vlan-transportzone-108", "segmentation_id": 108, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapec4e83ca-03", "ovs_interfaceid": "ec4e83ca-0312-4fc5-8c4f-475c5235bd4e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1991.528634] env[62824]: DEBUG nova.compute.manager [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 2afecec5-763d-4616-b690-41d3101cfc47] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1991.634246] env[62824]: DEBUG nova.network.neutron [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 2afecec5-763d-4616-b690-41d3101cfc47] Successfully created port: 3d71b7f8-4ccd-4ae9-87c1-c4f03a8a22ed {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1991.712445] env[62824]: DEBUG oslo_vmware.api [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52c712f4-0bbb-530f-04aa-997f19b156f1, 'name': SearchDatastore_Task, 'duration_secs': 0.023744} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1991.716931] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1991.716931] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2/387b7f14-f3c1-43ed-8c4e-6de3ce9822e2.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1991.719722] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-81d8ebf7-33fa-470f-88f1-6835543ceb31 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1991.729134] env[62824]: DEBUG oslo_vmware.api [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Waiting for the task: (returnval){ [ 1991.729134] env[62824]: value = "task-2145953" [ 1991.729134] env[62824]: _type = "Task" [ 1991.729134] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1991.739194] env[62824]: DEBUG oslo_vmware.api [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2145953, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1991.769260] env[62824]: DEBUG oslo_concurrency.lockutils [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1992.016345] env[62824]: DEBUG oslo_vmware.api [None req-870c96bf-3657-45b9-96d5-c5182b858ac4 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Task: {'id': task-2145952, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.321473} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1992.017031] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Releasing lock "refresh_cache-8bb90a4d-93c9-4f54-b15e-48fe966e6c97" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1992.017228] env[62824]: DEBUG nova.compute.manager [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 8bb90a4d-93c9-4f54-b15e-48fe966e6c97] Instance network_info: |[{"id": "ec4e83ca-0312-4fc5-8c4f-475c5235bd4e", "address": "fa:16:3e:27:78:10", "network": {"id": "f099da67-9797-4509-a5fe-7c1d5d8560d5", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-13908051-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "31bcd298e89d49da977dda4cefd20432", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "023d6500-887e-4dc4-bec5-06b40450d9c0", "external-id": "nsx-vlan-transportzone-108", "segmentation_id": 108, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapec4e83ca-03", "ovs_interfaceid": "ec4e83ca-0312-4fc5-8c4f-475c5235bd4e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1992.017593] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-870c96bf-3657-45b9-96d5-c5182b858ac4 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1992.017854] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-870c96bf-3657-45b9-96d5-c5182b858ac4 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: dc3317e4-6958-4df6-8e39-c89b4a5d861c] Deleted contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1992.018114] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-870c96bf-3657-45b9-96d5-c5182b858ac4 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: dc3317e4-6958-4df6-8e39-c89b4a5d861c] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1992.018373] env[62824]: INFO nova.compute.manager [None req-870c96bf-3657-45b9-96d5-c5182b858ac4 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] [instance: dc3317e4-6958-4df6-8e39-c89b4a5d861c] Took 1.40 seconds to destroy the instance on the hypervisor. [ 1992.018696] env[62824]: DEBUG oslo.service.loopingcall [None req-870c96bf-3657-45b9-96d5-c5182b858ac4 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1992.020178] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 8bb90a4d-93c9-4f54-b15e-48fe966e6c97] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:27:78:10', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '023d6500-887e-4dc4-bec5-06b40450d9c0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ec4e83ca-0312-4fc5-8c4f-475c5235bd4e', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1992.028036] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Creating folder: Project (31bcd298e89d49da977dda4cefd20432). Parent ref: group-v438503. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1992.028474] env[62824]: DEBUG nova.compute.manager [-] [instance: dc3317e4-6958-4df6-8e39-c89b4a5d861c] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1992.028635] env[62824]: DEBUG nova.network.neutron [-] [instance: dc3317e4-6958-4df6-8e39-c89b4a5d861c] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1992.034032] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0b376583-bbc4-453a-a854-b97fb1678b00 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1992.060021] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Created folder: Project (31bcd298e89d49da977dda4cefd20432) in parent group-v438503. [ 1992.060264] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Creating folder: Instances. Parent ref: group-v438599. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1992.060762] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-20638be7-e353-4d5b-b812-bb924cd949dc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1992.078931] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Created folder: Instances in parent group-v438599. [ 1992.079215] env[62824]: DEBUG oslo.service.loopingcall [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1992.080558] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8bb90a4d-93c9-4f54-b15e-48fe966e6c97] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1992.080558] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7e564adc-5432-4041-aa67-78437353b4b9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1992.108264] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1992.108264] env[62824]: value = "task-2145956" [ 1992.108264] env[62824]: _type = "Task" [ 1992.108264] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1992.118052] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145956, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1992.236226] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-244e0aa7-fb5a-40f2-8b15-e7b37c382c0a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1992.246211] env[62824]: DEBUG oslo_vmware.api [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2145953, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1992.249341] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0a958aa-f2c5-42e2-b47f-f5594462ec57 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1992.298246] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bb51aea-9aea-462c-afff-d9720a530765 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1992.308692] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-607fb2c6-17d4-4cf9-8394-409a3791f0d6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1992.314210] env[62824]: DEBUG nova.compute.manager [req-36eaafd2-8650-4268-bbb7-f0e966be0b7b req-4a2e595b-9a31-44db-8960-47b23f6a3ee4 service nova] [instance: 8bb90a4d-93c9-4f54-b15e-48fe966e6c97] Received event network-vif-plugged-ec4e83ca-0312-4fc5-8c4f-475c5235bd4e {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1992.314397] env[62824]: DEBUG oslo_concurrency.lockutils [req-36eaafd2-8650-4268-bbb7-f0e966be0b7b req-4a2e595b-9a31-44db-8960-47b23f6a3ee4 service nova] Acquiring lock "8bb90a4d-93c9-4f54-b15e-48fe966e6c97-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1992.314615] env[62824]: DEBUG oslo_concurrency.lockutils [req-36eaafd2-8650-4268-bbb7-f0e966be0b7b req-4a2e595b-9a31-44db-8960-47b23f6a3ee4 service nova] Lock "8bb90a4d-93c9-4f54-b15e-48fe966e6c97-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1992.314785] env[62824]: DEBUG oslo_concurrency.lockutils [req-36eaafd2-8650-4268-bbb7-f0e966be0b7b req-4a2e595b-9a31-44db-8960-47b23f6a3ee4 service nova] Lock "8bb90a4d-93c9-4f54-b15e-48fe966e6c97-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1992.314972] env[62824]: DEBUG nova.compute.manager [req-36eaafd2-8650-4268-bbb7-f0e966be0b7b req-4a2e595b-9a31-44db-8960-47b23f6a3ee4 service nova] [instance: 8bb90a4d-93c9-4f54-b15e-48fe966e6c97] No waiting events found dispatching network-vif-plugged-ec4e83ca-0312-4fc5-8c4f-475c5235bd4e {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1992.315166] env[62824]: WARNING nova.compute.manager [req-36eaafd2-8650-4268-bbb7-f0e966be0b7b req-4a2e595b-9a31-44db-8960-47b23f6a3ee4 service nova] [instance: 8bb90a4d-93c9-4f54-b15e-48fe966e6c97] Received unexpected event network-vif-plugged-ec4e83ca-0312-4fc5-8c4f-475c5235bd4e for instance with vm_state building and task_state spawning. [ 1992.315331] env[62824]: DEBUG nova.compute.manager [req-36eaafd2-8650-4268-bbb7-f0e966be0b7b req-4a2e595b-9a31-44db-8960-47b23f6a3ee4 service nova] [instance: 8bb90a4d-93c9-4f54-b15e-48fe966e6c97] Received event network-changed-ec4e83ca-0312-4fc5-8c4f-475c5235bd4e {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1992.315479] env[62824]: DEBUG nova.compute.manager [req-36eaafd2-8650-4268-bbb7-f0e966be0b7b req-4a2e595b-9a31-44db-8960-47b23f6a3ee4 service nova] [instance: 8bb90a4d-93c9-4f54-b15e-48fe966e6c97] Refreshing instance network info cache due to event network-changed-ec4e83ca-0312-4fc5-8c4f-475c5235bd4e. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1992.315654] env[62824]: DEBUG oslo_concurrency.lockutils [req-36eaafd2-8650-4268-bbb7-f0e966be0b7b req-4a2e595b-9a31-44db-8960-47b23f6a3ee4 service nova] Acquiring lock "refresh_cache-8bb90a4d-93c9-4f54-b15e-48fe966e6c97" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1992.315799] env[62824]: DEBUG oslo_concurrency.lockutils [req-36eaafd2-8650-4268-bbb7-f0e966be0b7b req-4a2e595b-9a31-44db-8960-47b23f6a3ee4 service nova] Acquired lock "refresh_cache-8bb90a4d-93c9-4f54-b15e-48fe966e6c97" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1992.315937] env[62824]: DEBUG nova.network.neutron [req-36eaafd2-8650-4268-bbb7-f0e966be0b7b req-4a2e595b-9a31-44db-8960-47b23f6a3ee4 service nova] [instance: 8bb90a4d-93c9-4f54-b15e-48fe966e6c97] Refreshing network info cache for port ec4e83ca-0312-4fc5-8c4f-475c5235bd4e {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1992.321545] env[62824]: DEBUG oslo_concurrency.lockutils [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Acquiring lock "2f315f9e-94e4-47ef-9503-ee92b59e5452" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1992.322489] env[62824]: DEBUG oslo_concurrency.lockutils [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Lock "2f315f9e-94e4-47ef-9503-ee92b59e5452" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1992.333296] env[62824]: DEBUG nova.compute.provider_tree [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1992.542312] env[62824]: DEBUG nova.compute.manager [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 2afecec5-763d-4616-b690-41d3101cfc47] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1992.547214] env[62824]: DEBUG nova.compute.manager [req-25943042-d3cf-463c-a185-221cd6157002 req-73f124c0-16ce-4c33-aad2-c4c81f36398a service nova] [instance: dc3317e4-6958-4df6-8e39-c89b4a5d861c] Received event network-vif-deleted-ee0328da-cfb9-4f03-94fc-1acb84f8c553 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1992.547414] env[62824]: INFO nova.compute.manager [req-25943042-d3cf-463c-a185-221cd6157002 req-73f124c0-16ce-4c33-aad2-c4c81f36398a service nova] [instance: dc3317e4-6958-4df6-8e39-c89b4a5d861c] Neutron deleted interface ee0328da-cfb9-4f03-94fc-1acb84f8c553; detaching it from the instance and deleting it from the info cache [ 1992.547647] env[62824]: DEBUG nova.network.neutron [req-25943042-d3cf-463c-a185-221cd6157002 req-73f124c0-16ce-4c33-aad2-c4c81f36398a service nova] [instance: dc3317e4-6958-4df6-8e39-c89b4a5d861c] Updating instance_info_cache with network_info: [{"id": "4c03a7c0-a685-4fe9-9dce-635f6dda43b0", "address": "fa:16:3e:3a:30:1d", "network": {"id": "a1dc1a68-2912-4ec5-8bc2-e7294c0563c1", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1056165949", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.118", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "4cb963b19c1549ab8a11e5a77e4f202d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "650f7968-4522-4ba5-8304-1b9949951ed7", "external-id": "nsx-vlan-transportzone-568", "segmentation_id": 568, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c03a7c0-a6", "ovs_interfaceid": "4c03a7c0-a685-4fe9-9dce-635f6dda43b0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1992.578351] env[62824]: DEBUG nova.virt.hardware [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1992.578664] env[62824]: DEBUG nova.virt.hardware [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1992.578847] env[62824]: DEBUG nova.virt.hardware [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1992.579046] env[62824]: DEBUG nova.virt.hardware [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1992.579201] env[62824]: DEBUG nova.virt.hardware [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1992.579350] env[62824]: DEBUG nova.virt.hardware [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1992.579797] env[62824]: DEBUG nova.virt.hardware [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1992.579972] env[62824]: DEBUG nova.virt.hardware [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1992.580201] env[62824]: DEBUG nova.virt.hardware [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1992.580566] env[62824]: DEBUG nova.virt.hardware [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1992.580771] env[62824]: DEBUG nova.virt.hardware [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1992.582216] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1df2797-1e17-41fe-9068-08bdf60258a8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1992.592667] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0825643e-2bc2-443a-99fe-646c082b2547 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1992.618484] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145956, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1992.741990] env[62824]: DEBUG oslo_vmware.api [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2145953, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.55414} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1992.741990] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2/387b7f14-f3c1-43ed-8c4e-6de3ce9822e2.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1992.741990] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1992.741990] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f316b7fc-60f7-4c81-949c-f6edbc7ca7fa {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1992.751441] env[62824]: DEBUG oslo_vmware.api [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Waiting for the task: (returnval){ [ 1992.751441] env[62824]: value = "task-2145957" [ 1992.751441] env[62824]: _type = "Task" [ 1992.751441] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1992.762259] env[62824]: DEBUG oslo_vmware.api [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2145957, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1992.836618] env[62824]: DEBUG nova.scheduler.client.report [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1993.051103] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8988d395-5029-4495-8611-5cb1c2c06721 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1993.064273] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9852a729-8e49-4063-8deb-648d9478b671 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1993.102541] env[62824]: DEBUG nova.compute.manager [req-25943042-d3cf-463c-a185-221cd6157002 req-73f124c0-16ce-4c33-aad2-c4c81f36398a service nova] [instance: dc3317e4-6958-4df6-8e39-c89b4a5d861c] Detach interface failed, port_id=ee0328da-cfb9-4f03-94fc-1acb84f8c553, reason: Instance dc3317e4-6958-4df6-8e39-c89b4a5d861c could not be found. {{(pid=62824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11482}} [ 1993.118385] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145956, 'name': CreateVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1993.127840] env[62824]: DEBUG nova.network.neutron [req-36eaafd2-8650-4268-bbb7-f0e966be0b7b req-4a2e595b-9a31-44db-8960-47b23f6a3ee4 service nova] [instance: 8bb90a4d-93c9-4f54-b15e-48fe966e6c97] Updated VIF entry in instance network info cache for port ec4e83ca-0312-4fc5-8c4f-475c5235bd4e. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1993.128209] env[62824]: DEBUG nova.network.neutron [req-36eaafd2-8650-4268-bbb7-f0e966be0b7b req-4a2e595b-9a31-44db-8960-47b23f6a3ee4 service nova] [instance: 8bb90a4d-93c9-4f54-b15e-48fe966e6c97] Updating instance_info_cache with network_info: [{"id": "ec4e83ca-0312-4fc5-8c4f-475c5235bd4e", "address": "fa:16:3e:27:78:10", "network": {"id": "f099da67-9797-4509-a5fe-7c1d5d8560d5", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-13908051-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "31bcd298e89d49da977dda4cefd20432", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "023d6500-887e-4dc4-bec5-06b40450d9c0", "external-id": "nsx-vlan-transportzone-108", "segmentation_id": 108, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapec4e83ca-03", "ovs_interfaceid": "ec4e83ca-0312-4fc5-8c4f-475c5235bd4e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1993.250389] env[62824]: DEBUG nova.network.neutron [-] [instance: dc3317e4-6958-4df6-8e39-c89b4a5d861c] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1993.267556] env[62824]: DEBUG oslo_vmware.api [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2145957, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.165875} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1993.267930] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1993.268738] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-371f80d6-b9f6-4f2a-abf5-dcd8ed46fb17 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1993.291524] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2] Reconfiguring VM instance instance-00000029 to attach disk [datastore2] 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2/387b7f14-f3c1-43ed-8c4e-6de3ce9822e2.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1993.292070] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8494dac4-7a40-4678-9a30-7fe974878ab3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1993.306894] env[62824]: DEBUG nova.network.neutron [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 2afecec5-763d-4616-b690-41d3101cfc47] Successfully updated port: 3d71b7f8-4ccd-4ae9-87c1-c4f03a8a22ed {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1993.314144] env[62824]: DEBUG oslo_vmware.api [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Waiting for the task: (returnval){ [ 1993.314144] env[62824]: value = "task-2145958" [ 1993.314144] env[62824]: _type = "Task" [ 1993.314144] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1993.323328] env[62824]: DEBUG oslo_vmware.api [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2145958, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1993.344047] env[62824]: DEBUG oslo_concurrency.lockutils [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.831s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1993.345071] env[62824]: DEBUG nova.compute.manager [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 41f068d3-2c8d-46f0-8d84-78531d28b0dd] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1993.347649] env[62824]: DEBUG oslo_concurrency.lockutils [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.944s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1993.349355] env[62824]: INFO nova.compute.claims [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1993.621214] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145956, 'name': CreateVM_Task, 'duration_secs': 1.205736} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1993.621394] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8bb90a4d-93c9-4f54-b15e-48fe966e6c97] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1993.622109] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1993.622259] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1993.622588] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1993.622835] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cd61c5f7-9bb3-47bf-adf2-ec86b75553f5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1993.627398] env[62824]: DEBUG oslo_vmware.api [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Waiting for the task: (returnval){ [ 1993.627398] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52f3f614-5ad6-1c70-9139-0336c2e18b19" [ 1993.627398] env[62824]: _type = "Task" [ 1993.627398] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1993.630862] env[62824]: DEBUG oslo_concurrency.lockutils [req-36eaafd2-8650-4268-bbb7-f0e966be0b7b req-4a2e595b-9a31-44db-8960-47b23f6a3ee4 service nova] Releasing lock "refresh_cache-8bb90a4d-93c9-4f54-b15e-48fe966e6c97" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1993.631085] env[62824]: DEBUG nova.compute.manager [req-36eaafd2-8650-4268-bbb7-f0e966be0b7b req-4a2e595b-9a31-44db-8960-47b23f6a3ee4 service nova] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Received event network-changed-b27bda53-5f70-40ff-bb04-9b22f2241093 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1993.631260] env[62824]: DEBUG nova.compute.manager [req-36eaafd2-8650-4268-bbb7-f0e966be0b7b req-4a2e595b-9a31-44db-8960-47b23f6a3ee4 service nova] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Refreshing instance network info cache due to event network-changed-b27bda53-5f70-40ff-bb04-9b22f2241093. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1993.631460] env[62824]: DEBUG oslo_concurrency.lockutils [req-36eaafd2-8650-4268-bbb7-f0e966be0b7b req-4a2e595b-9a31-44db-8960-47b23f6a3ee4 service nova] Acquiring lock "refresh_cache-7febb8b4-f984-4d79-a888-b2829f2a9df6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1993.631596] env[62824]: DEBUG oslo_concurrency.lockutils [req-36eaafd2-8650-4268-bbb7-f0e966be0b7b req-4a2e595b-9a31-44db-8960-47b23f6a3ee4 service nova] Acquired lock "refresh_cache-7febb8b4-f984-4d79-a888-b2829f2a9df6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1993.631752] env[62824]: DEBUG nova.network.neutron [req-36eaafd2-8650-4268-bbb7-f0e966be0b7b req-4a2e595b-9a31-44db-8960-47b23f6a3ee4 service nova] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Refreshing network info cache for port b27bda53-5f70-40ff-bb04-9b22f2241093 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1993.638093] env[62824]: DEBUG oslo_vmware.api [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52f3f614-5ad6-1c70-9139-0336c2e18b19, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1993.753180] env[62824]: INFO nova.compute.manager [-] [instance: dc3317e4-6958-4df6-8e39-c89b4a5d861c] Took 1.72 seconds to deallocate network for instance. [ 1993.809378] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Acquiring lock "refresh_cache-2afecec5-763d-4616-b690-41d3101cfc47" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1993.809546] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Acquired lock "refresh_cache-2afecec5-763d-4616-b690-41d3101cfc47" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1993.809699] env[62824]: DEBUG nova.network.neutron [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 2afecec5-763d-4616-b690-41d3101cfc47] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1993.826612] env[62824]: DEBUG oslo_vmware.api [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2145958, 'name': ReconfigVM_Task, 'duration_secs': 0.284924} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1993.826884] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2] Reconfigured VM instance instance-00000029 to attach disk [datastore2] 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2/387b7f14-f3c1-43ed-8c4e-6de3ce9822e2.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1993.827649] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4682a6f7-775e-4c39-87a8-52b85b2ff2eb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1993.834253] env[62824]: DEBUG oslo_vmware.api [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Waiting for the task: (returnval){ [ 1993.834253] env[62824]: value = "task-2145959" [ 1993.834253] env[62824]: _type = "Task" [ 1993.834253] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1993.842184] env[62824]: DEBUG oslo_vmware.api [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2145959, 'name': Rename_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1993.855503] env[62824]: DEBUG nova.compute.utils [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1993.857804] env[62824]: DEBUG nova.compute.manager [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 41f068d3-2c8d-46f0-8d84-78531d28b0dd] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1993.858014] env[62824]: DEBUG nova.network.neutron [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 41f068d3-2c8d-46f0-8d84-78531d28b0dd] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1993.904359] env[62824]: DEBUG nova.policy [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd8c161eab02941fabfcc430c451c682a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '31bcd298e89d49da977dda4cefd20432', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 1994.143622] env[62824]: DEBUG oslo_vmware.api [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52f3f614-5ad6-1c70-9139-0336c2e18b19, 'name': SearchDatastore_Task, 'duration_secs': 0.037143} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1994.144393] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1994.144393] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 8bb90a4d-93c9-4f54-b15e-48fe966e6c97] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1994.144393] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1994.144631] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1994.144706] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1994.144993] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-58faa28e-a181-4777-b895-78ddacc80cbb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1994.156368] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1994.156555] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1994.157413] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e0c649b3-da1b-4a65-a3c2-4bb7d1160c5f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1994.164343] env[62824]: DEBUG oslo_vmware.api [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Waiting for the task: (returnval){ [ 1994.164343] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52823cba-c148-4416-df3f-6c89f37891bd" [ 1994.164343] env[62824]: _type = "Task" [ 1994.164343] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1994.164343] env[62824]: DEBUG nova.network.neutron [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 41f068d3-2c8d-46f0-8d84-78531d28b0dd] Successfully created port: 81a37f77-8b23-4fc0-a40d-13c5d36fe6ce {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1994.173651] env[62824]: DEBUG oslo_vmware.api [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52823cba-c148-4416-df3f-6c89f37891bd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1994.260521] env[62824]: DEBUG oslo_concurrency.lockutils [None req-870c96bf-3657-45b9-96d5-c5182b858ac4 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1994.344583] env[62824]: DEBUG oslo_vmware.api [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2145959, 'name': Rename_Task, 'duration_secs': 0.139474} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1994.344920] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1994.345176] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9dc6819c-5aed-4541-986e-fc230aa8a5f3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1994.351494] env[62824]: DEBUG oslo_vmware.api [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Waiting for the task: (returnval){ [ 1994.351494] env[62824]: value = "task-2145960" [ 1994.351494] env[62824]: _type = "Task" [ 1994.351494] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1994.357921] env[62824]: DEBUG nova.compute.manager [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 41f068d3-2c8d-46f0-8d84-78531d28b0dd] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1994.368696] env[62824]: DEBUG oslo_vmware.api [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2145960, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1994.374710] env[62824]: DEBUG nova.network.neutron [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 2afecec5-763d-4616-b690-41d3101cfc47] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1994.379500] env[62824]: DEBUG nova.compute.manager [req-48d84468-563c-4956-8f41-54963119cfe9 req-ea3bc7cb-5ac0-465a-8341-a733c62f23c3 service nova] [instance: 2afecec5-763d-4616-b690-41d3101cfc47] Received event network-vif-plugged-3d71b7f8-4ccd-4ae9-87c1-c4f03a8a22ed {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1994.379800] env[62824]: DEBUG oslo_concurrency.lockutils [req-48d84468-563c-4956-8f41-54963119cfe9 req-ea3bc7cb-5ac0-465a-8341-a733c62f23c3 service nova] Acquiring lock "2afecec5-763d-4616-b690-41d3101cfc47-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1994.379904] env[62824]: DEBUG oslo_concurrency.lockutils [req-48d84468-563c-4956-8f41-54963119cfe9 req-ea3bc7cb-5ac0-465a-8341-a733c62f23c3 service nova] Lock "2afecec5-763d-4616-b690-41d3101cfc47-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1994.380110] env[62824]: DEBUG oslo_concurrency.lockutils [req-48d84468-563c-4956-8f41-54963119cfe9 req-ea3bc7cb-5ac0-465a-8341-a733c62f23c3 service nova] Lock "2afecec5-763d-4616-b690-41d3101cfc47-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1994.380656] env[62824]: DEBUG nova.compute.manager [req-48d84468-563c-4956-8f41-54963119cfe9 req-ea3bc7cb-5ac0-465a-8341-a733c62f23c3 service nova] [instance: 2afecec5-763d-4616-b690-41d3101cfc47] No waiting events found dispatching network-vif-plugged-3d71b7f8-4ccd-4ae9-87c1-c4f03a8a22ed {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1994.380656] env[62824]: WARNING nova.compute.manager [req-48d84468-563c-4956-8f41-54963119cfe9 req-ea3bc7cb-5ac0-465a-8341-a733c62f23c3 service nova] [instance: 2afecec5-763d-4616-b690-41d3101cfc47] Received unexpected event network-vif-plugged-3d71b7f8-4ccd-4ae9-87c1-c4f03a8a22ed for instance with vm_state building and task_state spawning. [ 1994.380656] env[62824]: DEBUG nova.compute.manager [req-48d84468-563c-4956-8f41-54963119cfe9 req-ea3bc7cb-5ac0-465a-8341-a733c62f23c3 service nova] [instance: 2afecec5-763d-4616-b690-41d3101cfc47] Received event network-changed-3d71b7f8-4ccd-4ae9-87c1-c4f03a8a22ed {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1994.380861] env[62824]: DEBUG nova.compute.manager [req-48d84468-563c-4956-8f41-54963119cfe9 req-ea3bc7cb-5ac0-465a-8341-a733c62f23c3 service nova] [instance: 2afecec5-763d-4616-b690-41d3101cfc47] Refreshing instance network info cache due to event network-changed-3d71b7f8-4ccd-4ae9-87c1-c4f03a8a22ed. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1994.380861] env[62824]: DEBUG oslo_concurrency.lockutils [req-48d84468-563c-4956-8f41-54963119cfe9 req-ea3bc7cb-5ac0-465a-8341-a733c62f23c3 service nova] Acquiring lock "refresh_cache-2afecec5-763d-4616-b690-41d3101cfc47" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1994.677522] env[62824]: DEBUG oslo_vmware.api [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52823cba-c148-4416-df3f-6c89f37891bd, 'name': SearchDatastore_Task, 'duration_secs': 0.028215} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1994.680566] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d4a40f68-2822-4bed-8aec-126ee42349c2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1994.685632] env[62824]: DEBUG oslo_vmware.api [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Waiting for the task: (returnval){ [ 1994.685632] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52d0fdd5-edf8-21ff-178f-a397466e74b3" [ 1994.685632] env[62824]: _type = "Task" [ 1994.685632] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1994.696052] env[62824]: DEBUG oslo_vmware.api [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52d0fdd5-edf8-21ff-178f-a397466e74b3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1994.785829] env[62824]: DEBUG nova.network.neutron [req-36eaafd2-8650-4268-bbb7-f0e966be0b7b req-4a2e595b-9a31-44db-8960-47b23f6a3ee4 service nova] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Updated VIF entry in instance network info cache for port b27bda53-5f70-40ff-bb04-9b22f2241093. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1994.786236] env[62824]: DEBUG nova.network.neutron [req-36eaafd2-8650-4268-bbb7-f0e966be0b7b req-4a2e595b-9a31-44db-8960-47b23f6a3ee4 service nova] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Updating instance_info_cache with network_info: [{"id": "b27bda53-5f70-40ff-bb04-9b22f2241093", "address": "fa:16:3e:4d:09:75", "network": {"id": "3d2f0f85-4d51-4459-ad99-443b0821fd9d", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1138734647-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "adf40520c1ef4d97b1ffe2c8cc9649b9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "10b81051-1eb1-406b-888c-4548c470c77e", "external-id": "nsx-vlan-transportzone-207", "segmentation_id": 207, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb27bda53-5f", "ovs_interfaceid": "b27bda53-5f70-40ff-bb04-9b22f2241093", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1994.863793] env[62824]: DEBUG oslo_vmware.api [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2145960, 'name': PowerOnVM_Task, 'duration_secs': 0.446625} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1994.866696] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1994.866893] env[62824]: INFO nova.compute.manager [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2] Took 7.52 seconds to spawn the instance on the hypervisor. [ 1994.867096] env[62824]: DEBUG nova.compute.manager [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1994.871890] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c714060-4cec-469c-9cef-5aca37af0ee2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1994.875528] env[62824]: DEBUG nova.network.neutron [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 2afecec5-763d-4616-b690-41d3101cfc47] Updating instance_info_cache with network_info: [{"id": "3d71b7f8-4ccd-4ae9-87c1-c4f03a8a22ed", "address": "fa:16:3e:74:ae:c4", "network": {"id": "f099da67-9797-4509-a5fe-7c1d5d8560d5", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-13908051-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "31bcd298e89d49da977dda4cefd20432", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "023d6500-887e-4dc4-bec5-06b40450d9c0", "external-id": "nsx-vlan-transportzone-108", "segmentation_id": 108, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d71b7f8-4c", "ovs_interfaceid": "3d71b7f8-4ccd-4ae9-87c1-c4f03a8a22ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1994.928947] env[62824]: DEBUG nova.compute.manager [req-493032fe-8fd7-42cc-9588-5f08e8829a0b req-658380cc-c410-4057-8229-64faa283e5da service nova] [instance: dc3317e4-6958-4df6-8e39-c89b4a5d861c] Received event network-vif-deleted-4c03a7c0-a685-4fe9-9dce-635f6dda43b0 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1994.931494] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa464b40-f124-49fd-bcc8-a76a4fb8a91d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1994.940110] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9456067-9a54-434c-ac13-b160ec3d4221 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1994.972430] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac98b670-3daf-49a7-8ff2-51c9a489de0f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1994.980080] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07fb145d-74a4-46e5-a4f0-6be99e6252f6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1994.994192] env[62824]: DEBUG nova.compute.provider_tree [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1995.196225] env[62824]: DEBUG oslo_vmware.api [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52d0fdd5-edf8-21ff-178f-a397466e74b3, 'name': SearchDatastore_Task, 'duration_secs': 0.010251} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1995.196552] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1995.196744] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 8bb90a4d-93c9-4f54-b15e-48fe966e6c97/8bb90a4d-93c9-4f54-b15e-48fe966e6c97.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1995.196997] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-dc90e961-8c8e-4411-9c88-f628d210f4ea {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1995.203392] env[62824]: DEBUG oslo_vmware.api [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Waiting for the task: (returnval){ [ 1995.203392] env[62824]: value = "task-2145961" [ 1995.203392] env[62824]: _type = "Task" [ 1995.203392] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1995.212480] env[62824]: DEBUG oslo_vmware.api [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': task-2145961, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1995.289475] env[62824]: DEBUG oslo_concurrency.lockutils [req-36eaafd2-8650-4268-bbb7-f0e966be0b7b req-4a2e595b-9a31-44db-8960-47b23f6a3ee4 service nova] Releasing lock "refresh_cache-7febb8b4-f984-4d79-a888-b2829f2a9df6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1995.373115] env[62824]: DEBUG nova.compute.manager [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 41f068d3-2c8d-46f0-8d84-78531d28b0dd] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1995.378742] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Releasing lock "refresh_cache-2afecec5-763d-4616-b690-41d3101cfc47" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1995.379054] env[62824]: DEBUG nova.compute.manager [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 2afecec5-763d-4616-b690-41d3101cfc47] Instance network_info: |[{"id": "3d71b7f8-4ccd-4ae9-87c1-c4f03a8a22ed", "address": "fa:16:3e:74:ae:c4", "network": {"id": "f099da67-9797-4509-a5fe-7c1d5d8560d5", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-13908051-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "31bcd298e89d49da977dda4cefd20432", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "023d6500-887e-4dc4-bec5-06b40450d9c0", "external-id": "nsx-vlan-transportzone-108", "segmentation_id": 108, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d71b7f8-4c", "ovs_interfaceid": "3d71b7f8-4ccd-4ae9-87c1-c4f03a8a22ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1995.379337] env[62824]: DEBUG oslo_concurrency.lockutils [req-48d84468-563c-4956-8f41-54963119cfe9 req-ea3bc7cb-5ac0-465a-8341-a733c62f23c3 service nova] Acquired lock "refresh_cache-2afecec5-763d-4616-b690-41d3101cfc47" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1995.379514] env[62824]: DEBUG nova.network.neutron [req-48d84468-563c-4956-8f41-54963119cfe9 req-ea3bc7cb-5ac0-465a-8341-a733c62f23c3 service nova] [instance: 2afecec5-763d-4616-b690-41d3101cfc47] Refreshing network info cache for port 3d71b7f8-4ccd-4ae9-87c1-c4f03a8a22ed {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1995.380680] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 2afecec5-763d-4616-b690-41d3101cfc47] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:74:ae:c4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '023d6500-887e-4dc4-bec5-06b40450d9c0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3d71b7f8-4ccd-4ae9-87c1-c4f03a8a22ed', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1995.389162] env[62824]: DEBUG oslo.service.loopingcall [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1995.393666] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2afecec5-763d-4616-b690-41d3101cfc47] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1995.397507] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1fcb2ed3-dc48-4009-b3d4-ef09e49a2c9d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1995.412442] env[62824]: INFO nova.compute.manager [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2] Took 37.79 seconds to build instance. [ 1995.419699] env[62824]: DEBUG nova.virt.hardware [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:24Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1995.419988] env[62824]: DEBUG nova.virt.hardware [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1995.420182] env[62824]: DEBUG nova.virt.hardware [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1995.420398] env[62824]: DEBUG nova.virt.hardware [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1995.420549] env[62824]: DEBUG nova.virt.hardware [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1995.420715] env[62824]: DEBUG nova.virt.hardware [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1995.420931] env[62824]: DEBUG nova.virt.hardware [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1995.421108] env[62824]: DEBUG nova.virt.hardware [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1995.421277] env[62824]: DEBUG nova.virt.hardware [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1995.421430] env[62824]: DEBUG nova.virt.hardware [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1995.421601] env[62824]: DEBUG nova.virt.hardware [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1995.422719] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f626db3-25d7-4ffc-90b0-bdd177360711 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1995.427421] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1995.427421] env[62824]: value = "task-2145962" [ 1995.427421] env[62824]: _type = "Task" [ 1995.427421] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1995.434739] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae41b171-11b4-42c2-8b4e-b1d76c155555 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1995.442134] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145962, 'name': CreateVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1995.497026] env[62824]: DEBUG nova.scheduler.client.report [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1995.690451] env[62824]: DEBUG nova.network.neutron [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 41f068d3-2c8d-46f0-8d84-78531d28b0dd] Successfully updated port: 81a37f77-8b23-4fc0-a40d-13c5d36fe6ce {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1995.715984] env[62824]: DEBUG oslo_vmware.api [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': task-2145961, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1995.915180] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fc62e5ce-cf82-4686-88e3-a53850efd30a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Lock "387b7f14-f3c1-43ed-8c4e-6de3ce9822e2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 106.623s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1995.939184] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145962, 'name': CreateVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1996.003266] env[62824]: DEBUG oslo_concurrency.lockutils [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.655s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1996.003703] env[62824]: DEBUG nova.compute.manager [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1996.008998] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 24.860s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1996.110786] env[62824]: DEBUG nova.network.neutron [req-48d84468-563c-4956-8f41-54963119cfe9 req-ea3bc7cb-5ac0-465a-8341-a733c62f23c3 service nova] [instance: 2afecec5-763d-4616-b690-41d3101cfc47] Updated VIF entry in instance network info cache for port 3d71b7f8-4ccd-4ae9-87c1-c4f03a8a22ed. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1996.111168] env[62824]: DEBUG nova.network.neutron [req-48d84468-563c-4956-8f41-54963119cfe9 req-ea3bc7cb-5ac0-465a-8341-a733c62f23c3 service nova] [instance: 2afecec5-763d-4616-b690-41d3101cfc47] Updating instance_info_cache with network_info: [{"id": "3d71b7f8-4ccd-4ae9-87c1-c4f03a8a22ed", "address": "fa:16:3e:74:ae:c4", "network": {"id": "f099da67-9797-4509-a5fe-7c1d5d8560d5", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-13908051-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "31bcd298e89d49da977dda4cefd20432", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "023d6500-887e-4dc4-bec5-06b40450d9c0", "external-id": "nsx-vlan-transportzone-108", "segmentation_id": 108, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d71b7f8-4c", "ovs_interfaceid": "3d71b7f8-4ccd-4ae9-87c1-c4f03a8a22ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1996.193827] env[62824]: DEBUG oslo_concurrency.lockutils [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Acquiring lock "refresh_cache-41f068d3-2c8d-46f0-8d84-78531d28b0dd" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1996.193827] env[62824]: DEBUG oslo_concurrency.lockutils [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Acquired lock "refresh_cache-41f068d3-2c8d-46f0-8d84-78531d28b0dd" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1996.194068] env[62824]: DEBUG nova.network.neutron [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 41f068d3-2c8d-46f0-8d84-78531d28b0dd] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1996.218811] env[62824]: DEBUG oslo_vmware.api [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': task-2145961, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.781732} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1996.219187] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 8bb90a4d-93c9-4f54-b15e-48fe966e6c97/8bb90a4d-93c9-4f54-b15e-48fe966e6c97.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1996.219446] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 8bb90a4d-93c9-4f54-b15e-48fe966e6c97] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1996.219744] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fc8f2592-1f22-413b-a938-d5b3ec366ce0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1996.229508] env[62824]: DEBUG oslo_vmware.api [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Waiting for the task: (returnval){ [ 1996.229508] env[62824]: value = "task-2145963" [ 1996.229508] env[62824]: _type = "Task" [ 1996.229508] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1996.239952] env[62824]: DEBUG oslo_vmware.api [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': task-2145963, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1996.417838] env[62824]: DEBUG nova.compute.manager [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 88bad2dd-dce7-41df-b56c-93a5d054c11e] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1996.438939] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145962, 'name': CreateVM_Task, 'duration_secs': 0.590274} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1996.439132] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2afecec5-763d-4616-b690-41d3101cfc47] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1996.439840] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1996.439986] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1996.440358] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1996.440612] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-97746a54-ff6c-4862-8cf2-500d311f5e66 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1996.445300] env[62824]: DEBUG oslo_vmware.api [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Waiting for the task: (returnval){ [ 1996.445300] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52bf11df-e099-6055-95a6-410bca015760" [ 1996.445300] env[62824]: _type = "Task" [ 1996.445300] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1996.453119] env[62824]: DEBUG oslo_vmware.api [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52bf11df-e099-6055-95a6-410bca015760, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1996.502077] env[62824]: DEBUG nova.compute.manager [req-316f6d11-b78e-45b5-b6c5-185576b106bb req-1f5dc63a-d895-41eb-aa47-a8511e164e21 service nova] [instance: 41f068d3-2c8d-46f0-8d84-78531d28b0dd] Received event network-vif-plugged-81a37f77-8b23-4fc0-a40d-13c5d36fe6ce {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1996.502313] env[62824]: DEBUG oslo_concurrency.lockutils [req-316f6d11-b78e-45b5-b6c5-185576b106bb req-1f5dc63a-d895-41eb-aa47-a8511e164e21 service nova] Acquiring lock "41f068d3-2c8d-46f0-8d84-78531d28b0dd-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1996.502790] env[62824]: DEBUG oslo_concurrency.lockutils [req-316f6d11-b78e-45b5-b6c5-185576b106bb req-1f5dc63a-d895-41eb-aa47-a8511e164e21 service nova] Lock "41f068d3-2c8d-46f0-8d84-78531d28b0dd-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1996.502790] env[62824]: DEBUG oslo_concurrency.lockutils [req-316f6d11-b78e-45b5-b6c5-185576b106bb req-1f5dc63a-d895-41eb-aa47-a8511e164e21 service nova] Lock "41f068d3-2c8d-46f0-8d84-78531d28b0dd-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1996.502888] env[62824]: DEBUG nova.compute.manager [req-316f6d11-b78e-45b5-b6c5-185576b106bb req-1f5dc63a-d895-41eb-aa47-a8511e164e21 service nova] [instance: 41f068d3-2c8d-46f0-8d84-78531d28b0dd] No waiting events found dispatching network-vif-plugged-81a37f77-8b23-4fc0-a40d-13c5d36fe6ce {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1996.502982] env[62824]: WARNING nova.compute.manager [req-316f6d11-b78e-45b5-b6c5-185576b106bb req-1f5dc63a-d895-41eb-aa47-a8511e164e21 service nova] [instance: 41f068d3-2c8d-46f0-8d84-78531d28b0dd] Received unexpected event network-vif-plugged-81a37f77-8b23-4fc0-a40d-13c5d36fe6ce for instance with vm_state building and task_state spawning. [ 1996.503153] env[62824]: DEBUG nova.compute.manager [req-316f6d11-b78e-45b5-b6c5-185576b106bb req-1f5dc63a-d895-41eb-aa47-a8511e164e21 service nova] [instance: 41f068d3-2c8d-46f0-8d84-78531d28b0dd] Received event network-changed-81a37f77-8b23-4fc0-a40d-13c5d36fe6ce {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1996.503307] env[62824]: DEBUG nova.compute.manager [req-316f6d11-b78e-45b5-b6c5-185576b106bb req-1f5dc63a-d895-41eb-aa47-a8511e164e21 service nova] [instance: 41f068d3-2c8d-46f0-8d84-78531d28b0dd] Refreshing instance network info cache due to event network-changed-81a37f77-8b23-4fc0-a40d-13c5d36fe6ce. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 1996.503470] env[62824]: DEBUG oslo_concurrency.lockutils [req-316f6d11-b78e-45b5-b6c5-185576b106bb req-1f5dc63a-d895-41eb-aa47-a8511e164e21 service nova] Acquiring lock "refresh_cache-41f068d3-2c8d-46f0-8d84-78531d28b0dd" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1996.511533] env[62824]: DEBUG nova.compute.utils [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1996.513181] env[62824]: DEBUG nova.objects.instance [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Lazy-loading 'migration_context' on Instance uuid e158b5d4-c120-4e6c-89c4-7668e097926d {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1996.514121] env[62824]: DEBUG nova.compute.manager [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1996.514300] env[62824]: DEBUG nova.network.neutron [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1996.554932] env[62824]: DEBUG nova.policy [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '67ad7b01e840426f84959c233af46382', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e8bdb40d071f48c38a96e4d63db2114a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 1996.613614] env[62824]: DEBUG oslo_concurrency.lockutils [req-48d84468-563c-4956-8f41-54963119cfe9 req-ea3bc7cb-5ac0-465a-8341-a733c62f23c3 service nova] Releasing lock "refresh_cache-2afecec5-763d-4616-b690-41d3101cfc47" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1996.741128] env[62824]: DEBUG oslo_vmware.api [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': task-2145963, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062109} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1996.742361] env[62824]: DEBUG nova.network.neutron [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 41f068d3-2c8d-46f0-8d84-78531d28b0dd] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1996.745306] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 8bb90a4d-93c9-4f54-b15e-48fe966e6c97] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1996.746406] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cea2f14e-366b-47cf-885c-15fd822047a1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1996.771423] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 8bb90a4d-93c9-4f54-b15e-48fe966e6c97] Reconfiguring VM instance instance-0000002a to attach disk [datastore2] 8bb90a4d-93c9-4f54-b15e-48fe966e6c97/8bb90a4d-93c9-4f54-b15e-48fe966e6c97.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1996.774073] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b71d908a-0510-4473-a751-648850844818 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1996.798749] env[62824]: DEBUG oslo_vmware.api [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Waiting for the task: (returnval){ [ 1996.798749] env[62824]: value = "task-2145964" [ 1996.798749] env[62824]: _type = "Task" [ 1996.798749] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1996.803847] env[62824]: DEBUG nova.network.neutron [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Successfully created port: a1d55778-1594-4d8c-943a-350198200d3a {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1996.809948] env[62824]: DEBUG oslo_vmware.api [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': task-2145964, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1996.945829] env[62824]: DEBUG nova.network.neutron [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 41f068d3-2c8d-46f0-8d84-78531d28b0dd] Updating instance_info_cache with network_info: [{"id": "81a37f77-8b23-4fc0-a40d-13c5d36fe6ce", "address": "fa:16:3e:1b:b7:f9", "network": {"id": "f099da67-9797-4509-a5fe-7c1d5d8560d5", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-13908051-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "31bcd298e89d49da977dda4cefd20432", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "023d6500-887e-4dc4-bec5-06b40450d9c0", "external-id": "nsx-vlan-transportzone-108", "segmentation_id": 108, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap81a37f77-8b", "ovs_interfaceid": "81a37f77-8b23-4fc0-a40d-13c5d36fe6ce", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1996.949474] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1996.962826] env[62824]: DEBUG oslo_vmware.api [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52bf11df-e099-6055-95a6-410bca015760, 'name': SearchDatastore_Task, 'duration_secs': 0.009357} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1996.963739] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1996.963999] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 2afecec5-763d-4616-b690-41d3101cfc47] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1996.964408] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1996.964564] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1996.964747] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1996.965429] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-41b13cef-8f9d-427a-bf35-80dc0746ba4c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1996.977812] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1996.977812] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1996.978567] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-535635ce-db16-4e02-9561-dacfdad51520 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1996.986839] env[62824]: DEBUG oslo_vmware.api [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Waiting for the task: (returnval){ [ 1996.986839] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52e8ee3e-7ba8-7ddf-4290-62102ba3c1b1" [ 1996.986839] env[62824]: _type = "Task" [ 1996.986839] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1996.995314] env[62824]: DEBUG oslo_vmware.api [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52e8ee3e-7ba8-7ddf-4290-62102ba3c1b1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1997.020720] env[62824]: DEBUG nova.compute.manager [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1997.310766] env[62824]: DEBUG oslo_vmware.api [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': task-2145964, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1997.450053] env[62824]: DEBUG oslo_concurrency.lockutils [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Releasing lock "refresh_cache-41f068d3-2c8d-46f0-8d84-78531d28b0dd" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1997.450214] env[62824]: DEBUG nova.compute.manager [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 41f068d3-2c8d-46f0-8d84-78531d28b0dd] Instance network_info: |[{"id": "81a37f77-8b23-4fc0-a40d-13c5d36fe6ce", "address": "fa:16:3e:1b:b7:f9", "network": {"id": "f099da67-9797-4509-a5fe-7c1d5d8560d5", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-13908051-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "31bcd298e89d49da977dda4cefd20432", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "023d6500-887e-4dc4-bec5-06b40450d9c0", "external-id": "nsx-vlan-transportzone-108", "segmentation_id": 108, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap81a37f77-8b", "ovs_interfaceid": "81a37f77-8b23-4fc0-a40d-13c5d36fe6ce", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1997.454598] env[62824]: DEBUG oslo_concurrency.lockutils [req-316f6d11-b78e-45b5-b6c5-185576b106bb req-1f5dc63a-d895-41eb-aa47-a8511e164e21 service nova] Acquired lock "refresh_cache-41f068d3-2c8d-46f0-8d84-78531d28b0dd" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1997.454598] env[62824]: DEBUG nova.network.neutron [req-316f6d11-b78e-45b5-b6c5-185576b106bb req-1f5dc63a-d895-41eb-aa47-a8511e164e21 service nova] [instance: 41f068d3-2c8d-46f0-8d84-78531d28b0dd] Refreshing network info cache for port 81a37f77-8b23-4fc0-a40d-13c5d36fe6ce {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1997.455285] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 41f068d3-2c8d-46f0-8d84-78531d28b0dd] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1b:b7:f9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '023d6500-887e-4dc4-bec5-06b40450d9c0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '81a37f77-8b23-4fc0-a40d-13c5d36fe6ce', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1997.463118] env[62824]: DEBUG oslo.service.loopingcall [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1997.465985] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 41f068d3-2c8d-46f0-8d84-78531d28b0dd] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1997.465985] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-de8625ce-d1eb-424d-81be-f4fb85ef4a14 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1997.493117] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1997.493117] env[62824]: value = "task-2145965" [ 1997.493117] env[62824]: _type = "Task" [ 1997.493117] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1997.504627] env[62824]: DEBUG oslo_vmware.api [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52e8ee3e-7ba8-7ddf-4290-62102ba3c1b1, 'name': SearchDatastore_Task, 'duration_secs': 0.026663} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1997.511301] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145965, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1997.511826] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f541f85b-2977-413d-b482-1c077f990fec {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1997.517918] env[62824]: DEBUG oslo_vmware.api [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Waiting for the task: (returnval){ [ 1997.517918] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5219982e-fc1b-b101-c055-8b2df60dcd2e" [ 1997.517918] env[62824]: _type = "Task" [ 1997.517918] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1997.533547] env[62824]: DEBUG oslo_vmware.api [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5219982e-fc1b-b101-c055-8b2df60dcd2e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1997.587068] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c84be6a-a9e4-4e20-8f87-2872c8b5aa67 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1997.596270] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31418368-d3c1-4ab3-9071-08c514ea7b0c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1997.629802] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84dce6ee-ca91-4d63-8432-917c00f31e4e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1997.637821] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12e90937-6cac-47bd-b9b6-5511fe95e480 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1997.651500] env[62824]: DEBUG nova.compute.provider_tree [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1997.721538] env[62824]: DEBUG nova.compute.manager [None req-21197ef3-67a1-4f5a-a6f1-7b606e12edaf tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1997.722434] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6398ec08-7927-406d-8fe6-3f3759880ed0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1997.810885] env[62824]: DEBUG oslo_vmware.api [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': task-2145964, 'name': ReconfigVM_Task, 'duration_secs': 0.715813} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1997.811169] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 8bb90a4d-93c9-4f54-b15e-48fe966e6c97] Reconfigured VM instance instance-0000002a to attach disk [datastore2] 8bb90a4d-93c9-4f54-b15e-48fe966e6c97/8bb90a4d-93c9-4f54-b15e-48fe966e6c97.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1997.811785] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-36bfcf60-8c27-4457-8988-b3ce440894bc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1997.818063] env[62824]: DEBUG oslo_vmware.api [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Waiting for the task: (returnval){ [ 1997.818063] env[62824]: value = "task-2145966" [ 1997.818063] env[62824]: _type = "Task" [ 1997.818063] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1997.825749] env[62824]: DEBUG oslo_vmware.api [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': task-2145966, 'name': Rename_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1998.004365] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145965, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1998.027569] env[62824]: DEBUG oslo_vmware.api [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5219982e-fc1b-b101-c055-8b2df60dcd2e, 'name': SearchDatastore_Task, 'duration_secs': 0.014063} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1998.027840] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1998.028114] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 2afecec5-763d-4616-b690-41d3101cfc47/2afecec5-763d-4616-b690-41d3101cfc47.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1998.028375] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8d4f1538-25b6-4628-905a-c5bef1bccacc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1998.036808] env[62824]: DEBUG nova.compute.manager [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1998.038879] env[62824]: DEBUG oslo_vmware.api [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Waiting for the task: (returnval){ [ 1998.038879] env[62824]: value = "task-2145967" [ 1998.038879] env[62824]: _type = "Task" [ 1998.038879] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1998.047747] env[62824]: DEBUG oslo_vmware.api [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': task-2145967, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1998.059223] env[62824]: DEBUG nova.virt.hardware [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1998.059401] env[62824]: DEBUG nova.virt.hardware [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1998.059554] env[62824]: DEBUG nova.virt.hardware [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1998.059738] env[62824]: DEBUG nova.virt.hardware [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1998.059890] env[62824]: DEBUG nova.virt.hardware [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1998.060041] env[62824]: DEBUG nova.virt.hardware [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1998.060335] env[62824]: DEBUG nova.virt.hardware [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1998.060598] env[62824]: DEBUG nova.virt.hardware [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1998.060792] env[62824]: DEBUG nova.virt.hardware [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1998.060998] env[62824]: DEBUG nova.virt.hardware [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1998.061228] env[62824]: DEBUG nova.virt.hardware [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1998.062130] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7535cbca-210d-418f-8e28-d3b801a5cf55 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1998.072539] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb8f0e51-cb1c-4f03-9390-4e5af78ec9ad {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1998.154806] env[62824]: DEBUG nova.scheduler.client.report [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1998.233949] env[62824]: INFO nova.compute.manager [None req-21197ef3-67a1-4f5a-a6f1-7b606e12edaf tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] instance snapshotting [ 1998.236966] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-834761e5-10af-44af-a995-fcb13bac2e04 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1998.258896] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-400bc144-108c-4803-b025-a581700cb9b5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1998.281031] env[62824]: DEBUG nova.network.neutron [req-316f6d11-b78e-45b5-b6c5-185576b106bb req-1f5dc63a-d895-41eb-aa47-a8511e164e21 service nova] [instance: 41f068d3-2c8d-46f0-8d84-78531d28b0dd] Updated VIF entry in instance network info cache for port 81a37f77-8b23-4fc0-a40d-13c5d36fe6ce. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1998.281402] env[62824]: DEBUG nova.network.neutron [req-316f6d11-b78e-45b5-b6c5-185576b106bb req-1f5dc63a-d895-41eb-aa47-a8511e164e21 service nova] [instance: 41f068d3-2c8d-46f0-8d84-78531d28b0dd] Updating instance_info_cache with network_info: [{"id": "81a37f77-8b23-4fc0-a40d-13c5d36fe6ce", "address": "fa:16:3e:1b:b7:f9", "network": {"id": "f099da67-9797-4509-a5fe-7c1d5d8560d5", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-13908051-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "31bcd298e89d49da977dda4cefd20432", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "023d6500-887e-4dc4-bec5-06b40450d9c0", "external-id": "nsx-vlan-transportzone-108", "segmentation_id": 108, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap81a37f77-8b", "ovs_interfaceid": "81a37f77-8b23-4fc0-a40d-13c5d36fe6ce", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1998.330640] env[62824]: DEBUG oslo_vmware.api [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': task-2145966, 'name': Rename_Task, 'duration_secs': 0.142539} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1998.331109] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 8bb90a4d-93c9-4f54-b15e-48fe966e6c97] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1998.331226] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dd4d91b9-e765-432f-a077-83bc2d2c5e22 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1998.338949] env[62824]: DEBUG oslo_vmware.api [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Waiting for the task: (returnval){ [ 1998.338949] env[62824]: value = "task-2145968" [ 1998.338949] env[62824]: _type = "Task" [ 1998.338949] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1998.348040] env[62824]: DEBUG oslo_vmware.api [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': task-2145968, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1998.507801] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145965, 'name': CreateVM_Task, 'duration_secs': 0.639077} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1998.507801] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 41f068d3-2c8d-46f0-8d84-78531d28b0dd] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1998.508841] env[62824]: DEBUG oslo_concurrency.lockutils [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1998.509053] env[62824]: DEBUG oslo_concurrency.lockutils [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1998.509790] env[62824]: DEBUG oslo_concurrency.lockutils [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1998.510901] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-171a10e8-dddd-4fb3-b386-e5ed03797528 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1998.515957] env[62824]: DEBUG oslo_vmware.api [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Waiting for the task: (returnval){ [ 1998.515957] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52175a57-eccf-bc8b-e824-86a5d682fa5a" [ 1998.515957] env[62824]: _type = "Task" [ 1998.515957] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1998.524584] env[62824]: DEBUG oslo_vmware.api [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52175a57-eccf-bc8b-e824-86a5d682fa5a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1998.548921] env[62824]: DEBUG oslo_vmware.api [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': task-2145967, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.491283} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1998.549495] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 2afecec5-763d-4616-b690-41d3101cfc47/2afecec5-763d-4616-b690-41d3101cfc47.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1998.549495] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 2afecec5-763d-4616-b690-41d3101cfc47] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1998.549746] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b85efdad-b132-4ae3-a6c4-24ad4722e161 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1998.556277] env[62824]: DEBUG oslo_vmware.api [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Waiting for the task: (returnval){ [ 1998.556277] env[62824]: value = "task-2145969" [ 1998.556277] env[62824]: _type = "Task" [ 1998.556277] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1998.563738] env[62824]: DEBUG oslo_vmware.api [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': task-2145969, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1998.756508] env[62824]: DEBUG nova.compute.manager [req-4f9b02ee-ab22-440a-ba71-fbb43d256ee1 req-131355c4-82ef-4385-94d8-6c0137bd662d service nova] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Received event network-vif-plugged-a1d55778-1594-4d8c-943a-350198200d3a {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 1998.756734] env[62824]: DEBUG oslo_concurrency.lockutils [req-4f9b02ee-ab22-440a-ba71-fbb43d256ee1 req-131355c4-82ef-4385-94d8-6c0137bd662d service nova] Acquiring lock "35f0435c-fe46-46ab-bc71-c21bb0ad3d55-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1998.757016] env[62824]: DEBUG oslo_concurrency.lockutils [req-4f9b02ee-ab22-440a-ba71-fbb43d256ee1 req-131355c4-82ef-4385-94d8-6c0137bd662d service nova] Lock "35f0435c-fe46-46ab-bc71-c21bb0ad3d55-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1998.757222] env[62824]: DEBUG oslo_concurrency.lockutils [req-4f9b02ee-ab22-440a-ba71-fbb43d256ee1 req-131355c4-82ef-4385-94d8-6c0137bd662d service nova] Lock "35f0435c-fe46-46ab-bc71-c21bb0ad3d55-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1998.757384] env[62824]: DEBUG nova.compute.manager [req-4f9b02ee-ab22-440a-ba71-fbb43d256ee1 req-131355c4-82ef-4385-94d8-6c0137bd662d service nova] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] No waiting events found dispatching network-vif-plugged-a1d55778-1594-4d8c-943a-350198200d3a {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1998.757548] env[62824]: WARNING nova.compute.manager [req-4f9b02ee-ab22-440a-ba71-fbb43d256ee1 req-131355c4-82ef-4385-94d8-6c0137bd662d service nova] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Received unexpected event network-vif-plugged-a1d55778-1594-4d8c-943a-350198200d3a for instance with vm_state building and task_state spawning. [ 1998.770766] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-21197ef3-67a1-4f5a-a6f1-7b606e12edaf tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Creating Snapshot of the VM instance {{(pid=62824) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1998.771089] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-8b4c19c6-12f0-4a51-b7ea-2b405ff8a730 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1998.779173] env[62824]: DEBUG oslo_vmware.api [None req-21197ef3-67a1-4f5a-a6f1-7b606e12edaf tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Waiting for the task: (returnval){ [ 1998.779173] env[62824]: value = "task-2145970" [ 1998.779173] env[62824]: _type = "Task" [ 1998.779173] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1998.787591] env[62824]: DEBUG oslo_concurrency.lockutils [req-316f6d11-b78e-45b5-b6c5-185576b106bb req-1f5dc63a-d895-41eb-aa47-a8511e164e21 service nova] Releasing lock "refresh_cache-41f068d3-2c8d-46f0-8d84-78531d28b0dd" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1998.787969] env[62824]: DEBUG oslo_vmware.api [None req-21197ef3-67a1-4f5a-a6f1-7b606e12edaf tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2145970, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1998.851848] env[62824]: DEBUG oslo_vmware.api [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': task-2145968, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1998.852833] env[62824]: DEBUG nova.network.neutron [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Successfully updated port: a1d55778-1594-4d8c-943a-350198200d3a {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1999.027276] env[62824]: DEBUG oslo_vmware.api [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52175a57-eccf-bc8b-e824-86a5d682fa5a, 'name': SearchDatastore_Task, 'duration_secs': 0.015213} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1999.027650] env[62824]: DEBUG oslo_concurrency.lockutils [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1999.027855] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 41f068d3-2c8d-46f0-8d84-78531d28b0dd] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1999.028108] env[62824]: DEBUG oslo_concurrency.lockutils [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1999.028247] env[62824]: DEBUG oslo_concurrency.lockutils [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1999.028439] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1999.028705] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4112e751-9708-494f-bbf4-ff05d45c7c13 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1999.039127] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1999.039344] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1999.040134] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f2d3ba63-44fc-4225-a7c1-9baeef8fd7cf {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1999.045271] env[62824]: DEBUG oslo_vmware.api [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Waiting for the task: (returnval){ [ 1999.045271] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]525be509-ca6e-553b-f9ac-3e6274229127" [ 1999.045271] env[62824]: _type = "Task" [ 1999.045271] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1999.053290] env[62824]: DEBUG oslo_vmware.api [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]525be509-ca6e-553b-f9ac-3e6274229127, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1999.065361] env[62824]: DEBUG oslo_vmware.api [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': task-2145969, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070572} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1999.065621] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 2afecec5-763d-4616-b690-41d3101cfc47] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1999.066669] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddf42bb7-ca4d-49c3-8dd3-67d0941f59a0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1999.090656] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 2afecec5-763d-4616-b690-41d3101cfc47] Reconfiguring VM instance instance-0000002b to attach disk [datastore2] 2afecec5-763d-4616-b690-41d3101cfc47/2afecec5-763d-4616-b690-41d3101cfc47.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1999.090956] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1f52be61-16ed-4d22-9640-cd32db396da3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1999.111811] env[62824]: DEBUG oslo_vmware.api [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Waiting for the task: (returnval){ [ 1999.111811] env[62824]: value = "task-2145971" [ 1999.111811] env[62824]: _type = "Task" [ 1999.111811] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1999.120492] env[62824]: DEBUG oslo_vmware.api [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': task-2145971, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1999.166043] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 3.157s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1999.171855] env[62824]: DEBUG oslo_concurrency.lockutils [None req-847fb409-5740-4b52-92bf-c077485b2771 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.899s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1999.172203] env[62824]: DEBUG nova.objects.instance [None req-847fb409-5740-4b52-92bf-c077485b2771 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lazy-loading 'resources' on Instance uuid 487c2c9d-2cd2-4912-8613-e1bfac732c40 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1999.289658] env[62824]: DEBUG oslo_vmware.api [None req-21197ef3-67a1-4f5a-a6f1-7b606e12edaf tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2145970, 'name': CreateSnapshot_Task, 'duration_secs': 0.489673} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1999.289941] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-21197ef3-67a1-4f5a-a6f1-7b606e12edaf tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Created Snapshot of the VM instance {{(pid=62824) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1999.290703] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea44c4e9-0574-44dc-85fe-0ffed204ff34 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1999.349422] env[62824]: DEBUG oslo_vmware.api [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': task-2145968, 'name': PowerOnVM_Task, 'duration_secs': 0.6566} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1999.349843] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 8bb90a4d-93c9-4f54-b15e-48fe966e6c97] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1999.350034] env[62824]: INFO nova.compute.manager [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 8bb90a4d-93c9-4f54-b15e-48fe966e6c97] Took 9.43 seconds to spawn the instance on the hypervisor. [ 1999.350093] env[62824]: DEBUG nova.compute.manager [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 8bb90a4d-93c9-4f54-b15e-48fe966e6c97] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1999.350917] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b143d7f-174b-490e-9472-69867f1a2865 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1999.355631] env[62824]: DEBUG oslo_concurrency.lockutils [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Acquiring lock "refresh_cache-35f0435c-fe46-46ab-bc71-c21bb0ad3d55" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1999.355779] env[62824]: DEBUG oslo_concurrency.lockutils [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Acquired lock "refresh_cache-35f0435c-fe46-46ab-bc71-c21bb0ad3d55" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1999.356057] env[62824]: DEBUG nova.network.neutron [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1999.556338] env[62824]: DEBUG oslo_vmware.api [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]525be509-ca6e-553b-f9ac-3e6274229127, 'name': SearchDatastore_Task, 'duration_secs': 0.017311} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1999.557186] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8ac0af93-3fdf-41c1-aca0-cd27cb153d81 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1999.562318] env[62824]: DEBUG oslo_vmware.api [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Waiting for the task: (returnval){ [ 1999.562318] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52d12891-87d9-33fe-48a2-f5bc6a176a2a" [ 1999.562318] env[62824]: _type = "Task" [ 1999.562318] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1999.571316] env[62824]: DEBUG oslo_vmware.api [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52d12891-87d9-33fe-48a2-f5bc6a176a2a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1999.624174] env[62824]: DEBUG oslo_vmware.api [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': task-2145971, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1999.811987] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-21197ef3-67a1-4f5a-a6f1-7b606e12edaf tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Creating linked-clone VM from snapshot {{(pid=62824) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1999.814866] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-6906c4aa-8958-4125-8ad4-f55d79ba91f9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1999.824906] env[62824]: DEBUG oslo_vmware.api [None req-21197ef3-67a1-4f5a-a6f1-7b606e12edaf tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Waiting for the task: (returnval){ [ 1999.824906] env[62824]: value = "task-2145972" [ 1999.824906] env[62824]: _type = "Task" [ 1999.824906] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1999.834651] env[62824]: DEBUG oslo_vmware.api [None req-21197ef3-67a1-4f5a-a6f1-7b606e12edaf tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2145972, 'name': CloneVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1999.877086] env[62824]: INFO nova.compute.manager [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 8bb90a4d-93c9-4f54-b15e-48fe966e6c97] Took 41.93 seconds to build instance. [ 1999.909053] env[62824]: DEBUG nova.network.neutron [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2000.075626] env[62824]: DEBUG oslo_vmware.api [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52d12891-87d9-33fe-48a2-f5bc6a176a2a, 'name': SearchDatastore_Task, 'duration_secs': 0.059474} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2000.076583] env[62824]: DEBUG nova.network.neutron [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Updating instance_info_cache with network_info: [{"id": "a1d55778-1594-4d8c-943a-350198200d3a", "address": "fa:16:3e:d6:c1:1b", "network": {"id": "64e5f639-e16b-4c14-bffb-5d30fb0cb35e", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-1724557834-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e8bdb40d071f48c38a96e4d63db2114a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5514c5a3-1294-40ad-ae96-29d5c24a3d95", "external-id": "nsx-vlan-transportzone-179", "segmentation_id": 179, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa1d55778-15", "ovs_interfaceid": "a1d55778-1594-4d8c-943a-350198200d3a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2000.077804] env[62824]: DEBUG oslo_concurrency.lockutils [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2000.078080] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 41f068d3-2c8d-46f0-8d84-78531d28b0dd/41f068d3-2c8d-46f0-8d84-78531d28b0dd.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2000.078628] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6ea0ee94-3217-43a3-84da-a2051360fdfa {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2000.087676] env[62824]: DEBUG oslo_vmware.api [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Waiting for the task: (returnval){ [ 2000.087676] env[62824]: value = "task-2145973" [ 2000.087676] env[62824]: _type = "Task" [ 2000.087676] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2000.098702] env[62824]: DEBUG oslo_vmware.api [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': task-2145973, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2000.126488] env[62824]: DEBUG oslo_vmware.api [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': task-2145971, 'name': ReconfigVM_Task, 'duration_secs': 0.924567} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2000.126780] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 2afecec5-763d-4616-b690-41d3101cfc47] Reconfigured VM instance instance-0000002b to attach disk [datastore2] 2afecec5-763d-4616-b690-41d3101cfc47/2afecec5-763d-4616-b690-41d3101cfc47.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2000.127585] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-40093423-037b-424a-8823-1369487d3acb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2000.134472] env[62824]: DEBUG oslo_vmware.api [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Waiting for the task: (returnval){ [ 2000.134472] env[62824]: value = "task-2145974" [ 2000.134472] env[62824]: _type = "Task" [ 2000.134472] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2000.143257] env[62824]: DEBUG oslo_vmware.api [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': task-2145974, 'name': Rename_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2000.233422] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc514f11-f6d8-4481-b99f-c3e579e44636 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2000.242203] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d198dc8-fab5-4821-b4cf-a4900eb358cd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2001.005390] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5eb320f2-e0a5-4a53-9a09-eb5627ab346e tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Lock "8bb90a4d-93c9-4f54-b15e-48fe966e6c97" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 110.997s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2001.005797] env[62824]: DEBUG oslo_concurrency.lockutils [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Releasing lock "refresh_cache-35f0435c-fe46-46ab-bc71-c21bb0ad3d55" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2001.006075] env[62824]: DEBUG nova.compute.manager [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Instance network_info: |[{"id": "a1d55778-1594-4d8c-943a-350198200d3a", "address": "fa:16:3e:d6:c1:1b", "network": {"id": "64e5f639-e16b-4c14-bffb-5d30fb0cb35e", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-1724557834-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e8bdb40d071f48c38a96e4d63db2114a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5514c5a3-1294-40ad-ae96-29d5c24a3d95", "external-id": "nsx-vlan-transportzone-179", "segmentation_id": 179, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa1d55778-15", "ovs_interfaceid": "a1d55778-1594-4d8c-943a-350198200d3a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2001.019023] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d6:c1:1b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5514c5a3-1294-40ad-ae96-29d5c24a3d95', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a1d55778-1594-4d8c-943a-350198200d3a', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2001.024556] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Creating folder: Project (e8bdb40d071f48c38a96e4d63db2114a). Parent ref: group-v438503. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 2001.030746] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66ed83d2-1eeb-442b-9530-b76bb919090f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2001.033961] env[62824]: DEBUG nova.compute.manager [req-1b1717df-d4f4-45d2-9d9f-44e82e64492f req-032e1ea9-440f-4810-b0aa-6eb122c60194 service nova] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Received event network-changed-a1d55778-1594-4d8c-943a-350198200d3a {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2001.034146] env[62824]: DEBUG nova.compute.manager [req-1b1717df-d4f4-45d2-9d9f-44e82e64492f req-032e1ea9-440f-4810-b0aa-6eb122c60194 service nova] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Refreshing instance network info cache due to event network-changed-a1d55778-1594-4d8c-943a-350198200d3a. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2001.034359] env[62824]: DEBUG oslo_concurrency.lockutils [req-1b1717df-d4f4-45d2-9d9f-44e82e64492f req-032e1ea9-440f-4810-b0aa-6eb122c60194 service nova] Acquiring lock "refresh_cache-35f0435c-fe46-46ab-bc71-c21bb0ad3d55" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2001.034500] env[62824]: DEBUG oslo_concurrency.lockutils [req-1b1717df-d4f4-45d2-9d9f-44e82e64492f req-032e1ea9-440f-4810-b0aa-6eb122c60194 service nova] Acquired lock "refresh_cache-35f0435c-fe46-46ab-bc71-c21bb0ad3d55" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2001.034658] env[62824]: DEBUG nova.network.neutron [req-1b1717df-d4f4-45d2-9d9f-44e82e64492f req-032e1ea9-440f-4810-b0aa-6eb122c60194 service nova] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Refreshing network info cache for port a1d55778-1594-4d8c-943a-350198200d3a {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2001.036566] env[62824]: DEBUG nova.compute.manager [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] [instance: c63b05b7-1dbf-4637-8728-33673be0d3ea] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2001.042018] env[62824]: INFO nova.compute.manager [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Swapping old allocation on dict_keys(['bf95157b-a274-42de-9ccf-9851128a44a1']) held by migration bc55ced6-f1ca-4506-869b-2d7d5683b841 for instance [ 2001.042018] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ecc0f010-fdff-48c8-a577-6e065e245d51 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2001.051063] env[62824]: DEBUG oslo_vmware.api [None req-21197ef3-67a1-4f5a-a6f1-7b606e12edaf tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2145972, 'name': CloneVM_Task} progress is 95%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2001.060249] env[62824]: DEBUG oslo_vmware.api [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': task-2145973, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.532622} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2001.060854] env[62824]: DEBUG oslo_vmware.api [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': task-2145974, 'name': Rename_Task, 'duration_secs': 0.149436} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2001.065118] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 41f068d3-2c8d-46f0-8d84-78531d28b0dd/41f068d3-2c8d-46f0-8d84-78531d28b0dd.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2001.065118] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 41f068d3-2c8d-46f0-8d84-78531d28b0dd] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2001.065239] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-041f5460-61c0-42a9-89b4-b27f8534736d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2001.069440] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 2afecec5-763d-4616-b690-41d3101cfc47] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2001.070074] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-52083d4a-0eb9-4de6-839b-b3c578d76d93 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2001.071759] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3a789f8e-ab61-486f-837a-a6615e5cf126 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2001.074595] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Created folder: Project (e8bdb40d071f48c38a96e4d63db2114a) in parent group-v438503. [ 2001.074795] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Creating folder: Instances. Parent ref: group-v438606. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 2001.075583] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-929edf4c-fba4-4f84-8dfe-c72beaf9ac6e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2001.086474] env[62824]: DEBUG nova.compute.provider_tree [None req-847fb409-5740-4b52-92bf-c077485b2771 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2001.089551] env[62824]: DEBUG nova.scheduler.client.report [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Overwriting current allocation {'allocations': {'bf95157b-a274-42de-9ccf-9851128a44a1': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 71}}, 'project_id': 'b2f9084931bc4b94b0ae85e3e4b0ba90', 'user_id': '2cbf8eb928484d54a5371365386e73f6', 'consumer_generation': 1} on consumer e158b5d4-c120-4e6c-89c4-7668e097926d {{(pid=62824) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2036}} [ 2001.093409] env[62824]: DEBUG oslo_vmware.api [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Waiting for the task: (returnval){ [ 2001.093409] env[62824]: value = "task-2145976" [ 2001.093409] env[62824]: _type = "Task" [ 2001.093409] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2001.093657] env[62824]: DEBUG oslo_vmware.api [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Waiting for the task: (returnval){ [ 2001.093657] env[62824]: value = "task-2145977" [ 2001.093657] env[62824]: _type = "Task" [ 2001.093657] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2001.102404] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Created folder: Instances in parent group-v438606. [ 2001.102666] env[62824]: DEBUG oslo.service.loopingcall [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2001.103521] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2001.104078] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-234b51d5-6a95-4cd7-b785-592608e38fc0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2001.125440] env[62824]: DEBUG oslo_vmware.api [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': task-2145976, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2001.125890] env[62824]: DEBUG oslo_vmware.api [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': task-2145977, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2001.130817] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2001.130817] env[62824]: value = "task-2145979" [ 2001.130817] env[62824]: _type = "Task" [ 2001.130817] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2001.139212] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145979, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2001.198011] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Acquiring lock "refresh_cache-e158b5d4-c120-4e6c-89c4-7668e097926d" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2001.198211] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Acquired lock "refresh_cache-e158b5d4-c120-4e6c-89c4-7668e097926d" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2001.198385] env[62824]: DEBUG nova.network.neutron [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2001.517273] env[62824]: DEBUG oslo_vmware.api [None req-21197ef3-67a1-4f5a-a6f1-7b606e12edaf tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2145972, 'name': CloneVM_Task, 'duration_secs': 1.395769} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2001.517622] env[62824]: INFO nova.virt.vmwareapi.vmops [None req-21197ef3-67a1-4f5a-a6f1-7b606e12edaf tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Created linked-clone VM from snapshot [ 2001.518475] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da1ab59f-aecf-459e-a570-61364c5e004e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2001.526257] env[62824]: DEBUG nova.virt.vmwareapi.images [None req-21197ef3-67a1-4f5a-a6f1-7b606e12edaf tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Uploading image 9b84651b-bc0f-4778-9d83-183e9ac78e55 {{(pid=62824) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 2001.546940] env[62824]: DEBUG oslo_vmware.rw_handles [None req-21197ef3-67a1-4f5a-a6f1-7b606e12edaf tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 2001.546940] env[62824]: value = "vm-438605" [ 2001.546940] env[62824]: _type = "VirtualMachine" [ 2001.546940] env[62824]: }. {{(pid=62824) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 2001.551453] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-834cadf5-707b-49b3-84a3-fd7036977579 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2001.560551] env[62824]: DEBUG oslo_vmware.rw_handles [None req-21197ef3-67a1-4f5a-a6f1-7b606e12edaf tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Lease: (returnval){ [ 2001.560551] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5209c73a-2ab2-8df1-4d68-dffee1c46b41" [ 2001.560551] env[62824]: _type = "HttpNfcLease" [ 2001.560551] env[62824]: } obtained for exporting VM: (result){ [ 2001.560551] env[62824]: value = "vm-438605" [ 2001.560551] env[62824]: _type = "VirtualMachine" [ 2001.560551] env[62824]: }. {{(pid=62824) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 2001.560872] env[62824]: DEBUG oslo_vmware.api [None req-21197ef3-67a1-4f5a-a6f1-7b606e12edaf tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Waiting for the lease: (returnval){ [ 2001.560872] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5209c73a-2ab2-8df1-4d68-dffee1c46b41" [ 2001.560872] env[62824]: _type = "HttpNfcLease" [ 2001.560872] env[62824]: } to be ready. {{(pid=62824) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 2001.564800] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2001.566843] env[62824]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 2001.566843] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5209c73a-2ab2-8df1-4d68-dffee1c46b41" [ 2001.566843] env[62824]: _type = "HttpNfcLease" [ 2001.566843] env[62824]: } is initializing. {{(pid=62824) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 2001.595775] env[62824]: DEBUG nova.scheduler.client.report [None req-847fb409-5740-4b52-92bf-c077485b2771 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2001.610739] env[62824]: DEBUG oslo_vmware.api [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': task-2145977, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2001.613625] env[62824]: DEBUG oslo_vmware.api [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': task-2145976, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.286721} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2001.614486] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 41f068d3-2c8d-46f0-8d84-78531d28b0dd] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2001.615312] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-381599dc-2801-4158-94f7-667993243492 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2001.648089] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 41f068d3-2c8d-46f0-8d84-78531d28b0dd] Reconfiguring VM instance instance-0000002c to attach disk [datastore2] 41f068d3-2c8d-46f0-8d84-78531d28b0dd/41f068d3-2c8d-46f0-8d84-78531d28b0dd.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2001.655710] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-16149159-306c-4924-8e25-82dcce577a36 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2001.680189] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145979, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2001.681407] env[62824]: DEBUG oslo_vmware.api [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Waiting for the task: (returnval){ [ 2001.681407] env[62824]: value = "task-2145981" [ 2001.681407] env[62824]: _type = "Task" [ 2001.681407] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2001.689686] env[62824]: DEBUG oslo_vmware.api [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': task-2145981, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2001.812325] env[62824]: DEBUG nova.network.neutron [req-1b1717df-d4f4-45d2-9d9f-44e82e64492f req-032e1ea9-440f-4810-b0aa-6eb122c60194 service nova] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Updated VIF entry in instance network info cache for port a1d55778-1594-4d8c-943a-350198200d3a. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2001.812325] env[62824]: DEBUG nova.network.neutron [req-1b1717df-d4f4-45d2-9d9f-44e82e64492f req-032e1ea9-440f-4810-b0aa-6eb122c60194 service nova] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Updating instance_info_cache with network_info: [{"id": "a1d55778-1594-4d8c-943a-350198200d3a", "address": "fa:16:3e:d6:c1:1b", "network": {"id": "64e5f639-e16b-4c14-bffb-5d30fb0cb35e", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-1724557834-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e8bdb40d071f48c38a96e4d63db2114a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5514c5a3-1294-40ad-ae96-29d5c24a3d95", "external-id": "nsx-vlan-transportzone-179", "segmentation_id": 179, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa1d55778-15", "ovs_interfaceid": "a1d55778-1594-4d8c-943a-350198200d3a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2002.052222] env[62824]: DEBUG nova.network.neutron [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Updating instance_info_cache with network_info: [{"id": "02e90c28-7f36-4d83-84bb-fec194a04dab", "address": "fa:16:3e:53:1a:4c", "network": {"id": "b6403dc3-30ef-4c76-b521-380c179f345d", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.19", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "22c5c40c2e004babaa15896c89307303", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1002b79b-224e-41e3-a484-4245a767147a", "external-id": "nsx-vlan-transportzone-353", "segmentation_id": 353, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap02e90c28-7f", "ovs_interfaceid": "02e90c28-7f36-4d83-84bb-fec194a04dab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2002.069097] env[62824]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 2002.069097] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5209c73a-2ab2-8df1-4d68-dffee1c46b41" [ 2002.069097] env[62824]: _type = "HttpNfcLease" [ 2002.069097] env[62824]: } is ready. {{(pid=62824) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 2002.069399] env[62824]: DEBUG oslo_vmware.rw_handles [None req-21197ef3-67a1-4f5a-a6f1-7b606e12edaf tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 2002.069399] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5209c73a-2ab2-8df1-4d68-dffee1c46b41" [ 2002.069399] env[62824]: _type = "HttpNfcLease" [ 2002.069399] env[62824]: }. {{(pid=62824) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 2002.070219] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8e25e0b-bb55-4c02-acc5-21c0521fc05e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2002.077953] env[62824]: DEBUG oslo_vmware.rw_handles [None req-21197ef3-67a1-4f5a-a6f1-7b606e12edaf tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5243b7f5-5cb5-611a-2d77-60a4ce26c479/disk-0.vmdk from lease info. {{(pid=62824) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 2002.078145] env[62824]: DEBUG oslo_vmware.rw_handles [None req-21197ef3-67a1-4f5a-a6f1-7b606e12edaf tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5243b7f5-5cb5-611a-2d77-60a4ce26c479/disk-0.vmdk for reading. {{(pid=62824) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 2002.136689] env[62824]: DEBUG oslo_concurrency.lockutils [None req-847fb409-5740-4b52-92bf-c077485b2771 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.965s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2002.143693] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.220s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2002.145325] env[62824]: INFO nova.compute.claims [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] [instance: db018af0-7bdf-40c5-b129-fde15dfb8ab4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2002.158570] env[62824]: DEBUG oslo_vmware.api [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': task-2145977, 'name': PowerOnVM_Task, 'duration_secs': 0.842476} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2002.159076] env[62824]: INFO nova.scheduler.client.report [None req-847fb409-5740-4b52-92bf-c077485b2771 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Deleted allocations for instance 487c2c9d-2cd2-4912-8613-e1bfac732c40 [ 2002.166330] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 2afecec5-763d-4616-b690-41d3101cfc47] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2002.166561] env[62824]: INFO nova.compute.manager [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 2afecec5-763d-4616-b690-41d3101cfc47] Took 9.62 seconds to spawn the instance on the hypervisor. [ 2002.166982] env[62824]: DEBUG nova.compute.manager [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 2afecec5-763d-4616-b690-41d3101cfc47] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2002.167055] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2145979, 'name': CreateVM_Task, 'duration_secs': 0.646475} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2002.168110] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb0e3c07-099f-4a81-9baf-71d700ebc3c0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2002.170803] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2002.171489] env[62824]: DEBUG oslo_concurrency.lockutils [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2002.171785] env[62824]: DEBUG oslo_concurrency.lockutils [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2002.172138] env[62824]: DEBUG oslo_concurrency.lockutils [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2002.172832] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ecc3086b-1e35-4845-8cb5-0ed8f25641fe {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2002.181152] env[62824]: DEBUG oslo_vmware.api [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Waiting for the task: (returnval){ [ 2002.181152] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5296d0a8-e601-a09d-afce-8cf122e9b9b3" [ 2002.181152] env[62824]: _type = "Task" [ 2002.181152] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2002.195524] env[62824]: DEBUG oslo_vmware.api [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': task-2145981, 'name': ReconfigVM_Task, 'duration_secs': 0.286693} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2002.203149] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 41f068d3-2c8d-46f0-8d84-78531d28b0dd] Reconfigured VM instance instance-0000002c to attach disk [datastore2] 41f068d3-2c8d-46f0-8d84-78531d28b0dd/41f068d3-2c8d-46f0-8d84-78531d28b0dd.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2002.204390] env[62824]: DEBUG oslo_vmware.api [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5296d0a8-e601-a09d-afce-8cf122e9b9b3, 'name': SearchDatastore_Task, 'duration_secs': 0.010218} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2002.205287] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4163e5db-cd53-485e-81ba-1d9dc482c9c0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2002.206842] env[62824]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-dab3fd50-947d-4a4b-a5b5-b79c38707289 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2002.208443] env[62824]: DEBUG oslo_concurrency.lockutils [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2002.208686] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2002.209148] env[62824]: DEBUG oslo_concurrency.lockutils [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2002.209148] env[62824]: DEBUG oslo_concurrency.lockutils [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2002.209230] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2002.209497] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a9838d9d-46e1-4f00-80ea-241cec991e31 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2002.216443] env[62824]: DEBUG oslo_vmware.api [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Waiting for the task: (returnval){ [ 2002.216443] env[62824]: value = "task-2145982" [ 2002.216443] env[62824]: _type = "Task" [ 2002.216443] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2002.221580] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2002.221900] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2002.223607] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1800310f-0a68-41ba-9d3d-7cdb11c89542 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2002.232408] env[62824]: DEBUG oslo_vmware.api [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': task-2145982, 'name': Rename_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2002.232843] env[62824]: DEBUG oslo_vmware.api [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Waiting for the task: (returnval){ [ 2002.232843] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]528d348d-f685-306b-ff61-bd70ac77484a" [ 2002.232843] env[62824]: _type = "Task" [ 2002.232843] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2002.244088] env[62824]: DEBUG oslo_vmware.api [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]528d348d-f685-306b-ff61-bd70ac77484a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2002.313685] env[62824]: DEBUG oslo_concurrency.lockutils [req-1b1717df-d4f4-45d2-9d9f-44e82e64492f req-032e1ea9-440f-4810-b0aa-6eb122c60194 service nova] Releasing lock "refresh_cache-35f0435c-fe46-46ab-bc71-c21bb0ad3d55" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2002.554859] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Releasing lock "refresh_cache-e158b5d4-c120-4e6c-89c4-7668e097926d" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2002.555436] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2002.555747] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6c3481a5-bf52-483a-8632-a67277e3bc93 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2002.565554] env[62824]: DEBUG oslo_vmware.api [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Waiting for the task: (returnval){ [ 2002.565554] env[62824]: value = "task-2145983" [ 2002.565554] env[62824]: _type = "Task" [ 2002.565554] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2002.576803] env[62824]: DEBUG oslo_vmware.api [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Task: {'id': task-2145983, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2002.671384] env[62824]: DEBUG oslo_concurrency.lockutils [None req-847fb409-5740-4b52-92bf-c077485b2771 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "487c2c9d-2cd2-4912-8613-e1bfac732c40" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.924s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2002.692515] env[62824]: INFO nova.compute.manager [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 2afecec5-763d-4616-b690-41d3101cfc47] Took 40.67 seconds to build instance. [ 2002.726990] env[62824]: DEBUG oslo_vmware.api [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': task-2145982, 'name': Rename_Task, 'duration_secs': 0.143092} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2002.727874] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 41f068d3-2c8d-46f0-8d84-78531d28b0dd] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2002.729253] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-501f014e-4dca-4b59-9d3c-57c6baa3a70f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2002.741617] env[62824]: DEBUG oslo_vmware.api [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Waiting for the task: (returnval){ [ 2002.741617] env[62824]: value = "task-2145984" [ 2002.741617] env[62824]: _type = "Task" [ 2002.741617] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2002.755738] env[62824]: DEBUG oslo_vmware.api [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]528d348d-f685-306b-ff61-bd70ac77484a, 'name': SearchDatastore_Task, 'duration_secs': 0.010735} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2002.758437] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8b0dbbca-ab90-4630-bce9-8a858cc44dd5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2002.768340] env[62824]: DEBUG oslo_vmware.api [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': task-2145984, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2002.772691] env[62824]: DEBUG oslo_vmware.api [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Waiting for the task: (returnval){ [ 2002.772691] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52f585b6-109e-b8ad-9e24-6e334e3b992a" [ 2002.772691] env[62824]: _type = "Task" [ 2002.772691] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2002.781979] env[62824]: DEBUG oslo_vmware.api [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52f585b6-109e-b8ad-9e24-6e334e3b992a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2003.078917] env[62824]: DEBUG oslo_vmware.api [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Task: {'id': task-2145983, 'name': PowerOffVM_Task, 'duration_secs': 0.212749} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2003.080284] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2003.080677] env[62824]: DEBUG nova.virt.hardware [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:50:59Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='1bdf5fc7-1e8e-4668-88fe-6c47a0097f11',id=29,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-1536379991',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2003.081528] env[62824]: DEBUG nova.virt.hardware [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2003.081703] env[62824]: DEBUG nova.virt.hardware [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2003.081899] env[62824]: DEBUG nova.virt.hardware [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2003.082070] env[62824]: DEBUG nova.virt.hardware [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2003.082233] env[62824]: DEBUG nova.virt.hardware [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2003.082469] env[62824]: DEBUG nova.virt.hardware [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2003.082761] env[62824]: DEBUG nova.virt.hardware [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2003.083086] env[62824]: DEBUG nova.virt.hardware [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2003.083356] env[62824]: DEBUG nova.virt.hardware [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2003.083550] env[62824]: DEBUG nova.virt.hardware [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2003.090158] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-93b9fa2d-b18d-4b7f-b4bd-f38abb0d4829 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2003.106982] env[62824]: DEBUG oslo_vmware.api [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Waiting for the task: (returnval){ [ 2003.106982] env[62824]: value = "task-2145985" [ 2003.106982] env[62824]: _type = "Task" [ 2003.106982] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2003.115291] env[62824]: DEBUG oslo_vmware.api [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Task: {'id': task-2145985, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2003.195383] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e5b0c40e-1e7c-4d35-b0c4-0cc20d872b76 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Lock "2afecec5-763d-4616-b690-41d3101cfc47" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 112.412s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2003.257916] env[62824]: DEBUG oslo_vmware.api [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': task-2145984, 'name': PowerOnVM_Task} progress is 87%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2003.290662] env[62824]: DEBUG oslo_vmware.api [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52f585b6-109e-b8ad-9e24-6e334e3b992a, 'name': SearchDatastore_Task, 'duration_secs': 0.014926} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2003.292608] env[62824]: DEBUG oslo_concurrency.lockutils [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2003.295769] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 35f0435c-fe46-46ab-bc71-c21bb0ad3d55/35f0435c-fe46-46ab-bc71-c21bb0ad3d55.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2003.299640] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f1cfd985-97f0-48a1-8600-6339446c3aec {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2003.310074] env[62824]: DEBUG oslo_vmware.api [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Waiting for the task: (returnval){ [ 2003.310074] env[62824]: value = "task-2145986" [ 2003.310074] env[62824]: _type = "Task" [ 2003.310074] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2003.323567] env[62824]: DEBUG oslo_vmware.api [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Task: {'id': task-2145986, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2003.618337] env[62824]: DEBUG oslo_vmware.api [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Task: {'id': task-2145985, 'name': ReconfigVM_Task, 'duration_secs': 0.184076} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2003.621948] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cb490e6-afd3-43d5-bc6b-d882b6cb49cd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2003.641194] env[62824]: DEBUG nova.virt.hardware [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:50:59Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='1bdf5fc7-1e8e-4668-88fe-6c47a0097f11',id=29,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-1536379991',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2003.641561] env[62824]: DEBUG nova.virt.hardware [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2003.642311] env[62824]: DEBUG nova.virt.hardware [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2003.642886] env[62824]: DEBUG nova.virt.hardware [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2003.643150] env[62824]: DEBUG nova.virt.hardware [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2003.643794] env[62824]: DEBUG nova.virt.hardware [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2003.645312] env[62824]: DEBUG nova.virt.hardware [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2003.645499] env[62824]: DEBUG nova.virt.hardware [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2003.645677] env[62824]: DEBUG nova.virt.hardware [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2003.646151] env[62824]: DEBUG nova.virt.hardware [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2003.646151] env[62824]: DEBUG nova.virt.hardware [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2003.649637] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-93e9a460-cf2f-4f52-bdd4-0335e70e3671 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2003.656270] env[62824]: DEBUG oslo_vmware.api [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Waiting for the task: (returnval){ [ 2003.656270] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5259509b-093b-fb39-408d-378d584ed83c" [ 2003.656270] env[62824]: _type = "Task" [ 2003.656270] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2003.666603] env[62824]: DEBUG oslo_vmware.api [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5259509b-093b-fb39-408d-378d584ed83c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2003.698524] env[62824]: DEBUG nova.compute.manager [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 6a724332-a165-4a2b-9dd8-85f27e7b7637] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2003.701946] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9d3b7b1-1685-4c35-8900-c2c0f04b620a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2003.710477] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1e15a21-932c-404d-88ba-d50841629a35 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2003.748339] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6439d8c5-bcb7-48b8-97fb-2a83f62b26f8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2003.759694] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d92c7205-10d4-4a53-b261-9115f86626c9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2003.763889] env[62824]: DEBUG oslo_vmware.api [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': task-2145984, 'name': PowerOnVM_Task, 'duration_secs': 0.662853} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2003.764593] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 41f068d3-2c8d-46f0-8d84-78531d28b0dd] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2003.764685] env[62824]: INFO nova.compute.manager [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 41f068d3-2c8d-46f0-8d84-78531d28b0dd] Took 8.39 seconds to spawn the instance on the hypervisor. [ 2003.764807] env[62824]: DEBUG nova.compute.manager [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 41f068d3-2c8d-46f0-8d84-78531d28b0dd] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2003.765915] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adea5baf-a258-4a5e-8fe1-fa325e9d29f1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2003.778185] env[62824]: DEBUG nova.compute.provider_tree [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2003.797913] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f8aad8ab-6936-4c8d-8d26-3f40750d45f7 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Acquiring lock "9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2003.798161] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f8aad8ab-6936-4c8d-8d26-3f40750d45f7 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Lock "9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2003.798383] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f8aad8ab-6936-4c8d-8d26-3f40750d45f7 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Acquiring lock "9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2003.798566] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f8aad8ab-6936-4c8d-8d26-3f40750d45f7 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Lock "9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2003.798734] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f8aad8ab-6936-4c8d-8d26-3f40750d45f7 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Lock "9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2003.801564] env[62824]: INFO nova.compute.manager [None req-f8aad8ab-6936-4c8d-8d26-3f40750d45f7 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c] Terminating instance [ 2003.822273] env[62824]: DEBUG oslo_vmware.api [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Task: {'id': task-2145986, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2004.109882] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquiring lock "919e8a45-7810-4a8d-a1aa-5046b5ab059c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2004.110338] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "919e8a45-7810-4a8d-a1aa-5046b5ab059c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2004.168947] env[62824]: DEBUG oslo_vmware.api [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5259509b-093b-fb39-408d-378d584ed83c, 'name': SearchDatastore_Task, 'duration_secs': 0.024622} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2004.174507] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Reconfiguring VM instance instance-00000017 to detach disk 2000 {{(pid=62824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 2004.175279] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5c2a7c1e-b99d-4be1-8b05-b9bfe6c2654e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2004.194701] env[62824]: DEBUG oslo_vmware.api [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Waiting for the task: (returnval){ [ 2004.194701] env[62824]: value = "task-2145987" [ 2004.194701] env[62824]: _type = "Task" [ 2004.194701] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2004.204753] env[62824]: DEBUG oslo_vmware.api [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Task: {'id': task-2145987, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2004.221624] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2004.282127] env[62824]: DEBUG nova.scheduler.client.report [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2004.293579] env[62824]: INFO nova.compute.manager [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 41f068d3-2c8d-46f0-8d84-78531d28b0dd] Took 35.65 seconds to build instance. [ 2004.305314] env[62824]: DEBUG nova.compute.manager [None req-f8aad8ab-6936-4c8d-8d26-3f40750d45f7 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2004.305532] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-f8aad8ab-6936-4c8d-8d26-3f40750d45f7 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2004.306962] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b78ca17b-df60-4e3e-b7e1-b3f4b5d35a93 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2004.317705] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8aad8ab-6936-4c8d-8d26-3f40750d45f7 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2004.318370] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5dfbfd04-1200-44f5-86b8-69fbdf122c2f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2004.323256] env[62824]: DEBUG oslo_vmware.api [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Task: {'id': task-2145986, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.931805} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2004.323890] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 35f0435c-fe46-46ab-bc71-c21bb0ad3d55/35f0435c-fe46-46ab-bc71-c21bb0ad3d55.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2004.324475] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2004.324475] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dbf1b738-b1f0-4d47-962c-8cf3cfd3a301 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2004.328446] env[62824]: DEBUG oslo_vmware.api [None req-f8aad8ab-6936-4c8d-8d26-3f40750d45f7 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Waiting for the task: (returnval){ [ 2004.328446] env[62824]: value = "task-2145988" [ 2004.328446] env[62824]: _type = "Task" [ 2004.328446] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2004.333075] env[62824]: DEBUG oslo_vmware.api [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Waiting for the task: (returnval){ [ 2004.333075] env[62824]: value = "task-2145989" [ 2004.333075] env[62824]: _type = "Task" [ 2004.333075] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2004.339343] env[62824]: DEBUG oslo_vmware.api [None req-f8aad8ab-6936-4c8d-8d26-3f40750d45f7 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Task: {'id': task-2145988, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2004.344029] env[62824]: DEBUG oslo_vmware.api [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Task: {'id': task-2145989, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2004.515031] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Acquiring lock "15479062-af75-4925-99b3-77d6a49751ad" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2004.515294] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Lock "15479062-af75-4925-99b3-77d6a49751ad" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2004.705209] env[62824]: DEBUG oslo_vmware.api [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Task: {'id': task-2145987, 'name': ReconfigVM_Task, 'duration_secs': 0.374066} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2004.705529] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Reconfigured VM instance instance-00000017 to detach disk 2000 {{(pid=62824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 2004.706373] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3215c042-a974-44d2-bbed-15141e56eca0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2004.740020] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Reconfiguring VM instance instance-00000017 to attach disk [datastore2] e158b5d4-c120-4e6c-89c4-7668e097926d/e158b5d4-c120-4e6c-89c4-7668e097926d.vmdk or device None with type thin {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2004.740020] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e2fa5bc9-b6ea-4c31-82e2-2763c998b4fb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2004.760470] env[62824]: DEBUG oslo_vmware.api [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Waiting for the task: (returnval){ [ 2004.760470] env[62824]: value = "task-2145990" [ 2004.760470] env[62824]: _type = "Task" [ 2004.760470] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2004.773319] env[62824]: DEBUG oslo_vmware.api [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Task: {'id': task-2145990, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2004.786841] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.643s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2004.787406] env[62824]: DEBUG nova.compute.manager [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] [instance: db018af0-7bdf-40c5-b129-fde15dfb8ab4] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2004.790029] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0a53b27b-71f4-4fee-b778-9e0958eeed63 tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.695s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2004.790254] env[62824]: DEBUG nova.objects.instance [None req-0a53b27b-71f4-4fee-b778-9e0958eeed63 tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Lazy-loading 'resources' on Instance uuid 69bcac33-d6ac-4ee7-b674-58ab25389149 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2004.795992] env[62824]: DEBUG oslo_concurrency.lockutils [None req-407af591-05a3-45f7-9c44-9a58be2e59d4 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Lock "41f068d3-2c8d-46f0-8d84-78531d28b0dd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 113.398s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2004.841215] env[62824]: DEBUG oslo_vmware.api [None req-f8aad8ab-6936-4c8d-8d26-3f40750d45f7 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Task: {'id': task-2145988, 'name': PowerOffVM_Task, 'duration_secs': 0.276581} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2004.841881] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8aad8ab-6936-4c8d-8d26-3f40750d45f7 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2004.842042] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-f8aad8ab-6936-4c8d-8d26-3f40750d45f7 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2004.842306] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-61bc2096-7baa-4d95-a662-be3ca9cf1bc3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2004.846926] env[62824]: DEBUG oslo_vmware.api [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Task: {'id': task-2145989, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069775} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2004.847495] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2004.848261] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-380af41e-0432-4280-80e7-cf53e9c3cc5c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2004.869968] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Reconfiguring VM instance instance-0000002d to attach disk [datastore2] 35f0435c-fe46-46ab-bc71-c21bb0ad3d55/35f0435c-fe46-46ab-bc71-c21bb0ad3d55.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2004.870696] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b3d199d5-d0ed-4cb9-a51f-0634cb540950 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2004.890600] env[62824]: DEBUG oslo_vmware.api [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Waiting for the task: (returnval){ [ 2004.890600] env[62824]: value = "task-2145992" [ 2004.890600] env[62824]: _type = "Task" [ 2004.890600] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2004.898889] env[62824]: DEBUG oslo_vmware.api [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Task: {'id': task-2145992, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2004.939051] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-f8aad8ab-6936-4c8d-8d26-3f40750d45f7 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2004.939051] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-f8aad8ab-6936-4c8d-8d26-3f40750d45f7 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c] Deleting contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2004.939051] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-f8aad8ab-6936-4c8d-8d26-3f40750d45f7 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Deleting the datastore file [datastore1] 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2004.939051] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-97d8521e-4d5b-4722-bdfe-843db90100f6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2004.945381] env[62824]: DEBUG oslo_vmware.api [None req-f8aad8ab-6936-4c8d-8d26-3f40750d45f7 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Waiting for the task: (returnval){ [ 2004.945381] env[62824]: value = "task-2145993" [ 2004.945381] env[62824]: _type = "Task" [ 2004.945381] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2004.953488] env[62824]: DEBUG oslo_vmware.api [None req-f8aad8ab-6936-4c8d-8d26-3f40750d45f7 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Task: {'id': task-2145993, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2005.272847] env[62824]: DEBUG oslo_vmware.api [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Task: {'id': task-2145990, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2005.293761] env[62824]: DEBUG nova.compute.utils [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2005.298739] env[62824]: DEBUG nova.compute.manager [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] [instance: db018af0-7bdf-40c5-b129-fde15dfb8ab4] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2005.299165] env[62824]: DEBUG nova.network.neutron [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] [instance: db018af0-7bdf-40c5-b129-fde15dfb8ab4] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2005.303863] env[62824]: DEBUG nova.compute.manager [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 21c05725-3dad-4965-98d0-0622ebcc8ece] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2005.391132] env[62824]: DEBUG nova.policy [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6afdbefb672f4b77b0e708aa95218121', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b7f15dcd636744cbb08ec8c2f4ef36a2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2005.401909] env[62824]: DEBUG oslo_vmware.api [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Task: {'id': task-2145992, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2005.456773] env[62824]: DEBUG oslo_vmware.api [None req-f8aad8ab-6936-4c8d-8d26-3f40750d45f7 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Task: {'id': task-2145993, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2005.770560] env[62824]: DEBUG oslo_vmware.api [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Task: {'id': task-2145990, 'name': ReconfigVM_Task, 'duration_secs': 0.685943} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2005.770960] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Reconfigured VM instance instance-00000017 to attach disk [datastore2] e158b5d4-c120-4e6c-89c4-7668e097926d/e158b5d4-c120-4e6c-89c4-7668e097926d.vmdk or device None with type thin {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2005.774180] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bef49a1-45ab-459a-80d5-ba37b2afb167 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2005.798644] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-876ad538-e945-4c52-819a-a998520dab97 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2005.801653] env[62824]: DEBUG nova.compute.manager [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] [instance: db018af0-7bdf-40c5-b129-fde15dfb8ab4] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2005.823028] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2005.827499] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fb240df-544d-41b3-b3b2-ac4c93ae7093 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2005.848486] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f463eaeb-caa6-4dfa-adc8-5b36b3a5da78 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2005.851958] env[62824]: DEBUG nova.network.neutron [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] [instance: db018af0-7bdf-40c5-b129-fde15dfb8ab4] Successfully created port: d94a1564-ae07-4d73-bae0-2cf718dc318c {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2005.859240] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2005.859525] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3f57f760-07f8-4c87-8575-cae01e4b8f6a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2005.866485] env[62824]: DEBUG oslo_vmware.api [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Waiting for the task: (returnval){ [ 2005.866485] env[62824]: value = "task-2145994" [ 2005.866485] env[62824]: _type = "Task" [ 2005.866485] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2005.877198] env[62824]: DEBUG oslo_vmware.api [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Task: {'id': task-2145994, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2005.904831] env[62824]: DEBUG oslo_vmware.api [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Task: {'id': task-2145992, 'name': ReconfigVM_Task, 'duration_secs': 0.865262} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2005.905262] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Reconfigured VM instance instance-0000002d to attach disk [datastore2] 35f0435c-fe46-46ab-bc71-c21bb0ad3d55/35f0435c-fe46-46ab-bc71-c21bb0ad3d55.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2005.906009] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1da4b551-d0a6-44ff-b199-8cedbb11cef1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2005.916685] env[62824]: DEBUG oslo_vmware.api [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Waiting for the task: (returnval){ [ 2005.916685] env[62824]: value = "task-2145995" [ 2005.916685] env[62824]: _type = "Task" [ 2005.916685] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2005.925770] env[62824]: DEBUG oslo_vmware.api [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Task: {'id': task-2145995, 'name': Rename_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2005.944343] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7a9a279-89db-4022-92c4-83e03455ed48 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2005.955689] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b696ca1-f7f5-48df-b8ca-accd0f6fa371 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2005.962890] env[62824]: DEBUG oslo_vmware.api [None req-f8aad8ab-6936-4c8d-8d26-3f40750d45f7 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Task: {'id': task-2145993, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.596885} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2005.964035] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-f8aad8ab-6936-4c8d-8d26-3f40750d45f7 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2005.964180] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-f8aad8ab-6936-4c8d-8d26-3f40750d45f7 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c] Deleted contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2005.964345] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-f8aad8ab-6936-4c8d-8d26-3f40750d45f7 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2005.964519] env[62824]: INFO nova.compute.manager [None req-f8aad8ab-6936-4c8d-8d26-3f40750d45f7 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c] Took 1.66 seconds to destroy the instance on the hypervisor. [ 2005.964784] env[62824]: DEBUG oslo.service.loopingcall [None req-f8aad8ab-6936-4c8d-8d26-3f40750d45f7 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2005.965010] env[62824]: DEBUG nova.compute.manager [-] [instance: 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2005.965140] env[62824]: DEBUG nova.network.neutron [-] [instance: 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2005.994166] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cd457b6-1a39-40a5-af88-15122410c833 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2006.003431] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f10a6343-a9e4-4adb-bc87-ba25b16640a6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2006.018431] env[62824]: DEBUG nova.compute.provider_tree [None req-0a53b27b-71f4-4fee-b778-9e0958eeed63 tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2006.379561] env[62824]: DEBUG oslo_vmware.api [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Task: {'id': task-2145994, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2006.432450] env[62824]: DEBUG oslo_vmware.api [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Task: {'id': task-2145995, 'name': Rename_Task, 'duration_secs': 0.162845} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2006.432864] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2006.433175] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-026826b6-f733-4f3b-ba30-b1384a9aac25 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2006.440213] env[62824]: DEBUG oslo_vmware.api [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Waiting for the task: (returnval){ [ 2006.440213] env[62824]: value = "task-2145996" [ 2006.440213] env[62824]: _type = "Task" [ 2006.440213] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2006.448132] env[62824]: DEBUG oslo_vmware.api [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Task: {'id': task-2145996, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2006.523288] env[62824]: DEBUG nova.scheduler.client.report [None req-0a53b27b-71f4-4fee-b778-9e0958eeed63 tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2006.815895] env[62824]: DEBUG nova.compute.manager [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] [instance: db018af0-7bdf-40c5-b129-fde15dfb8ab4] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2006.842885] env[62824]: DEBUG nova.virt.hardware [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2006.843417] env[62824]: DEBUG nova.virt.hardware [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2006.843676] env[62824]: DEBUG nova.virt.hardware [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2006.844060] env[62824]: DEBUG nova.virt.hardware [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2006.844346] env[62824]: DEBUG nova.virt.hardware [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2006.844597] env[62824]: DEBUG nova.virt.hardware [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2006.845057] env[62824]: DEBUG nova.virt.hardware [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2006.845283] env[62824]: DEBUG nova.virt.hardware [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2006.845495] env[62824]: DEBUG nova.virt.hardware [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2006.845698] env[62824]: DEBUG nova.virt.hardware [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2006.845910] env[62824]: DEBUG nova.virt.hardware [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2006.846891] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f14c6855-4393-4ad6-9e70-3ce1b1c3df58 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2006.855134] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1127334d-db08-4c35-8832-ac95ab7eee75 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2006.886243] env[62824]: DEBUG oslo_vmware.api [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Task: {'id': task-2145994, 'name': PowerOnVM_Task, 'duration_secs': 0.55506} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2006.886243] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2006.899460] env[62824]: DEBUG oslo_concurrency.lockutils [None req-33e5ba91-26b9-4a1a-965a-7b34fe0c5be2 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Acquiring lock "8bb90a4d-93c9-4f54-b15e-48fe966e6c97" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2006.900241] env[62824]: DEBUG oslo_concurrency.lockutils [None req-33e5ba91-26b9-4a1a-965a-7b34fe0c5be2 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Lock "8bb90a4d-93c9-4f54-b15e-48fe966e6c97" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2006.900477] env[62824]: DEBUG nova.compute.manager [None req-33e5ba91-26b9-4a1a-965a-7b34fe0c5be2 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 8bb90a4d-93c9-4f54-b15e-48fe966e6c97] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2006.901450] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8b00240-5f12-4b0d-b0ec-6f4e88171506 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2006.909256] env[62824]: DEBUG nova.compute.manager [None req-33e5ba91-26b9-4a1a-965a-7b34fe0c5be2 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 8bb90a4d-93c9-4f54-b15e-48fe966e6c97] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62824) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3404}} [ 2006.909894] env[62824]: DEBUG nova.objects.instance [None req-33e5ba91-26b9-4a1a-965a-7b34fe0c5be2 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Lazy-loading 'flavor' on Instance uuid 8bb90a4d-93c9-4f54-b15e-48fe966e6c97 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2006.951105] env[62824]: DEBUG oslo_vmware.api [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Task: {'id': task-2145996, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2006.987262] env[62824]: DEBUG nova.compute.manager [req-9439ca54-b5c8-45e0-9440-cb5039e89cc5 req-87f4aff4-f729-4dd1-a457-f1e1070af958 service nova] [instance: 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c] Received event network-vif-deleted-084f765d-6af5-4867-bb76-af443b0bf834 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2006.987262] env[62824]: INFO nova.compute.manager [req-9439ca54-b5c8-45e0-9440-cb5039e89cc5 req-87f4aff4-f729-4dd1-a457-f1e1070af958 service nova] [instance: 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c] Neutron deleted interface 084f765d-6af5-4867-bb76-af443b0bf834; detaching it from the instance and deleting it from the info cache [ 2006.987262] env[62824]: DEBUG nova.network.neutron [req-9439ca54-b5c8-45e0-9440-cb5039e89cc5 req-87f4aff4-f729-4dd1-a457-f1e1070af958 service nova] [instance: 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2007.034918] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0a53b27b-71f4-4fee-b778-9e0958eeed63 tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.242s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2007.035880] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 24.325s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2007.059817] env[62824]: INFO nova.scheduler.client.report [None req-0a53b27b-71f4-4fee-b778-9e0958eeed63 tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Deleted allocations for instance 69bcac33-d6ac-4ee7-b674-58ab25389149 [ 2007.100463] env[62824]: DEBUG nova.network.neutron [-] [instance: 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2007.453455] env[62824]: DEBUG oslo_vmware.api [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Task: {'id': task-2145996, 'name': PowerOnVM_Task, 'duration_secs': 0.672854} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2007.453788] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2007.454029] env[62824]: INFO nova.compute.manager [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Took 9.42 seconds to spawn the instance on the hypervisor. [ 2007.454701] env[62824]: DEBUG nova.compute.manager [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2007.455559] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3d11ac8-9e08-43d7-90c3-2d4d6e49f70a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2007.490606] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a3c97e76-ff7d-4fd8-87ef-e5487d194d7a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2007.501677] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5baa7a39-09a1-4f13-bf5c-e5cd1db832a7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2007.541148] env[62824]: DEBUG nova.compute.manager [req-9439ca54-b5c8-45e0-9440-cb5039e89cc5 req-87f4aff4-f729-4dd1-a457-f1e1070af958 service nova] [instance: 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c] Detach interface failed, port_id=084f765d-6af5-4867-bb76-af443b0bf834, reason: Instance 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c could not be found. {{(pid=62824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11482}} [ 2007.543781] env[62824]: DEBUG nova.compute.manager [req-82e4c125-cf69-4f32-93eb-d597b0320bae req-1460201d-4e16-4978-a9b1-19c2b66a84d7 service nova] [instance: db018af0-7bdf-40c5-b129-fde15dfb8ab4] Received event network-vif-plugged-d94a1564-ae07-4d73-bae0-2cf718dc318c {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2007.543987] env[62824]: DEBUG oslo_concurrency.lockutils [req-82e4c125-cf69-4f32-93eb-d597b0320bae req-1460201d-4e16-4978-a9b1-19c2b66a84d7 service nova] Acquiring lock "db018af0-7bdf-40c5-b129-fde15dfb8ab4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2007.544391] env[62824]: DEBUG oslo_concurrency.lockutils [req-82e4c125-cf69-4f32-93eb-d597b0320bae req-1460201d-4e16-4978-a9b1-19c2b66a84d7 service nova] Lock "db018af0-7bdf-40c5-b129-fde15dfb8ab4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2007.544571] env[62824]: DEBUG oslo_concurrency.lockutils [req-82e4c125-cf69-4f32-93eb-d597b0320bae req-1460201d-4e16-4978-a9b1-19c2b66a84d7 service nova] Lock "db018af0-7bdf-40c5-b129-fde15dfb8ab4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2007.544736] env[62824]: DEBUG nova.compute.manager [req-82e4c125-cf69-4f32-93eb-d597b0320bae req-1460201d-4e16-4978-a9b1-19c2b66a84d7 service nova] [instance: db018af0-7bdf-40c5-b129-fde15dfb8ab4] No waiting events found dispatching network-vif-plugged-d94a1564-ae07-4d73-bae0-2cf718dc318c {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2007.544898] env[62824]: WARNING nova.compute.manager [req-82e4c125-cf69-4f32-93eb-d597b0320bae req-1460201d-4e16-4978-a9b1-19c2b66a84d7 service nova] [instance: db018af0-7bdf-40c5-b129-fde15dfb8ab4] Received unexpected event network-vif-plugged-d94a1564-ae07-4d73-bae0-2cf718dc318c for instance with vm_state building and task_state spawning. [ 2007.569092] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0a53b27b-71f4-4fee-b778-9e0958eeed63 tempest-ServerAddressesTestJSON-1899557317 tempest-ServerAddressesTestJSON-1899557317-project-member] Lock "69bcac33-d6ac-4ee7-b674-58ab25389149" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.649s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2007.605745] env[62824]: INFO nova.compute.manager [-] [instance: 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c] Took 1.64 seconds to deallocate network for instance. [ 2007.637125] env[62824]: DEBUG nova.network.neutron [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] [instance: db018af0-7bdf-40c5-b129-fde15dfb8ab4] Successfully updated port: d94a1564-ae07-4d73-bae0-2cf718dc318c {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2007.898496] env[62824]: INFO nova.compute.manager [None req-b7da11ec-40a4-45e2-b086-d44fa99c0537 tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Updating instance to original state: 'active' [ 2007.916462] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-33e5ba91-26b9-4a1a-965a-7b34fe0c5be2 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 8bb90a4d-93c9-4f54-b15e-48fe966e6c97] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2007.916927] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d1caf955-a925-4b5f-a278-b8a5b66b8e4f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2007.924540] env[62824]: DEBUG oslo_vmware.api [None req-33e5ba91-26b9-4a1a-965a-7b34fe0c5be2 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Waiting for the task: (returnval){ [ 2007.924540] env[62824]: value = "task-2145997" [ 2007.924540] env[62824]: _type = "Task" [ 2007.924540] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2007.934673] env[62824]: DEBUG oslo_vmware.api [None req-33e5ba91-26b9-4a1a-965a-7b34fe0c5be2 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': task-2145997, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2007.976087] env[62824]: INFO nova.compute.manager [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Took 37.60 seconds to build instance. [ 2008.084044] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 29b8dd5f-1855-490c-a01b-54840073a753 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2008.084296] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 82382932-7302-4441-a6f8-9aa2300ec0f6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2008.084514] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance f4d63a93-23af-470c-b36a-662af81dc386 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2008.084593] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2008.084755] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2008.084939] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 07ce931f-45ef-409b-b714-9f1cd47a3a88 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2008.085122] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance f3740d9a-b21a-4a79-9e28-2a89ecd08bb6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2008.085296] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance b8cc8cd3-ea03-40bf-b867-7ad193365552 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2008.085455] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance c36ece43-3d70-4e67-a740-9057f413c722 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2008.085602] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance b96d1351-f3a7-4bac-998c-a34ab2606041 actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2008.085743] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance b3003c4b-ae5a-48df-8c12-a915a76253f4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2008.085881] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 6d80ec06-8559-4964-8577-a2512aa366ed actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2008.086071] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 59b5b883-4188-471c-8862-444f3ce08cb0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2008.086235] env[62824]: WARNING nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 5f1237ff-5841-41f1-8d77-df9fabaa62b6 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 2008.086460] env[62824]: WARNING nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance dc3317e4-6958-4df6-8e39-c89b4a5d861c is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 2008.086617] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance bbf1440b-8681-48c8-a178-9a83b925c695 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2008.086762] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 7febb8b4-f984-4d79-a888-b2829f2a9df6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2008.086939] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2008.087085] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 8bb90a4d-93c9-4f54-b15e-48fe966e6c97 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2008.087254] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 2afecec5-763d-4616-b690-41d3101cfc47 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2008.087404] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 41f068d3-2c8d-46f0-8d84-78531d28b0dd actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2008.087564] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 35f0435c-fe46-46ab-bc71-c21bb0ad3d55 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2008.087707] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance db018af0-7bdf-40c5-b129-fde15dfb8ab4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2008.087826] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance e158b5d4-c120-4e6c-89c4-7668e097926d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2008.113323] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f8aad8ab-6936-4c8d-8d26-3f40750d45f7 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2008.140094] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Acquiring lock "refresh_cache-db018af0-7bdf-40c5-b129-fde15dfb8ab4" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2008.140228] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Acquired lock "refresh_cache-db018af0-7bdf-40c5-b129-fde15dfb8ab4" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2008.140380] env[62824]: DEBUG nova.network.neutron [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] [instance: db018af0-7bdf-40c5-b129-fde15dfb8ab4] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2008.439096] env[62824]: DEBUG oslo_vmware.api [None req-33e5ba91-26b9-4a1a-965a-7b34fe0c5be2 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': task-2145997, 'name': PowerOffVM_Task, 'duration_secs': 0.298489} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2008.439096] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-33e5ba91-26b9-4a1a-965a-7b34fe0c5be2 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 8bb90a4d-93c9-4f54-b15e-48fe966e6c97] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2008.439096] env[62824]: DEBUG nova.compute.manager [None req-33e5ba91-26b9-4a1a-965a-7b34fe0c5be2 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 8bb90a4d-93c9-4f54-b15e-48fe966e6c97] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2008.440149] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3756c7cd-3ad2-4073-89e3-98178bfe3d4a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2008.479712] env[62824]: DEBUG oslo_concurrency.lockutils [None req-93d412de-7817-4091-8b1c-73409225eeae tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Lock "35f0435c-fe46-46ab-bc71-c21bb0ad3d55" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 109.443s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2008.591602] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance b9f30f87-3594-4468-9d29-70890d8761e3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 2008.702043] env[62824]: DEBUG nova.network.neutron [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] [instance: db018af0-7bdf-40c5-b129-fde15dfb8ab4] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2008.953327] env[62824]: DEBUG oslo_concurrency.lockutils [None req-33e5ba91-26b9-4a1a-965a-7b34fe0c5be2 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Lock "8bb90a4d-93c9-4f54-b15e-48fe966e6c97" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.053s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2008.968876] env[62824]: DEBUG nova.network.neutron [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] [instance: db018af0-7bdf-40c5-b129-fde15dfb8ab4] Updating instance_info_cache with network_info: [{"id": "d94a1564-ae07-4d73-bae0-2cf718dc318c", "address": "fa:16:3e:2d:d1:35", "network": {"id": "b1f70fef-09fd-4c2c-9225-64a9bf881f51", "bridge": "br-int", "label": "tempest-ImagesNegativeTestJSON-156503515-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b7f15dcd636744cbb08ec8c2f4ef36a2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "00a15667-7ca5-4dc9-be92-164750d87988", "external-id": "nsx-vlan-transportzone-933", "segmentation_id": 933, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd94a1564-ae", "ovs_interfaceid": "d94a1564-ae07-4d73-bae0-2cf718dc318c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2008.983401] env[62824]: DEBUG nova.compute.manager [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] [instance: ed7d3e95-f26e-40fd-a370-e17922bbff8e] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2009.095145] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 02405b17-7022-4e2c-a357-415de88f63d0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 2009.472066] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Releasing lock "refresh_cache-db018af0-7bdf-40c5-b129-fde15dfb8ab4" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2009.472358] env[62824]: DEBUG nova.compute.manager [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] [instance: db018af0-7bdf-40c5-b129-fde15dfb8ab4] Instance network_info: |[{"id": "d94a1564-ae07-4d73-bae0-2cf718dc318c", "address": "fa:16:3e:2d:d1:35", "network": {"id": "b1f70fef-09fd-4c2c-9225-64a9bf881f51", "bridge": "br-int", "label": "tempest-ImagesNegativeTestJSON-156503515-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b7f15dcd636744cbb08ec8c2f4ef36a2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "00a15667-7ca5-4dc9-be92-164750d87988", "external-id": "nsx-vlan-transportzone-933", "segmentation_id": 933, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd94a1564-ae", "ovs_interfaceid": "d94a1564-ae07-4d73-bae0-2cf718dc318c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2009.472778] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] [instance: db018af0-7bdf-40c5-b129-fde15dfb8ab4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2d:d1:35', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '00a15667-7ca5-4dc9-be92-164750d87988', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd94a1564-ae07-4d73-bae0-2cf718dc318c', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2009.482755] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Creating folder: Project (b7f15dcd636744cbb08ec8c2f4ef36a2). Parent ref: group-v438503. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 2009.483260] env[62824]: DEBUG oslo_concurrency.lockutils [None req-6008bdc7-ed30-4e87-8975-1eeef6a5d44f tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Acquiring lock "e158b5d4-c120-4e6c-89c4-7668e097926d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2009.486051] env[62824]: DEBUG oslo_concurrency.lockutils [None req-6008bdc7-ed30-4e87-8975-1eeef6a5d44f tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Lock "e158b5d4-c120-4e6c-89c4-7668e097926d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.003s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2009.486328] env[62824]: DEBUG oslo_concurrency.lockutils [None req-6008bdc7-ed30-4e87-8975-1eeef6a5d44f tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Acquiring lock "e158b5d4-c120-4e6c-89c4-7668e097926d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2009.486522] env[62824]: DEBUG oslo_concurrency.lockutils [None req-6008bdc7-ed30-4e87-8975-1eeef6a5d44f tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Lock "e158b5d4-c120-4e6c-89c4-7668e097926d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2009.486689] env[62824]: DEBUG oslo_concurrency.lockutils [None req-6008bdc7-ed30-4e87-8975-1eeef6a5d44f tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Lock "e158b5d4-c120-4e6c-89c4-7668e097926d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2009.488271] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e2ca9098-963d-426c-8b26-2050b29555e3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2009.491388] env[62824]: INFO nova.compute.manager [None req-6008bdc7-ed30-4e87-8975-1eeef6a5d44f tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Terminating instance [ 2009.506518] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Created folder: Project (b7f15dcd636744cbb08ec8c2f4ef36a2) in parent group-v438503. [ 2009.507804] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Creating folder: Instances. Parent ref: group-v438609. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 2009.507804] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0dd5680f-73e5-49a5-844f-46cd0c4f00cd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2009.517031] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Created folder: Instances in parent group-v438609. [ 2009.517570] env[62824]: DEBUG oslo.service.loopingcall [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2009.518243] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2009.518438] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: db018af0-7bdf-40c5-b129-fde15dfb8ab4] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2009.519141] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e5acc452-eec7-4228-b317-9613e8899717 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2009.540838] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2009.540838] env[62824]: value = "task-2146000" [ 2009.540838] env[62824]: _type = "Task" [ 2009.540838] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2009.548740] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146000, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2009.572242] env[62824]: DEBUG nova.compute.manager [req-d361411a-6ee2-4c93-a40f-35c2a4b2b0a7 req-fcbf1b7b-c5d0-4e1d-9f47-a6f3970a7510 service nova] [instance: db018af0-7bdf-40c5-b129-fde15dfb8ab4] Received event network-changed-d94a1564-ae07-4d73-bae0-2cf718dc318c {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2009.572476] env[62824]: DEBUG nova.compute.manager [req-d361411a-6ee2-4c93-a40f-35c2a4b2b0a7 req-fcbf1b7b-c5d0-4e1d-9f47-a6f3970a7510 service nova] [instance: db018af0-7bdf-40c5-b129-fde15dfb8ab4] Refreshing instance network info cache due to event network-changed-d94a1564-ae07-4d73-bae0-2cf718dc318c. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2009.572695] env[62824]: DEBUG oslo_concurrency.lockutils [req-d361411a-6ee2-4c93-a40f-35c2a4b2b0a7 req-fcbf1b7b-c5d0-4e1d-9f47-a6f3970a7510 service nova] Acquiring lock "refresh_cache-db018af0-7bdf-40c5-b129-fde15dfb8ab4" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2009.572948] env[62824]: DEBUG oslo_concurrency.lockutils [req-d361411a-6ee2-4c93-a40f-35c2a4b2b0a7 req-fcbf1b7b-c5d0-4e1d-9f47-a6f3970a7510 service nova] Acquired lock "refresh_cache-db018af0-7bdf-40c5-b129-fde15dfb8ab4" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2009.573040] env[62824]: DEBUG nova.network.neutron [req-d361411a-6ee2-4c93-a40f-35c2a4b2b0a7 req-fcbf1b7b-c5d0-4e1d-9f47-a6f3970a7510 service nova] [instance: db018af0-7bdf-40c5-b129-fde15dfb8ab4] Refreshing network info cache for port d94a1564-ae07-4d73-bae0-2cf718dc318c {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2009.575017] env[62824]: DEBUG nova.compute.manager [None req-58bcc65c-ba1f-4462-9f05-d0e06434f495 tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2009.576047] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00093c34-89aa-4b26-b852-de0310b34265 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2009.599169] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance c65ccd20-b682-420a-9c1a-47a45959197e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 2009.861230] env[62824]: DEBUG nova.objects.instance [None req-26bb01c0-bc09-4b0e-b772-39e2968113f7 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Lazy-loading 'flavor' on Instance uuid 8bb90a4d-93c9-4f54-b15e-48fe966e6c97 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2009.999607] env[62824]: DEBUG nova.compute.manager [None req-6008bdc7-ed30-4e87-8975-1eeef6a5d44f tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2009.999607] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-6008bdc7-ed30-4e87-8975-1eeef6a5d44f tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2010.000466] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d6f04b8-46cd-436a-85fe-2a77c78e4918 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2010.009962] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-6008bdc7-ed30-4e87-8975-1eeef6a5d44f tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2010.010295] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c10dc49d-e174-4104-a8f5-050aaa5e6b6e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2010.019026] env[62824]: DEBUG oslo_vmware.api [None req-6008bdc7-ed30-4e87-8975-1eeef6a5d44f tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Waiting for the task: (returnval){ [ 2010.019026] env[62824]: value = "task-2146001" [ 2010.019026] env[62824]: _type = "Task" [ 2010.019026] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2010.034106] env[62824]: DEBUG oslo_vmware.api [None req-6008bdc7-ed30-4e87-8975-1eeef6a5d44f tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Task: {'id': task-2146001, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2010.052243] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146000, 'name': CreateVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2010.087413] env[62824]: INFO nova.compute.manager [None req-58bcc65c-ba1f-4462-9f05-d0e06434f495 tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] instance snapshotting [ 2010.092593] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5452a4f9-73c2-4231-997e-1fa555467b33 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2010.114061] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 4423d172-acdd-4c69-b3b8-ff166e1b8548 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 2010.116372] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-247d6069-b503-4b4b-abce-2f25675a9d10 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2010.366661] env[62824]: DEBUG oslo_concurrency.lockutils [None req-26bb01c0-bc09-4b0e-b772-39e2968113f7 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Acquiring lock "refresh_cache-8bb90a4d-93c9-4f54-b15e-48fe966e6c97" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2010.366840] env[62824]: DEBUG oslo_concurrency.lockutils [None req-26bb01c0-bc09-4b0e-b772-39e2968113f7 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Acquired lock "refresh_cache-8bb90a4d-93c9-4f54-b15e-48fe966e6c97" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2010.367029] env[62824]: DEBUG nova.network.neutron [None req-26bb01c0-bc09-4b0e-b772-39e2968113f7 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 8bb90a4d-93c9-4f54-b15e-48fe966e6c97] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2010.367211] env[62824]: DEBUG nova.objects.instance [None req-26bb01c0-bc09-4b0e-b772-39e2968113f7 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Lazy-loading 'info_cache' on Instance uuid 8bb90a4d-93c9-4f54-b15e-48fe966e6c97 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2010.470124] env[62824]: DEBUG nova.network.neutron [req-d361411a-6ee2-4c93-a40f-35c2a4b2b0a7 req-fcbf1b7b-c5d0-4e1d-9f47-a6f3970a7510 service nova] [instance: db018af0-7bdf-40c5-b129-fde15dfb8ab4] Updated VIF entry in instance network info cache for port d94a1564-ae07-4d73-bae0-2cf718dc318c. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2010.470434] env[62824]: DEBUG nova.network.neutron [req-d361411a-6ee2-4c93-a40f-35c2a4b2b0a7 req-fcbf1b7b-c5d0-4e1d-9f47-a6f3970a7510 service nova] [instance: db018af0-7bdf-40c5-b129-fde15dfb8ab4] Updating instance_info_cache with network_info: [{"id": "d94a1564-ae07-4d73-bae0-2cf718dc318c", "address": "fa:16:3e:2d:d1:35", "network": {"id": "b1f70fef-09fd-4c2c-9225-64a9bf881f51", "bridge": "br-int", "label": "tempest-ImagesNegativeTestJSON-156503515-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b7f15dcd636744cbb08ec8c2f4ef36a2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "00a15667-7ca5-4dc9-be92-164750d87988", "external-id": "nsx-vlan-transportzone-933", "segmentation_id": 933, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd94a1564-ae", "ovs_interfaceid": "d94a1564-ae07-4d73-bae0-2cf718dc318c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2010.531835] env[62824]: DEBUG oslo_vmware.api [None req-6008bdc7-ed30-4e87-8975-1eeef6a5d44f tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Task: {'id': task-2146001, 'name': PowerOffVM_Task, 'duration_secs': 0.230357} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2010.532209] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-6008bdc7-ed30-4e87-8975-1eeef6a5d44f tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2010.532404] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-6008bdc7-ed30-4e87-8975-1eeef6a5d44f tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2010.532529] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dd3d89d6-a3d2-4700-953d-182a977a7a40 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2010.553328] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146000, 'name': CreateVM_Task, 'duration_secs': 0.556086} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2010.553559] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: db018af0-7bdf-40c5-b129-fde15dfb8ab4] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2010.554394] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2010.554464] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2010.555513] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2010.555513] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b6383cea-2b1c-4fb1-950e-d3f864bef3d9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2010.562750] env[62824]: DEBUG oslo_vmware.api [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Waiting for the task: (returnval){ [ 2010.562750] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52840321-cd87-3c37-3937-e8ab876336f4" [ 2010.562750] env[62824]: _type = "Task" [ 2010.562750] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2010.570818] env[62824]: DEBUG oslo_vmware.api [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52840321-cd87-3c37-3937-e8ab876336f4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2010.620771] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 88bad2dd-dce7-41df-b56c-93a5d054c11e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 2010.631413] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-58bcc65c-ba1f-4462-9f05-d0e06434f495 tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Creating Snapshot of the VM instance {{(pid=62824) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 2010.631721] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-d7d0be96-70c4-4191-ae74-3158325c43b7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2010.640402] env[62824]: DEBUG oslo_vmware.api [None req-58bcc65c-ba1f-4462-9f05-d0e06434f495 tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Waiting for the task: (returnval){ [ 2010.640402] env[62824]: value = "task-2146003" [ 2010.640402] env[62824]: _type = "Task" [ 2010.640402] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2010.649775] env[62824]: DEBUG oslo_vmware.api [None req-58bcc65c-ba1f-4462-9f05-d0e06434f495 tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Task: {'id': task-2146003, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2010.712062] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-6008bdc7-ed30-4e87-8975-1eeef6a5d44f tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2010.712366] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-6008bdc7-ed30-4e87-8975-1eeef6a5d44f tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Deleting contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2010.712562] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-6008bdc7-ed30-4e87-8975-1eeef6a5d44f tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Deleting the datastore file [datastore2] e158b5d4-c120-4e6c-89c4-7668e097926d {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2010.712840] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-940b97b6-6561-42f4-bdcb-a1aa94ef784a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2010.719437] env[62824]: DEBUG oslo_vmware.api [None req-6008bdc7-ed30-4e87-8975-1eeef6a5d44f tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Waiting for the task: (returnval){ [ 2010.719437] env[62824]: value = "task-2146004" [ 2010.719437] env[62824]: _type = "Task" [ 2010.719437] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2010.727733] env[62824]: DEBUG oslo_vmware.api [None req-6008bdc7-ed30-4e87-8975-1eeef6a5d44f tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Task: {'id': task-2146004, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2010.870755] env[62824]: DEBUG nova.objects.base [None req-26bb01c0-bc09-4b0e-b772-39e2968113f7 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Object Instance<8bb90a4d-93c9-4f54-b15e-48fe966e6c97> lazy-loaded attributes: flavor,info_cache {{(pid=62824) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 2010.972607] env[62824]: DEBUG oslo_concurrency.lockutils [req-d361411a-6ee2-4c93-a40f-35c2a4b2b0a7 req-fcbf1b7b-c5d0-4e1d-9f47-a6f3970a7510 service nova] Releasing lock "refresh_cache-db018af0-7bdf-40c5-b129-fde15dfb8ab4" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2011.072649] env[62824]: DEBUG oslo_vmware.api [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52840321-cd87-3c37-3937-e8ab876336f4, 'name': SearchDatastore_Task, 'duration_secs': 0.028328} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2011.072971] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2011.073219] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] [instance: db018af0-7bdf-40c5-b129-fde15dfb8ab4] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2011.073506] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2011.073691] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2011.073902] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2011.074577] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ea5f03a4-6eaa-41d3-ba15-ba37ef32e85a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2011.084015] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2011.084977] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2011.084977] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d0ef3d83-91c8-4d23-af56-6fcdcd995cfd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2011.091199] env[62824]: DEBUG oslo_vmware.api [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Waiting for the task: (returnval){ [ 2011.091199] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5253a17e-32d4-acd5-3fe4-21b11d5f639d" [ 2011.091199] env[62824]: _type = "Task" [ 2011.091199] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2011.099411] env[62824]: DEBUG oslo_vmware.api [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5253a17e-32d4-acd5-3fe4-21b11d5f639d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2011.123390] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance c63b05b7-1dbf-4637-8728-33673be0d3ea has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 2011.152015] env[62824]: DEBUG oslo_vmware.api [None req-58bcc65c-ba1f-4462-9f05-d0e06434f495 tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Task: {'id': task-2146003, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2011.229705] env[62824]: DEBUG oslo_vmware.api [None req-6008bdc7-ed30-4e87-8975-1eeef6a5d44f tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Task: {'id': task-2146004, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.229869} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2011.229962] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-6008bdc7-ed30-4e87-8975-1eeef6a5d44f tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2011.230176] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-6008bdc7-ed30-4e87-8975-1eeef6a5d44f tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Deleted contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2011.230422] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-6008bdc7-ed30-4e87-8975-1eeef6a5d44f tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2011.230693] env[62824]: INFO nova.compute.manager [None req-6008bdc7-ed30-4e87-8975-1eeef6a5d44f tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Took 1.23 seconds to destroy the instance on the hypervisor. [ 2011.231014] env[62824]: DEBUG oslo.service.loopingcall [None req-6008bdc7-ed30-4e87-8975-1eeef6a5d44f tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2011.231236] env[62824]: DEBUG nova.compute.manager [-] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2011.231337] env[62824]: DEBUG nova.network.neutron [-] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2011.609591] env[62824]: DEBUG oslo_vmware.api [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5253a17e-32d4-acd5-3fe4-21b11d5f639d, 'name': SearchDatastore_Task, 'duration_secs': 0.024533} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2011.610293] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-077f078f-c7f7-4888-b6b4-351c358e5c4f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2011.616760] env[62824]: DEBUG oslo_vmware.api [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Waiting for the task: (returnval){ [ 2011.616760] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]523e7add-32d7-0540-bb11-369413708f39" [ 2011.616760] env[62824]: _type = "Task" [ 2011.616760] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2011.628741] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 6a724332-a165-4a2b-9dd8-85f27e7b7637 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 2011.630452] env[62824]: DEBUG oslo_vmware.api [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]523e7add-32d7-0540-bb11-369413708f39, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2011.631115] env[62824]: DEBUG nova.network.neutron [None req-26bb01c0-bc09-4b0e-b772-39e2968113f7 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 8bb90a4d-93c9-4f54-b15e-48fe966e6c97] Updating instance_info_cache with network_info: [{"id": "ec4e83ca-0312-4fc5-8c4f-475c5235bd4e", "address": "fa:16:3e:27:78:10", "network": {"id": "f099da67-9797-4509-a5fe-7c1d5d8560d5", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-13908051-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "31bcd298e89d49da977dda4cefd20432", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "023d6500-887e-4dc4-bec5-06b40450d9c0", "external-id": "nsx-vlan-transportzone-108", "segmentation_id": 108, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapec4e83ca-03", "ovs_interfaceid": "ec4e83ca-0312-4fc5-8c4f-475c5235bd4e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2011.654036] env[62824]: DEBUG oslo_vmware.api [None req-58bcc65c-ba1f-4462-9f05-d0e06434f495 tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Task: {'id': task-2146003, 'name': CreateSnapshot_Task, 'duration_secs': 0.909506} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2011.654812] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-58bcc65c-ba1f-4462-9f05-d0e06434f495 tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Created Snapshot of the VM instance {{(pid=62824) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 2011.655449] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66c5a96e-2d80-4c16-a75b-60e796f3beec {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2011.671685] env[62824]: DEBUG nova.compute.manager [req-3bf4e54e-c9ab-44ff-b068-66da93ea2339 req-8614048a-0afb-4310-bf23-6bf3cd4bb3a9 service nova] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Received event network-vif-deleted-02e90c28-7f36-4d83-84bb-fec194a04dab {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2011.671958] env[62824]: INFO nova.compute.manager [req-3bf4e54e-c9ab-44ff-b068-66da93ea2339 req-8614048a-0afb-4310-bf23-6bf3cd4bb3a9 service nova] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Neutron deleted interface 02e90c28-7f36-4d83-84bb-fec194a04dab; detaching it from the instance and deleting it from the info cache [ 2011.672148] env[62824]: DEBUG nova.network.neutron [req-3bf4e54e-c9ab-44ff-b068-66da93ea2339 req-8614048a-0afb-4310-bf23-6bf3cd4bb3a9 service nova] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2012.094033] env[62824]: DEBUG nova.network.neutron [-] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2012.130020] env[62824]: DEBUG oslo_vmware.api [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]523e7add-32d7-0540-bb11-369413708f39, 'name': SearchDatastore_Task, 'duration_secs': 0.0138} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2012.130020] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2012.130020] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] db018af0-7bdf-40c5-b129-fde15dfb8ab4/db018af0-7bdf-40c5-b129-fde15dfb8ab4.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2012.130020] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7de26962-7eca-4d29-9ce6-07f845f68213 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2012.133701] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 21c05725-3dad-4965-98d0-0622ebcc8ece has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 2012.135170] env[62824]: DEBUG oslo_concurrency.lockutils [None req-26bb01c0-bc09-4b0e-b772-39e2968113f7 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Releasing lock "refresh_cache-8bb90a4d-93c9-4f54-b15e-48fe966e6c97" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2012.137647] env[62824]: DEBUG oslo_vmware.api [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Waiting for the task: (returnval){ [ 2012.137647] env[62824]: value = "task-2146005" [ 2012.137647] env[62824]: _type = "Task" [ 2012.137647] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2012.148202] env[62824]: DEBUG oslo_vmware.api [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Task: {'id': task-2146005, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2012.174498] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-58bcc65c-ba1f-4462-9f05-d0e06434f495 tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Creating linked-clone VM from snapshot {{(pid=62824) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 2012.175225] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-aab41101-6362-470b-8d44-ccdd21054e9e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2012.178845] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ddec3d82-a5fd-46e0-903d-022e276aa1da {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2012.190634] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecb7472a-752b-4cd9-b744-e6da8e48d389 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2012.201761] env[62824]: DEBUG oslo_vmware.api [None req-58bcc65c-ba1f-4462-9f05-d0e06434f495 tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Waiting for the task: (returnval){ [ 2012.201761] env[62824]: value = "task-2146006" [ 2012.201761] env[62824]: _type = "Task" [ 2012.201761] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2012.211013] env[62824]: DEBUG oslo_vmware.api [None req-58bcc65c-ba1f-4462-9f05-d0e06434f495 tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Task: {'id': task-2146006, 'name': CloneVM_Task} progress is 10%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2012.230455] env[62824]: DEBUG nova.compute.manager [req-3bf4e54e-c9ab-44ff-b068-66da93ea2339 req-8614048a-0afb-4310-bf23-6bf3cd4bb3a9 service nova] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Detach interface failed, port_id=02e90c28-7f36-4d83-84bb-fec194a04dab, reason: Instance e158b5d4-c120-4e6c-89c4-7668e097926d could not be found. {{(pid=62824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11482}} [ 2012.597295] env[62824]: INFO nova.compute.manager [-] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Took 1.37 seconds to deallocate network for instance. [ 2012.638855] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance ed7d3e95-f26e-40fd-a370-e17922bbff8e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 2012.656542] env[62824]: DEBUG oslo_vmware.api [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Task: {'id': task-2146005, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2012.712302] env[62824]: DEBUG oslo_vmware.api [None req-58bcc65c-ba1f-4462-9f05-d0e06434f495 tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Task: {'id': task-2146006, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2013.106011] env[62824]: DEBUG oslo_concurrency.lockutils [None req-6008bdc7-ed30-4e87-8975-1eeef6a5d44f tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2013.142122] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance b0456b8a-348a-4503-a92c-58e5ab455d1c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 2013.143762] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-26bb01c0-bc09-4b0e-b772-39e2968113f7 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 8bb90a4d-93c9-4f54-b15e-48fe966e6c97] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2013.144102] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dfe2db42-5d08-4e43-8e05-72b22cfb78ad {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2013.156258] env[62824]: DEBUG oslo_vmware.api [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Task: {'id': task-2146005, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.640367} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2013.158322] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] db018af0-7bdf-40c5-b129-fde15dfb8ab4/db018af0-7bdf-40c5-b129-fde15dfb8ab4.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2013.158496] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] [instance: db018af0-7bdf-40c5-b129-fde15dfb8ab4] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2013.159052] env[62824]: DEBUG oslo_vmware.api [None req-26bb01c0-bc09-4b0e-b772-39e2968113f7 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Waiting for the task: (returnval){ [ 2013.159052] env[62824]: value = "task-2146007" [ 2013.159052] env[62824]: _type = "Task" [ 2013.159052] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2013.159245] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0b568891-2fc7-4f7a-b906-3dff90f34314 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2013.170551] env[62824]: DEBUG oslo_vmware.api [None req-26bb01c0-bc09-4b0e-b772-39e2968113f7 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': task-2146007, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2013.171884] env[62824]: DEBUG oslo_vmware.api [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Waiting for the task: (returnval){ [ 2013.171884] env[62824]: value = "task-2146008" [ 2013.171884] env[62824]: _type = "Task" [ 2013.171884] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2013.182457] env[62824]: DEBUG oslo_vmware.api [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Task: {'id': task-2146008, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2013.213933] env[62824]: DEBUG oslo_vmware.api [None req-58bcc65c-ba1f-4462-9f05-d0e06434f495 tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Task: {'id': task-2146006, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2013.651802] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance bbfcb2e3-9326-4548-b15b-e054cbfd192e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 2013.672475] env[62824]: DEBUG oslo_vmware.api [None req-26bb01c0-bc09-4b0e-b772-39e2968113f7 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': task-2146007, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2013.680800] env[62824]: DEBUG oslo_vmware.api [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Task: {'id': task-2146008, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075614} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2013.681048] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] [instance: db018af0-7bdf-40c5-b129-fde15dfb8ab4] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2013.681824] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0aabbb4d-ab65-4aa5-9ad4-83fe014af403 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2013.704723] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] [instance: db018af0-7bdf-40c5-b129-fde15dfb8ab4] Reconfiguring VM instance instance-0000002e to attach disk [datastore1] db018af0-7bdf-40c5-b129-fde15dfb8ab4/db018af0-7bdf-40c5-b129-fde15dfb8ab4.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2013.705557] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2e67209a-4ec4-40fb-af9d-b39655589a3f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2013.728314] env[62824]: DEBUG oslo_vmware.api [None req-58bcc65c-ba1f-4462-9f05-d0e06434f495 tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Task: {'id': task-2146006, 'name': CloneVM_Task} progress is 95%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2013.730937] env[62824]: DEBUG oslo_vmware.api [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Waiting for the task: (returnval){ [ 2013.730937] env[62824]: value = "task-2146009" [ 2013.730937] env[62824]: _type = "Task" [ 2013.730937] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2013.740776] env[62824]: DEBUG oslo_vmware.api [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Task: {'id': task-2146009, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2014.156901] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 2f315f9e-94e4-47ef-9503-ee92b59e5452 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 2014.173921] env[62824]: DEBUG oslo_vmware.api [None req-26bb01c0-bc09-4b0e-b772-39e2968113f7 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': task-2146007, 'name': PowerOnVM_Task, 'duration_secs': 0.57773} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2014.174261] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-26bb01c0-bc09-4b0e-b772-39e2968113f7 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 8bb90a4d-93c9-4f54-b15e-48fe966e6c97] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2014.174455] env[62824]: DEBUG nova.compute.manager [None req-26bb01c0-bc09-4b0e-b772-39e2968113f7 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 8bb90a4d-93c9-4f54-b15e-48fe966e6c97] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2014.175271] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32d326df-aa4a-4716-b2d6-20cc92c199f0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2014.215871] env[62824]: DEBUG oslo_vmware.api [None req-58bcc65c-ba1f-4462-9f05-d0e06434f495 tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Task: {'id': task-2146006, 'name': CloneVM_Task, 'duration_secs': 1.899274} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2014.216966] env[62824]: INFO nova.virt.vmwareapi.vmops [None req-58bcc65c-ba1f-4462-9f05-d0e06434f495 tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Created linked-clone VM from snapshot [ 2014.217779] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28c76dc7-331a-4763-8c97-1c46f0f9e2b1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2014.226462] env[62824]: DEBUG nova.virt.vmwareapi.images [None req-58bcc65c-ba1f-4462-9f05-d0e06434f495 tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Uploading image c076cb8e-f451-482d-bd51-2014c6c0bb68 {{(pid=62824) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 2014.239105] env[62824]: DEBUG oslo_vmware.api [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Task: {'id': task-2146009, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2014.253666] env[62824]: DEBUG oslo_vmware.rw_handles [None req-58bcc65c-ba1f-4462-9f05-d0e06434f495 tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 2014.253666] env[62824]: value = "vm-438613" [ 2014.253666] env[62824]: _type = "VirtualMachine" [ 2014.253666] env[62824]: }. {{(pid=62824) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 2014.253954] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-02762331-e4bd-4e00-960e-01b3778feca3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2014.262605] env[62824]: DEBUG oslo_vmware.rw_handles [None req-58bcc65c-ba1f-4462-9f05-d0e06434f495 tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Lease: (returnval){ [ 2014.262605] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52f589f4-95c3-7847-420a-e634580e09a4" [ 2014.262605] env[62824]: _type = "HttpNfcLease" [ 2014.262605] env[62824]: } obtained for exporting VM: (result){ [ 2014.262605] env[62824]: value = "vm-438613" [ 2014.262605] env[62824]: _type = "VirtualMachine" [ 2014.262605] env[62824]: }. {{(pid=62824) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 2014.262957] env[62824]: DEBUG oslo_vmware.api [None req-58bcc65c-ba1f-4462-9f05-d0e06434f495 tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Waiting for the lease: (returnval){ [ 2014.262957] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52f589f4-95c3-7847-420a-e634580e09a4" [ 2014.262957] env[62824]: _type = "HttpNfcLease" [ 2014.262957] env[62824]: } to be ready. {{(pid=62824) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 2014.269277] env[62824]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 2014.269277] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52f589f4-95c3-7847-420a-e634580e09a4" [ 2014.269277] env[62824]: _type = "HttpNfcLease" [ 2014.269277] env[62824]: } is initializing. {{(pid=62824) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 2014.606989] env[62824]: DEBUG oslo_vmware.rw_handles [None req-21197ef3-67a1-4f5a-a6f1-7b606e12edaf tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5243b7f5-5cb5-611a-2d77-60a4ce26c479/disk-0.vmdk. {{(pid=62824) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 2014.607899] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-317fbcf4-09c6-4af9-a308-42d34c20fcb8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2014.614216] env[62824]: DEBUG oslo_vmware.rw_handles [None req-21197ef3-67a1-4f5a-a6f1-7b606e12edaf tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5243b7f5-5cb5-611a-2d77-60a4ce26c479/disk-0.vmdk is in state: ready. {{(pid=62824) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 2014.614358] env[62824]: ERROR oslo_vmware.rw_handles [None req-21197ef3-67a1-4f5a-a6f1-7b606e12edaf tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5243b7f5-5cb5-611a-2d77-60a4ce26c479/disk-0.vmdk due to incomplete transfer. [ 2014.614573] env[62824]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-1a792fe9-e34b-4fd0-a83f-1ac2c91b6a2e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2014.622013] env[62824]: DEBUG oslo_vmware.rw_handles [None req-21197ef3-67a1-4f5a-a6f1-7b606e12edaf tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5243b7f5-5cb5-611a-2d77-60a4ce26c479/disk-0.vmdk. {{(pid=62824) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 2014.622217] env[62824]: DEBUG nova.virt.vmwareapi.images [None req-21197ef3-67a1-4f5a-a6f1-7b606e12edaf tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Uploaded image 9b84651b-bc0f-4778-9d83-183e9ac78e55 to the Glance image server {{(pid=62824) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 2014.624206] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-21197ef3-67a1-4f5a-a6f1-7b606e12edaf tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Destroying the VM {{(pid=62824) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 2014.624461] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-49d5ed99-2891-4571-90a7-f04f681ceb62 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2014.629944] env[62824]: DEBUG oslo_vmware.api [None req-21197ef3-67a1-4f5a-a6f1-7b606e12edaf tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Waiting for the task: (returnval){ [ 2014.629944] env[62824]: value = "task-2146011" [ 2014.629944] env[62824]: _type = "Task" [ 2014.629944] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2014.637337] env[62824]: DEBUG oslo_vmware.api [None req-21197ef3-67a1-4f5a-a6f1-7b606e12edaf tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2146011, 'name': Destroy_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2014.659076] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 919e8a45-7810-4a8d-a1aa-5046b5ab059c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 2014.741305] env[62824]: DEBUG oslo_vmware.api [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Task: {'id': task-2146009, 'name': ReconfigVM_Task, 'duration_secs': 0.537396} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2014.741656] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] [instance: db018af0-7bdf-40c5-b129-fde15dfb8ab4] Reconfigured VM instance instance-0000002e to attach disk [datastore1] db018af0-7bdf-40c5-b129-fde15dfb8ab4/db018af0-7bdf-40c5-b129-fde15dfb8ab4.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2014.742568] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-62f4e580-5a2d-4900-a37b-7b49dd091656 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2014.749235] env[62824]: DEBUG oslo_vmware.api [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Waiting for the task: (returnval){ [ 2014.749235] env[62824]: value = "task-2146012" [ 2014.749235] env[62824]: _type = "Task" [ 2014.749235] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2014.757230] env[62824]: DEBUG oslo_vmware.api [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Task: {'id': task-2146012, 'name': Rename_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2014.769593] env[62824]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 2014.769593] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52f589f4-95c3-7847-420a-e634580e09a4" [ 2014.769593] env[62824]: _type = "HttpNfcLease" [ 2014.769593] env[62824]: } is ready. {{(pid=62824) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 2014.769892] env[62824]: DEBUG oslo_vmware.rw_handles [None req-58bcc65c-ba1f-4462-9f05-d0e06434f495 tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 2014.769892] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52f589f4-95c3-7847-420a-e634580e09a4" [ 2014.769892] env[62824]: _type = "HttpNfcLease" [ 2014.769892] env[62824]: }. {{(pid=62824) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 2014.770571] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a494806-e090-4d70-8c87-ed9ecf48dcb8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2014.778026] env[62824]: DEBUG oslo_vmware.rw_handles [None req-58bcc65c-ba1f-4462-9f05-d0e06434f495 tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e52e6b-1132-1e69-310a-5702becf9ef5/disk-0.vmdk from lease info. {{(pid=62824) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 2014.778189] env[62824]: DEBUG oslo_vmware.rw_handles [None req-58bcc65c-ba1f-4462-9f05-d0e06434f495 tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e52e6b-1132-1e69-310a-5702becf9ef5/disk-0.vmdk for reading. {{(pid=62824) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 2014.898224] env[62824]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-175acd38-ee71-40cc-bd47-f8afb3d46754 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2015.140269] env[62824]: DEBUG oslo_vmware.api [None req-21197ef3-67a1-4f5a-a6f1-7b606e12edaf tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2146011, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2015.166329] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 15479062-af75-4925-99b3-77d6a49751ad has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 2, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 2015.166663] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Total usable vcpus: 48, total allocated vcpus: 22 {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 2015.166813] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=4800MB phys_disk=200GB used_disk=21GB total_vcpus=48 used_vcpus=22 pci_stats=[] {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 2015.264424] env[62824]: DEBUG oslo_vmware.api [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Task: {'id': task-2146012, 'name': Rename_Task, 'duration_secs': 0.362573} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2015.265115] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] [instance: db018af0-7bdf-40c5-b129-fde15dfb8ab4] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2015.265622] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-10cb64e4-39a1-42f2-85cf-aafadb643135 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2015.274031] env[62824]: DEBUG oslo_vmware.api [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Waiting for the task: (returnval){ [ 2015.274031] env[62824]: value = "task-2146013" [ 2015.274031] env[62824]: _type = "Task" [ 2015.274031] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2015.282766] env[62824]: DEBUG oslo_vmware.api [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Task: {'id': task-2146013, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2015.644119] env[62824]: DEBUG oslo_vmware.api [None req-21197ef3-67a1-4f5a-a6f1-7b606e12edaf tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2146011, 'name': Destroy_Task, 'duration_secs': 0.585998} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2015.644119] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-21197ef3-67a1-4f5a-a6f1-7b606e12edaf tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Destroyed the VM [ 2015.644119] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-21197ef3-67a1-4f5a-a6f1-7b606e12edaf tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Deleting Snapshot of the VM instance {{(pid=62824) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 2015.646323] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-c77af4c9-0fde-4a51-b3d4-b25e716b75e2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2015.660583] env[62824]: DEBUG oslo_vmware.api [None req-21197ef3-67a1-4f5a-a6f1-7b606e12edaf tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Waiting for the task: (returnval){ [ 2015.660583] env[62824]: value = "task-2146014" [ 2015.660583] env[62824]: _type = "Task" [ 2015.660583] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2015.665489] env[62824]: DEBUG oslo_vmware.api [None req-21197ef3-67a1-4f5a-a6f1-7b606e12edaf tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2146014, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2015.787208] env[62824]: DEBUG oslo_vmware.api [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Task: {'id': task-2146013, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2015.807069] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61172371-b47f-48d2-b927-88e54b7a086e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2015.815098] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cfffcfd-3ea2-4bfb-a995-847f9134933e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2015.855186] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9ad36cc-ab65-492c-9be2-5935c93b9b61 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2015.864251] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5de33668-899e-49be-b764-8e8fe416c234 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2015.879623] env[62824]: DEBUG nova.compute.provider_tree [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2016.168750] env[62824]: DEBUG oslo_vmware.api [None req-21197ef3-67a1-4f5a-a6f1-7b606e12edaf tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2146014, 'name': RemoveSnapshot_Task} progress is 36%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2016.290796] env[62824]: DEBUG oslo_vmware.api [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Task: {'id': task-2146013, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2016.383886] env[62824]: DEBUG nova.scheduler.client.report [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2016.669378] env[62824]: DEBUG oslo_vmware.api [None req-21197ef3-67a1-4f5a-a6f1-7b606e12edaf tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2146014, 'name': RemoveSnapshot_Task, 'duration_secs': 0.771326} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2016.669779] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-21197ef3-67a1-4f5a-a6f1-7b606e12edaf tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Deleted Snapshot of the VM instance {{(pid=62824) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 2016.670038] env[62824]: INFO nova.compute.manager [None req-21197ef3-67a1-4f5a-a6f1-7b606e12edaf tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Took 18.43 seconds to snapshot the instance on the hypervisor. [ 2016.792529] env[62824]: DEBUG oslo_vmware.api [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Task: {'id': task-2146013, 'name': PowerOnVM_Task, 'duration_secs': 1.167923} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2016.792801] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] [instance: db018af0-7bdf-40c5-b129-fde15dfb8ab4] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2016.793054] env[62824]: INFO nova.compute.manager [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] [instance: db018af0-7bdf-40c5-b129-fde15dfb8ab4] Took 9.98 seconds to spawn the instance on the hypervisor. [ 2016.793229] env[62824]: DEBUG nova.compute.manager [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] [instance: db018af0-7bdf-40c5-b129-fde15dfb8ab4] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2016.794040] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4381d073-84b3-44b0-9c52-48048289cdfe {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2016.890495] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62824) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 2016.890671] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 9.855s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2016.891064] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.249s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2016.892746] env[62824]: INFO nova.compute.claims [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2016.896562] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2016.896718] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Cleaning up deleted instances with incomplete migration {{(pid=62824) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11783}} [ 2017.314045] env[62824]: INFO nova.compute.manager [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] [instance: db018af0-7bdf-40c5-b129-fde15dfb8ab4] Took 43.41 seconds to build instance. [ 2017.642720] env[62824]: DEBUG oslo_concurrency.lockutils [None req-6c2d5169-8d71-4209-a7d7-4b4ea2abbc8d tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Acquiring lock "db018af0-7bdf-40c5-b129-fde15dfb8ab4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2017.816248] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3a442e6f-1b51-4248-96f9-9fcc9f04041c tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Lock "db018af0-7bdf-40c5-b129-fde15dfb8ab4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 116.743s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2017.817572] env[62824]: DEBUG oslo_concurrency.lockutils [None req-6c2d5169-8d71-4209-a7d7-4b4ea2abbc8d tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Lock "db018af0-7bdf-40c5-b129-fde15dfb8ab4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.175s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2017.817821] env[62824]: DEBUG oslo_concurrency.lockutils [None req-6c2d5169-8d71-4209-a7d7-4b4ea2abbc8d tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Acquiring lock "db018af0-7bdf-40c5-b129-fde15dfb8ab4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2017.818051] env[62824]: DEBUG oslo_concurrency.lockutils [None req-6c2d5169-8d71-4209-a7d7-4b4ea2abbc8d tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Lock "db018af0-7bdf-40c5-b129-fde15dfb8ab4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2017.818246] env[62824]: DEBUG oslo_concurrency.lockutils [None req-6c2d5169-8d71-4209-a7d7-4b4ea2abbc8d tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Lock "db018af0-7bdf-40c5-b129-fde15dfb8ab4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2017.820289] env[62824]: INFO nova.compute.manager [None req-6c2d5169-8d71-4209-a7d7-4b4ea2abbc8d tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] [instance: db018af0-7bdf-40c5-b129-fde15dfb8ab4] Terminating instance [ 2018.262902] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a1ef20b1-87c8-4499-bf60-63e3997b1e64 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Acquiring lock "41f068d3-2c8d-46f0-8d84-78531d28b0dd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2018.263278] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a1ef20b1-87c8-4499-bf60-63e3997b1e64 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Lock "41f068d3-2c8d-46f0-8d84-78531d28b0dd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2018.263400] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a1ef20b1-87c8-4499-bf60-63e3997b1e64 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Acquiring lock "41f068d3-2c8d-46f0-8d84-78531d28b0dd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2018.263568] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a1ef20b1-87c8-4499-bf60-63e3997b1e64 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Lock "41f068d3-2c8d-46f0-8d84-78531d28b0dd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2018.263734] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a1ef20b1-87c8-4499-bf60-63e3997b1e64 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Lock "41f068d3-2c8d-46f0-8d84-78531d28b0dd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2018.268109] env[62824]: INFO nova.compute.manager [None req-a1ef20b1-87c8-4499-bf60-63e3997b1e64 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 41f068d3-2c8d-46f0-8d84-78531d28b0dd] Terminating instance [ 2018.283521] env[62824]: DEBUG nova.compute.manager [None req-b4894b47-a9c8-4b49-af68-93eace3f9a8c tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2018.284614] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dde1be54-953c-4cfe-b959-5b524343111e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2018.321274] env[62824]: DEBUG nova.compute.manager [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: b0456b8a-348a-4503-a92c-58e5ab455d1c] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2018.325027] env[62824]: DEBUG nova.compute.manager [None req-6c2d5169-8d71-4209-a7d7-4b4ea2abbc8d tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] [instance: db018af0-7bdf-40c5-b129-fde15dfb8ab4] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2018.325283] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-6c2d5169-8d71-4209-a7d7-4b4ea2abbc8d tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] [instance: db018af0-7bdf-40c5-b129-fde15dfb8ab4] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2018.326367] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fdc1fea-45a9-4d58-88d2-2e7c89bfea58 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2018.335040] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c2d5169-8d71-4209-a7d7-4b4ea2abbc8d tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] [instance: db018af0-7bdf-40c5-b129-fde15dfb8ab4] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2018.338023] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-99efa8e2-b3c5-4fce-a6c2-ed626d082eac {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2018.345531] env[62824]: DEBUG oslo_vmware.api [None req-6c2d5169-8d71-4209-a7d7-4b4ea2abbc8d tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Waiting for the task: (returnval){ [ 2018.345531] env[62824]: value = "task-2146015" [ 2018.345531] env[62824]: _type = "Task" [ 2018.345531] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2018.359157] env[62824]: DEBUG oslo_vmware.api [None req-6c2d5169-8d71-4209-a7d7-4b4ea2abbc8d tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Task: {'id': task-2146015, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2018.396676] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2018.397706] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2018.409250] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d000d2a5-0dc4-46de-84a4-2e559d750b6f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2018.418522] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3674e074-f95b-4ef6-ab3d-4bebb646908f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2018.454714] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdadd995-8230-4f43-b724-563444281130 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2018.465140] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bf55228-1325-4585-a52e-9aa624b56450 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2018.481353] env[62824]: DEBUG nova.compute.provider_tree [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2018.771459] env[62824]: DEBUG nova.compute.manager [None req-a1ef20b1-87c8-4499-bf60-63e3997b1e64 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 41f068d3-2c8d-46f0-8d84-78531d28b0dd] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2018.771733] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-a1ef20b1-87c8-4499-bf60-63e3997b1e64 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 41f068d3-2c8d-46f0-8d84-78531d28b0dd] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2018.772738] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5520b58f-3cf9-4caa-8235-397f8217600a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2018.782068] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1ef20b1-87c8-4499-bf60-63e3997b1e64 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 41f068d3-2c8d-46f0-8d84-78531d28b0dd] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2018.782360] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6ebec282-4daa-4d1f-8cf7-a0f2a5357144 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2018.789447] env[62824]: DEBUG oslo_vmware.api [None req-a1ef20b1-87c8-4499-bf60-63e3997b1e64 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Waiting for the task: (returnval){ [ 2018.789447] env[62824]: value = "task-2146016" [ 2018.789447] env[62824]: _type = "Task" [ 2018.789447] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2018.798636] env[62824]: INFO nova.compute.manager [None req-b4894b47-a9c8-4b49-af68-93eace3f9a8c tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2] instance snapshotting [ 2018.800378] env[62824]: DEBUG oslo_vmware.api [None req-a1ef20b1-87c8-4499-bf60-63e3997b1e64 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': task-2146016, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2018.802930] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c95d267a-ffd9-4b90-967f-e309afdd66ec {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2018.823601] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dea66cea-ea34-4868-9609-b1f91fcff127 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2018.844523] env[62824]: DEBUG oslo_concurrency.lockutils [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2018.855568] env[62824]: DEBUG oslo_vmware.api [None req-6c2d5169-8d71-4209-a7d7-4b4ea2abbc8d tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Task: {'id': task-2146015, 'name': PowerOffVM_Task, 'duration_secs': 0.316787} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2018.855783] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c2d5169-8d71-4209-a7d7-4b4ea2abbc8d tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] [instance: db018af0-7bdf-40c5-b129-fde15dfb8ab4] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2018.855954] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-6c2d5169-8d71-4209-a7d7-4b4ea2abbc8d tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] [instance: db018af0-7bdf-40c5-b129-fde15dfb8ab4] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2018.856256] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ed98b1c6-6282-4a04-afaa-aee55c42cda8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2018.908210] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2018.908210] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Starting heal instance info cache {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10445}} [ 2018.964632] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-6c2d5169-8d71-4209-a7d7-4b4ea2abbc8d tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] [instance: db018af0-7bdf-40c5-b129-fde15dfb8ab4] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2018.964861] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-6c2d5169-8d71-4209-a7d7-4b4ea2abbc8d tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] [instance: db018af0-7bdf-40c5-b129-fde15dfb8ab4] Deleting contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2018.965139] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c2d5169-8d71-4209-a7d7-4b4ea2abbc8d tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Deleting the datastore file [datastore1] db018af0-7bdf-40c5-b129-fde15dfb8ab4 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2018.965444] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-13d42023-f9ad-47a6-aced-62f1a0543d7d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2018.973868] env[62824]: DEBUG oslo_vmware.api [None req-6c2d5169-8d71-4209-a7d7-4b4ea2abbc8d tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Waiting for the task: (returnval){ [ 2018.973868] env[62824]: value = "task-2146018" [ 2018.973868] env[62824]: _type = "Task" [ 2018.973868] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2018.982559] env[62824]: DEBUG oslo_vmware.api [None req-6c2d5169-8d71-4209-a7d7-4b4ea2abbc8d tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Task: {'id': task-2146018, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2018.984520] env[62824]: DEBUG nova.scheduler.client.report [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2019.299741] env[62824]: DEBUG oslo_vmware.api [None req-a1ef20b1-87c8-4499-bf60-63e3997b1e64 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': task-2146016, 'name': PowerOffVM_Task, 'duration_secs': 0.346733} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2019.300020] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1ef20b1-87c8-4499-bf60-63e3997b1e64 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 41f068d3-2c8d-46f0-8d84-78531d28b0dd] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2019.300217] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-a1ef20b1-87c8-4499-bf60-63e3997b1e64 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 41f068d3-2c8d-46f0-8d84-78531d28b0dd] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2019.300476] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9f740892-6a96-4c5d-bf9a-298c4988f57c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2019.340474] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-b4894b47-a9c8-4b49-af68-93eace3f9a8c tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2] Creating Snapshot of the VM instance {{(pid=62824) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 2019.340772] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-1b517cff-4659-40ff-8f91-eb0c57be258e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2019.348049] env[62824]: DEBUG oslo_vmware.api [None req-b4894b47-a9c8-4b49-af68-93eace3f9a8c tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Waiting for the task: (returnval){ [ 2019.348049] env[62824]: value = "task-2146020" [ 2019.348049] env[62824]: _type = "Task" [ 2019.348049] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2019.356364] env[62824]: DEBUG oslo_vmware.api [None req-b4894b47-a9c8-4b49-af68-93eace3f9a8c tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2146020, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2019.375618] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-a1ef20b1-87c8-4499-bf60-63e3997b1e64 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 41f068d3-2c8d-46f0-8d84-78531d28b0dd] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2019.375618] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-a1ef20b1-87c8-4499-bf60-63e3997b1e64 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 41f068d3-2c8d-46f0-8d84-78531d28b0dd] Deleting contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2019.375618] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-a1ef20b1-87c8-4499-bf60-63e3997b1e64 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Deleting the datastore file [datastore2] 41f068d3-2c8d-46f0-8d84-78531d28b0dd {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2019.376039] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0aa88eb9-9b5b-4bd0-a486-18f403933048 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2019.383644] env[62824]: DEBUG oslo_vmware.api [None req-a1ef20b1-87c8-4499-bf60-63e3997b1e64 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Waiting for the task: (returnval){ [ 2019.383644] env[62824]: value = "task-2146021" [ 2019.383644] env[62824]: _type = "Task" [ 2019.383644] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2019.394315] env[62824]: DEBUG oslo_vmware.api [None req-a1ef20b1-87c8-4499-bf60-63e3997b1e64 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': task-2146021, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2019.440130] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "refresh_cache-82382932-7302-4441-a6f8-9aa2300ec0f6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2019.440269] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquired lock "refresh_cache-82382932-7302-4441-a6f8-9aa2300ec0f6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2019.440419] env[62824]: DEBUG nova.network.neutron [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 82382932-7302-4441-a6f8-9aa2300ec0f6] Forcefully refreshing network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 2019.485328] env[62824]: DEBUG oslo_vmware.api [None req-6c2d5169-8d71-4209-a7d7-4b4ea2abbc8d tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Task: {'id': task-2146018, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.418293} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2019.485969] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c2d5169-8d71-4209-a7d7-4b4ea2abbc8d tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2019.485969] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-6c2d5169-8d71-4209-a7d7-4b4ea2abbc8d tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] [instance: db018af0-7bdf-40c5-b129-fde15dfb8ab4] Deleted contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2019.485969] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-6c2d5169-8d71-4209-a7d7-4b4ea2abbc8d tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] [instance: db018af0-7bdf-40c5-b129-fde15dfb8ab4] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2019.486189] env[62824]: INFO nova.compute.manager [None req-6c2d5169-8d71-4209-a7d7-4b4ea2abbc8d tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] [instance: db018af0-7bdf-40c5-b129-fde15dfb8ab4] Took 1.16 seconds to destroy the instance on the hypervisor. [ 2019.486839] env[62824]: DEBUG oslo.service.loopingcall [None req-6c2d5169-8d71-4209-a7d7-4b4ea2abbc8d tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2019.486839] env[62824]: DEBUG nova.compute.manager [-] [instance: db018af0-7bdf-40c5-b129-fde15dfb8ab4] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2019.486839] env[62824]: DEBUG nova.network.neutron [-] [instance: db018af0-7bdf-40c5-b129-fde15dfb8ab4] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2019.488800] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.598s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2019.489298] env[62824]: DEBUG nova.compute.manager [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2019.493028] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 32.776s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2019.494221] env[62824]: INFO nova.compute.claims [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02405b17-7022-4e2c-a357-415de88f63d0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2019.778693] env[62824]: DEBUG nova.compute.manager [req-d510709f-1fa4-4751-9737-b3b45dfdbe8b req-0de84b58-1a7d-4733-a262-2ff30f5eb372 service nova] [instance: db018af0-7bdf-40c5-b129-fde15dfb8ab4] Received event network-vif-deleted-d94a1564-ae07-4d73-bae0-2cf718dc318c {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2019.778859] env[62824]: INFO nova.compute.manager [req-d510709f-1fa4-4751-9737-b3b45dfdbe8b req-0de84b58-1a7d-4733-a262-2ff30f5eb372 service nova] [instance: db018af0-7bdf-40c5-b129-fde15dfb8ab4] Neutron deleted interface d94a1564-ae07-4d73-bae0-2cf718dc318c; detaching it from the instance and deleting it from the info cache [ 2019.779074] env[62824]: DEBUG nova.network.neutron [req-d510709f-1fa4-4751-9737-b3b45dfdbe8b req-0de84b58-1a7d-4733-a262-2ff30f5eb372 service nova] [instance: db018af0-7bdf-40c5-b129-fde15dfb8ab4] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2019.859620] env[62824]: DEBUG oslo_vmware.api [None req-b4894b47-a9c8-4b49-af68-93eace3f9a8c tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2146020, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2019.894211] env[62824]: DEBUG oslo_vmware.api [None req-a1ef20b1-87c8-4499-bf60-63e3997b1e64 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': task-2146021, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.249552} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2019.894304] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-a1ef20b1-87c8-4499-bf60-63e3997b1e64 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2019.894428] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-a1ef20b1-87c8-4499-bf60-63e3997b1e64 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 41f068d3-2c8d-46f0-8d84-78531d28b0dd] Deleted contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2019.894704] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-a1ef20b1-87c8-4499-bf60-63e3997b1e64 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 41f068d3-2c8d-46f0-8d84-78531d28b0dd] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2019.894786] env[62824]: INFO nova.compute.manager [None req-a1ef20b1-87c8-4499-bf60-63e3997b1e64 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 41f068d3-2c8d-46f0-8d84-78531d28b0dd] Took 1.12 seconds to destroy the instance on the hypervisor. [ 2019.895013] env[62824]: DEBUG oslo.service.loopingcall [None req-a1ef20b1-87c8-4499-bf60-63e3997b1e64 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2019.895241] env[62824]: DEBUG nova.compute.manager [-] [instance: 41f068d3-2c8d-46f0-8d84-78531d28b0dd] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2019.895428] env[62824]: DEBUG nova.network.neutron [-] [instance: 41f068d3-2c8d-46f0-8d84-78531d28b0dd] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2019.999456] env[62824]: DEBUG nova.compute.utils [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2020.004392] env[62824]: DEBUG nova.compute.manager [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2020.004688] env[62824]: DEBUG nova.network.neutron [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2020.060094] env[62824]: DEBUG nova.policy [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6287facc8293419c943c0234ed798dac', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd180bbdc8e72414081363fe696ead6fb', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2020.257528] env[62824]: DEBUG nova.network.neutron [-] [instance: db018af0-7bdf-40c5-b129-fde15dfb8ab4] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2020.283769] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9b7d534b-54af-4d72-a8bc-b5ab2426a8d3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2020.291999] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edb8dc25-53da-4e86-b2c7-ab92af3a62f1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2020.331972] env[62824]: DEBUG nova.compute.manager [req-d510709f-1fa4-4751-9737-b3b45dfdbe8b req-0de84b58-1a7d-4733-a262-2ff30f5eb372 service nova] [instance: db018af0-7bdf-40c5-b129-fde15dfb8ab4] Detach interface failed, port_id=d94a1564-ae07-4d73-bae0-2cf718dc318c, reason: Instance db018af0-7bdf-40c5-b129-fde15dfb8ab4 could not be found. {{(pid=62824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11482}} [ 2020.358474] env[62824]: DEBUG oslo_vmware.api [None req-b4894b47-a9c8-4b49-af68-93eace3f9a8c tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2146020, 'name': CreateSnapshot_Task, 'duration_secs': 0.598392} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2020.358758] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-b4894b47-a9c8-4b49-af68-93eace3f9a8c tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2] Created Snapshot of the VM instance {{(pid=62824) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 2020.359514] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed83022f-bb5b-4ae8-91f9-d604fdd6067e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2020.427346] env[62824]: DEBUG nova.network.neutron [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] Successfully created port: 2d7ab792-8a41-4c8b-91a4-e4a677466a12 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2020.506201] env[62824]: DEBUG nova.compute.manager [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2020.679470] env[62824]: DEBUG nova.network.neutron [-] [instance: 41f068d3-2c8d-46f0-8d84-78531d28b0dd] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2020.711351] env[62824]: DEBUG nova.network.neutron [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 82382932-7302-4441-a6f8-9aa2300ec0f6] Updating instance_info_cache with network_info: [{"id": "760062c3-a8b6-4863-a557-0d44a27e3bb3", "address": "fa:16:3e:d4:15:ee", "network": {"id": "668ff2d5-a4e1-488b-b71f-7f45c130d0e0", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-616618532-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.169", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0cf3a345a7d54625885c2159edcf0877", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "57691231-2b8d-4d71-8f79-d4a6a1d95ec8", "external-id": "nsx-vlan-transportzone-373", "segmentation_id": 373, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap760062c3-a8", "ovs_interfaceid": "760062c3-a8b6-4863-a557-0d44a27e3bb3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2020.759882] env[62824]: INFO nova.compute.manager [-] [instance: db018af0-7bdf-40c5-b129-fde15dfb8ab4] Took 1.27 seconds to deallocate network for instance. [ 2020.878937] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-b4894b47-a9c8-4b49-af68-93eace3f9a8c tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2] Creating linked-clone VM from snapshot {{(pid=62824) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 2020.882197] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-a5566236-9399-41a1-8ef1-abac9d4b0caf {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2020.892233] env[62824]: DEBUG oslo_vmware.api [None req-b4894b47-a9c8-4b49-af68-93eace3f9a8c tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Waiting for the task: (returnval){ [ 2020.892233] env[62824]: value = "task-2146022" [ 2020.892233] env[62824]: _type = "Task" [ 2020.892233] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2020.903079] env[62824]: DEBUG oslo_vmware.api [None req-b4894b47-a9c8-4b49-af68-93eace3f9a8c tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2146022, 'name': CloneVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2021.079358] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49e91002-b833-4e9a-8e90-864413053d81 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2021.088169] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fef8fcf-47e0-4a5f-89d1-7b0f99d6a7fc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2021.123133] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01f39fd2-59ba-4c82-b9b6-e216b6ca1e9c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2021.131953] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ca277f5-2e0d-4e41-a6f8-c147fc8cdb32 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2021.149426] env[62824]: DEBUG nova.compute.provider_tree [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2021.184635] env[62824]: INFO nova.compute.manager [-] [instance: 41f068d3-2c8d-46f0-8d84-78531d28b0dd] Took 1.29 seconds to deallocate network for instance. [ 2021.214236] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Releasing lock "refresh_cache-82382932-7302-4441-a6f8-9aa2300ec0f6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2021.214512] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 82382932-7302-4441-a6f8-9aa2300ec0f6] Updated the network info_cache for instance {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10516}} [ 2021.214774] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2021.215120] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2021.215352] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2021.215636] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2021.215851] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2021.216070] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2021.216204] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62824) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11064}} [ 2021.216392] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager.update_available_resource {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2021.273784] env[62824]: DEBUG oslo_concurrency.lockutils [None req-6c2d5169-8d71-4209-a7d7-4b4ea2abbc8d tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2021.403672] env[62824]: DEBUG oslo_vmware.api [None req-b4894b47-a9c8-4b49-af68-93eace3f9a8c tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2146022, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2021.521046] env[62824]: DEBUG nova.compute.manager [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2021.544472] env[62824]: DEBUG nova.virt.hardware [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2021.544723] env[62824]: DEBUG nova.virt.hardware [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2021.544876] env[62824]: DEBUG nova.virt.hardware [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2021.545125] env[62824]: DEBUG nova.virt.hardware [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2021.545308] env[62824]: DEBUG nova.virt.hardware [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2021.545460] env[62824]: DEBUG nova.virt.hardware [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2021.545733] env[62824]: DEBUG nova.virt.hardware [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2021.545910] env[62824]: DEBUG nova.virt.hardware [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2021.546166] env[62824]: DEBUG nova.virt.hardware [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2021.546348] env[62824]: DEBUG nova.virt.hardware [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2021.546523] env[62824]: DEBUG nova.virt.hardware [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2021.547421] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0d48bef-90b7-4822-b2c8-89cb496eb012 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2021.555673] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d47380f-2d5d-44e0-90ef-5ef64f2ae7da {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2021.655930] env[62824]: DEBUG nova.scheduler.client.report [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2021.691689] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a1ef20b1-87c8-4499-bf60-63e3997b1e64 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2021.719519] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2021.810320] env[62824]: DEBUG nova.compute.manager [req-2d675216-982b-45e0-8669-c3b7d42a2fdc req-cf387f63-eb49-46dd-9d0d-c32a66e9d44a service nova] [instance: 41f068d3-2c8d-46f0-8d84-78531d28b0dd] Received event network-vif-deleted-81a37f77-8b23-4fc0-a40d-13c5d36fe6ce {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2021.903730] env[62824]: DEBUG oslo_vmware.api [None req-b4894b47-a9c8-4b49-af68-93eace3f9a8c tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2146022, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2021.973216] env[62824]: DEBUG nova.compute.manager [req-58746295-9d1b-4ec0-81ad-7dba63edefa0 req-19570620-7d66-42d7-8d49-71e0f8b0a2b5 service nova] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] Received event network-vif-plugged-2d7ab792-8a41-4c8b-91a4-e4a677466a12 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2021.973216] env[62824]: DEBUG oslo_concurrency.lockutils [req-58746295-9d1b-4ec0-81ad-7dba63edefa0 req-19570620-7d66-42d7-8d49-71e0f8b0a2b5 service nova] Acquiring lock "b9f30f87-3594-4468-9d29-70890d8761e3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2021.973216] env[62824]: DEBUG oslo_concurrency.lockutils [req-58746295-9d1b-4ec0-81ad-7dba63edefa0 req-19570620-7d66-42d7-8d49-71e0f8b0a2b5 service nova] Lock "b9f30f87-3594-4468-9d29-70890d8761e3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2021.973216] env[62824]: DEBUG oslo_concurrency.lockutils [req-58746295-9d1b-4ec0-81ad-7dba63edefa0 req-19570620-7d66-42d7-8d49-71e0f8b0a2b5 service nova] Lock "b9f30f87-3594-4468-9d29-70890d8761e3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2021.974891] env[62824]: DEBUG nova.compute.manager [req-58746295-9d1b-4ec0-81ad-7dba63edefa0 req-19570620-7d66-42d7-8d49-71e0f8b0a2b5 service nova] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] No waiting events found dispatching network-vif-plugged-2d7ab792-8a41-4c8b-91a4-e4a677466a12 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2021.974984] env[62824]: WARNING nova.compute.manager [req-58746295-9d1b-4ec0-81ad-7dba63edefa0 req-19570620-7d66-42d7-8d49-71e0f8b0a2b5 service nova] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] Received unexpected event network-vif-plugged-2d7ab792-8a41-4c8b-91a4-e4a677466a12 for instance with vm_state building and task_state spawning. [ 2022.112639] env[62824]: DEBUG nova.network.neutron [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] Successfully updated port: 2d7ab792-8a41-4c8b-91a4-e4a677466a12 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2022.162137] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.669s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2022.162137] env[62824]: DEBUG nova.compute.manager [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02405b17-7022-4e2c-a357-415de88f63d0] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2022.164726] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7a7b2101-f65d-4821-85ef-6e89be960d5e tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 33.115s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2022.164991] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7a7b2101-f65d-4821-85ef-6e89be960d5e tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2022.167658] env[62824]: DEBUG oslo_concurrency.lockutils [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 32.058s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2022.169865] env[62824]: INFO nova.compute.claims [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: c65ccd20-b682-420a-9c1a-47a45959197e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2022.202246] env[62824]: INFO nova.scheduler.client.report [None req-7a7b2101-f65d-4821-85ef-6e89be960d5e tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Deleted allocations for instance 5f1237ff-5841-41f1-8d77-df9fabaa62b6 [ 2022.407990] env[62824]: DEBUG oslo_vmware.api [None req-b4894b47-a9c8-4b49-af68-93eace3f9a8c tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2146022, 'name': CloneVM_Task, 'duration_secs': 1.476684} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2022.408728] env[62824]: INFO nova.virt.vmwareapi.vmops [None req-b4894b47-a9c8-4b49-af68-93eace3f9a8c tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2] Created linked-clone VM from snapshot [ 2022.409876] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e85c778-0b2a-4c8c-92a3-29243530a5c1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2022.418058] env[62824]: DEBUG nova.virt.vmwareapi.images [None req-b4894b47-a9c8-4b49-af68-93eace3f9a8c tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2] Uploading image 47f7084b-090d-4f6a-b175-4c8a7ded78e5 {{(pid=62824) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 2022.438027] env[62824]: DEBUG oslo_vmware.rw_handles [None req-b4894b47-a9c8-4b49-af68-93eace3f9a8c tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 2022.438027] env[62824]: value = "vm-438615" [ 2022.438027] env[62824]: _type = "VirtualMachine" [ 2022.438027] env[62824]: }. {{(pid=62824) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 2022.438298] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-1c85fe44-5f24-4b32-a9fd-81c67b8b53ca {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2022.444536] env[62824]: DEBUG oslo_vmware.rw_handles [None req-b4894b47-a9c8-4b49-af68-93eace3f9a8c tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Lease: (returnval){ [ 2022.444536] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52eae88f-5908-ce5c-432d-d9cbb05c34d1" [ 2022.444536] env[62824]: _type = "HttpNfcLease" [ 2022.444536] env[62824]: } obtained for exporting VM: (result){ [ 2022.444536] env[62824]: value = "vm-438615" [ 2022.444536] env[62824]: _type = "VirtualMachine" [ 2022.444536] env[62824]: }. {{(pid=62824) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 2022.444815] env[62824]: DEBUG oslo_vmware.api [None req-b4894b47-a9c8-4b49-af68-93eace3f9a8c tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Waiting for the lease: (returnval){ [ 2022.444815] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52eae88f-5908-ce5c-432d-d9cbb05c34d1" [ 2022.444815] env[62824]: _type = "HttpNfcLease" [ 2022.444815] env[62824]: } to be ready. {{(pid=62824) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 2022.451008] env[62824]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 2022.451008] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52eae88f-5908-ce5c-432d-d9cbb05c34d1" [ 2022.451008] env[62824]: _type = "HttpNfcLease" [ 2022.451008] env[62824]: } is initializing. {{(pid=62824) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 2022.615548] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Acquiring lock "refresh_cache-b9f30f87-3594-4468-9d29-70890d8761e3" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2022.615669] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Acquired lock "refresh_cache-b9f30f87-3594-4468-9d29-70890d8761e3" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2022.615827] env[62824]: DEBUG nova.network.neutron [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2022.676904] env[62824]: DEBUG nova.compute.utils [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2022.678328] env[62824]: DEBUG nova.compute.manager [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02405b17-7022-4e2c-a357-415de88f63d0] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2022.681390] env[62824]: DEBUG nova.network.neutron [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02405b17-7022-4e2c-a357-415de88f63d0] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2022.714485] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7a7b2101-f65d-4821-85ef-6e89be960d5e tempest-ServerGroupTestJSON-540441561 tempest-ServerGroupTestJSON-540441561-project-member] Lock "5f1237ff-5841-41f1-8d77-df9fabaa62b6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 37.562s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2022.741364] env[62824]: DEBUG nova.policy [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a6a63328f3534c7c912384254ccda1f0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '248af2bc3e7b41d2a74b0c6074b359e1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2022.953968] env[62824]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 2022.953968] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52eae88f-5908-ce5c-432d-d9cbb05c34d1" [ 2022.953968] env[62824]: _type = "HttpNfcLease" [ 2022.953968] env[62824]: } is ready. {{(pid=62824) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 2022.955749] env[62824]: DEBUG oslo_vmware.rw_handles [None req-b4894b47-a9c8-4b49-af68-93eace3f9a8c tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 2022.955749] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52eae88f-5908-ce5c-432d-d9cbb05c34d1" [ 2022.955749] env[62824]: _type = "HttpNfcLease" [ 2022.955749] env[62824]: }. {{(pid=62824) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 2022.956940] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a35fb543-1473-4a52-b467-c553f526191d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2022.965715] env[62824]: DEBUG oslo_vmware.rw_handles [None req-b4894b47-a9c8-4b49-af68-93eace3f9a8c tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c54850-614d-97c0-8298-0d56ca98ffd3/disk-0.vmdk from lease info. {{(pid=62824) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 2022.966340] env[62824]: DEBUG oslo_vmware.rw_handles [None req-b4894b47-a9c8-4b49-af68-93eace3f9a8c tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c54850-614d-97c0-8298-0d56ca98ffd3/disk-0.vmdk for reading. {{(pid=62824) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 2023.066019] env[62824]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-b25e2f5d-2318-4f0d-a887-95b775b7f80a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2023.176233] env[62824]: DEBUG nova.network.neutron [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2023.184848] env[62824]: DEBUG nova.compute.manager [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02405b17-7022-4e2c-a357-415de88f63d0] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2023.294164] env[62824]: DEBUG nova.network.neutron [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02405b17-7022-4e2c-a357-415de88f63d0] Successfully created port: cab3935d-90df-47ef-9198-10691fe660b4 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2023.390039] env[62824]: DEBUG nova.network.neutron [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] Updating instance_info_cache with network_info: [{"id": "2d7ab792-8a41-4c8b-91a4-e4a677466a12", "address": "fa:16:3e:20:e4:39", "network": {"id": "d299eecf-1628-4fce-bc67-f9f038d261cd", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1701982809-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d180bbdc8e72414081363fe696ead6fb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae4e3171-21cd-4094-b6cf-81bf366c75bd", "external-id": "nsx-vlan-transportzone-193", "segmentation_id": 193, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2d7ab792-8a", "ovs_interfaceid": "2d7ab792-8a41-4c8b-91a4-e4a677466a12", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2023.837602] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ecb0425-8c3e-4344-bf9a-8b649977798e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2023.848607] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4368b3fd-9d6b-43ee-b10a-49418dceb82d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2023.897214] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Releasing lock "refresh_cache-b9f30f87-3594-4468-9d29-70890d8761e3" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2023.897756] env[62824]: DEBUG nova.compute.manager [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] Instance network_info: |[{"id": "2d7ab792-8a41-4c8b-91a4-e4a677466a12", "address": "fa:16:3e:20:e4:39", "network": {"id": "d299eecf-1628-4fce-bc67-f9f038d261cd", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1701982809-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d180bbdc8e72414081363fe696ead6fb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae4e3171-21cd-4094-b6cf-81bf366c75bd", "external-id": "nsx-vlan-transportzone-193", "segmentation_id": 193, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2d7ab792-8a", "ovs_interfaceid": "2d7ab792-8a41-4c8b-91a4-e4a677466a12", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2023.898956] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:20:e4:39', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ae4e3171-21cd-4094-b6cf-81bf366c75bd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2d7ab792-8a41-4c8b-91a4-e4a677466a12', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2023.910158] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Creating folder: Project (d180bbdc8e72414081363fe696ead6fb). Parent ref: group-v438503. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 2023.911411] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26623072-23dd-4672-8955-7fe70fab3b69 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2023.915667] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-35854e30-067f-42df-b88d-0bfcca73be8b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2023.925541] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fec13b5f-af28-40ac-9d9f-786dfa9b1cdf {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2023.933204] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Created folder: Project (d180bbdc8e72414081363fe696ead6fb) in parent group-v438503. [ 2023.933439] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Creating folder: Instances. Parent ref: group-v438616. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 2023.934522] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-61fcf05d-c62b-4503-813f-7ad9a8f256d4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2023.945648] env[62824]: DEBUG nova.compute.provider_tree [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2023.955702] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Created folder: Instances in parent group-v438616. [ 2023.956218] env[62824]: DEBUG oslo.service.loopingcall [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2023.957161] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2023.957413] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-972328b4-dee9-4cd0-a7c8-1016a4949d44 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2023.978336] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2023.978336] env[62824]: value = "task-2146026" [ 2023.978336] env[62824]: _type = "Task" [ 2023.978336] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2023.987533] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146026, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2024.077761] env[62824]: DEBUG nova.compute.manager [req-92bd7af6-1a45-4148-b829-e7bfc4bacf99 req-48b273dc-9f7f-4334-abab-f9d4c3365b43 service nova] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] Received event network-changed-2d7ab792-8a41-4c8b-91a4-e4a677466a12 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2024.079160] env[62824]: DEBUG nova.compute.manager [req-92bd7af6-1a45-4148-b829-e7bfc4bacf99 req-48b273dc-9f7f-4334-abab-f9d4c3365b43 service nova] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] Refreshing instance network info cache due to event network-changed-2d7ab792-8a41-4c8b-91a4-e4a677466a12. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2024.079481] env[62824]: DEBUG oslo_concurrency.lockutils [req-92bd7af6-1a45-4148-b829-e7bfc4bacf99 req-48b273dc-9f7f-4334-abab-f9d4c3365b43 service nova] Acquiring lock "refresh_cache-b9f30f87-3594-4468-9d29-70890d8761e3" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2024.079641] env[62824]: DEBUG oslo_concurrency.lockutils [req-92bd7af6-1a45-4148-b829-e7bfc4bacf99 req-48b273dc-9f7f-4334-abab-f9d4c3365b43 service nova] Acquired lock "refresh_cache-b9f30f87-3594-4468-9d29-70890d8761e3" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2024.079808] env[62824]: DEBUG nova.network.neutron [req-92bd7af6-1a45-4148-b829-e7bfc4bacf99 req-48b273dc-9f7f-4334-abab-f9d4c3365b43 service nova] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] Refreshing network info cache for port 2d7ab792-8a41-4c8b-91a4-e4a677466a12 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2024.196214] env[62824]: DEBUG nova.compute.manager [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02405b17-7022-4e2c-a357-415de88f63d0] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2024.450566] env[62824]: DEBUG nova.scheduler.client.report [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2024.490885] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146026, 'name': CreateVM_Task, 'duration_secs': 0.507831} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2024.492602] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2024.495698] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2024.495905] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2024.496268] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2024.496533] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9dbb97e1-397a-4699-b6e7-fd117f868611 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2024.502824] env[62824]: DEBUG oslo_vmware.api [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Waiting for the task: (returnval){ [ 2024.502824] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52db8cb1-2193-ae86-a217-854b32c19cd9" [ 2024.502824] env[62824]: _type = "Task" [ 2024.502824] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2024.512419] env[62824]: DEBUG oslo_vmware.api [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52db8cb1-2193-ae86-a217-854b32c19cd9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2024.540750] env[62824]: DEBUG nova.virt.hardware [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2024.541132] env[62824]: DEBUG nova.virt.hardware [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2024.541246] env[62824]: DEBUG nova.virt.hardware [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2024.541393] env[62824]: DEBUG nova.virt.hardware [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2024.541636] env[62824]: DEBUG nova.virt.hardware [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2024.541779] env[62824]: DEBUG nova.virt.hardware [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2024.542148] env[62824]: DEBUG nova.virt.hardware [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2024.542455] env[62824]: DEBUG nova.virt.hardware [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2024.542708] env[62824]: DEBUG nova.virt.hardware [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2024.542902] env[62824]: DEBUG nova.virt.hardware [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2024.543197] env[62824]: DEBUG nova.virt.hardware [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2024.545821] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-557cf181-87e8-494e-8d30-d7d16334a4da {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2024.564424] env[62824]: DEBUG oslo_vmware.rw_handles [None req-58bcc65c-ba1f-4462-9f05-d0e06434f495 tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e52e6b-1132-1e69-310a-5702becf9ef5/disk-0.vmdk. {{(pid=62824) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 2024.565875] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8f75295-f764-47f4-a4a1-05fff5bca4c5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2024.572508] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8df50c5-6c11-4bb5-82f2-ee657e514115 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2024.579184] env[62824]: DEBUG oslo_vmware.rw_handles [None req-58bcc65c-ba1f-4462-9f05-d0e06434f495 tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e52e6b-1132-1e69-310a-5702becf9ef5/disk-0.vmdk is in state: ready. {{(pid=62824) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 2024.579529] env[62824]: ERROR oslo_vmware.rw_handles [None req-58bcc65c-ba1f-4462-9f05-d0e06434f495 tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e52e6b-1132-1e69-310a-5702becf9ef5/disk-0.vmdk due to incomplete transfer. [ 2024.588676] env[62824]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-4f1dc7d6-018b-4463-b4f6-b2cc19627c47 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2024.600594] env[62824]: DEBUG oslo_vmware.rw_handles [None req-58bcc65c-ba1f-4462-9f05-d0e06434f495 tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e52e6b-1132-1e69-310a-5702becf9ef5/disk-0.vmdk. {{(pid=62824) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 2024.600594] env[62824]: DEBUG nova.virt.vmwareapi.images [None req-58bcc65c-ba1f-4462-9f05-d0e06434f495 tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Uploaded image c076cb8e-f451-482d-bd51-2014c6c0bb68 to the Glance image server {{(pid=62824) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 2024.603649] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-58bcc65c-ba1f-4462-9f05-d0e06434f495 tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Destroying the VM {{(pid=62824) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 2024.604633] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-2cd2bf34-7d48-49bf-91e9-4c34f7273460 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2024.612272] env[62824]: DEBUG oslo_vmware.api [None req-58bcc65c-ba1f-4462-9f05-d0e06434f495 tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Waiting for the task: (returnval){ [ 2024.612272] env[62824]: value = "task-2146027" [ 2024.612272] env[62824]: _type = "Task" [ 2024.612272] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2024.621215] env[62824]: DEBUG oslo_vmware.api [None req-58bcc65c-ba1f-4462-9f05-d0e06434f495 tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Task: {'id': task-2146027, 'name': Destroy_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2024.634419] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1beb7be1-384e-40ae-965a-5630a0d7138d tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Acquiring lock "59b5b883-4188-471c-8862-444f3ce08cb0" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2024.634736] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1beb7be1-384e-40ae-965a-5630a0d7138d tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Lock "59b5b883-4188-471c-8862-444f3ce08cb0" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2024.962032] env[62824]: DEBUG oslo_concurrency.lockutils [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.792s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2024.962032] env[62824]: DEBUG nova.compute.manager [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: c65ccd20-b682-420a-9c1a-47a45959197e] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2024.965947] env[62824]: DEBUG oslo_concurrency.lockutils [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 33.197s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2024.969219] env[62824]: INFO nova.compute.claims [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] [instance: 4423d172-acdd-4c69-b3b8-ff166e1b8548] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2025.003364] env[62824]: DEBUG nova.network.neutron [req-92bd7af6-1a45-4148-b829-e7bfc4bacf99 req-48b273dc-9f7f-4334-abab-f9d4c3365b43 service nova] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] Updated VIF entry in instance network info cache for port 2d7ab792-8a41-4c8b-91a4-e4a677466a12. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2025.003742] env[62824]: DEBUG nova.network.neutron [req-92bd7af6-1a45-4148-b829-e7bfc4bacf99 req-48b273dc-9f7f-4334-abab-f9d4c3365b43 service nova] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] Updating instance_info_cache with network_info: [{"id": "2d7ab792-8a41-4c8b-91a4-e4a677466a12", "address": "fa:16:3e:20:e4:39", "network": {"id": "d299eecf-1628-4fce-bc67-f9f038d261cd", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1701982809-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d180bbdc8e72414081363fe696ead6fb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae4e3171-21cd-4094-b6cf-81bf366c75bd", "external-id": "nsx-vlan-transportzone-193", "segmentation_id": 193, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2d7ab792-8a", "ovs_interfaceid": "2d7ab792-8a41-4c8b-91a4-e4a677466a12", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2025.016807] env[62824]: DEBUG oslo_vmware.api [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52db8cb1-2193-ae86-a217-854b32c19cd9, 'name': SearchDatastore_Task, 'duration_secs': 0.023397} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2025.018028] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2025.018124] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2025.018741] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2025.018741] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2025.018741] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2025.019745] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-575171e1-944e-4ee7-81ff-d8d8ecb887e7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2025.029559] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2025.029858] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2025.030657] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f2efc2e9-0755-441f-bdd2-eb4d4ce3a1b5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2025.036780] env[62824]: DEBUG oslo_vmware.api [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Waiting for the task: (returnval){ [ 2025.036780] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5261536a-74b9-3f64-4578-b3587a911e08" [ 2025.036780] env[62824]: _type = "Task" [ 2025.036780] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2025.046421] env[62824]: DEBUG oslo_vmware.api [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5261536a-74b9-3f64-4578-b3587a911e08, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2025.122737] env[62824]: DEBUG oslo_vmware.api [None req-58bcc65c-ba1f-4462-9f05-d0e06434f495 tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Task: {'id': task-2146027, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2025.141172] env[62824]: DEBUG nova.compute.utils [None req-1beb7be1-384e-40ae-965a-5630a0d7138d tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2025.467731] env[62824]: DEBUG nova.compute.utils [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2025.468935] env[62824]: DEBUG nova.compute.manager [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: c65ccd20-b682-420a-9c1a-47a45959197e] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2025.469454] env[62824]: DEBUG nova.network.neutron [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: c65ccd20-b682-420a-9c1a-47a45959197e] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2025.511869] env[62824]: DEBUG oslo_concurrency.lockutils [req-92bd7af6-1a45-4148-b829-e7bfc4bacf99 req-48b273dc-9f7f-4334-abab-f9d4c3365b43 service nova] Releasing lock "refresh_cache-b9f30f87-3594-4468-9d29-70890d8761e3" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2025.552760] env[62824]: DEBUG oslo_vmware.api [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5261536a-74b9-3f64-4578-b3587a911e08, 'name': SearchDatastore_Task, 'duration_secs': 0.014003} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2025.554113] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-960ccf8f-cf6a-4aee-b42a-99afc0effa27 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2025.557872] env[62824]: DEBUG nova.policy [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '28495352107b45cd9cbd746c6affc4fe', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '64042a790d6c459186c68d73da32c019', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2025.563068] env[62824]: DEBUG oslo_vmware.api [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Waiting for the task: (returnval){ [ 2025.563068] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52b9b965-45fa-5823-e322-f4c1b049b7ae" [ 2025.563068] env[62824]: _type = "Task" [ 2025.563068] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2025.571769] env[62824]: DEBUG oslo_vmware.api [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52b9b965-45fa-5823-e322-f4c1b049b7ae, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2025.625599] env[62824]: DEBUG oslo_vmware.api [None req-58bcc65c-ba1f-4462-9f05-d0e06434f495 tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Task: {'id': task-2146027, 'name': Destroy_Task, 'duration_secs': 0.625323} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2025.625599] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-58bcc65c-ba1f-4462-9f05-d0e06434f495 tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Destroyed the VM [ 2025.625903] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-58bcc65c-ba1f-4462-9f05-d0e06434f495 tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Deleting Snapshot of the VM instance {{(pid=62824) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 2025.626172] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-3ec95928-fb14-4860-b1fa-d66ea15c5b19 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2025.632738] env[62824]: DEBUG oslo_vmware.api [None req-58bcc65c-ba1f-4462-9f05-d0e06434f495 tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Waiting for the task: (returnval){ [ 2025.632738] env[62824]: value = "task-2146028" [ 2025.632738] env[62824]: _type = "Task" [ 2025.632738] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2025.641878] env[62824]: DEBUG oslo_vmware.api [None req-58bcc65c-ba1f-4462-9f05-d0e06434f495 tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Task: {'id': task-2146028, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2025.646759] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1beb7be1-384e-40ae-965a-5630a0d7138d tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Lock "59b5b883-4188-471c-8862-444f3ce08cb0" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.012s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2025.661249] env[62824]: DEBUG nova.network.neutron [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02405b17-7022-4e2c-a357-415de88f63d0] Successfully updated port: cab3935d-90df-47ef-9198-10691fe660b4 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2025.875627] env[62824]: DEBUG nova.network.neutron [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: c65ccd20-b682-420a-9c1a-47a45959197e] Successfully created port: 344e044b-aab1-485b-bd89-5477bc394098 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2025.973193] env[62824]: DEBUG nova.compute.manager [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: c65ccd20-b682-420a-9c1a-47a45959197e] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2026.076722] env[62824]: DEBUG oslo_vmware.api [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52b9b965-45fa-5823-e322-f4c1b049b7ae, 'name': SearchDatastore_Task, 'duration_secs': 0.016031} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2026.076988] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2026.077398] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] b9f30f87-3594-4468-9d29-70890d8761e3/b9f30f87-3594-4468-9d29-70890d8761e3.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2026.077676] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-72071c6e-bc70-46ec-8979-a1ab22793c21 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2026.085352] env[62824]: DEBUG oslo_vmware.api [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Waiting for the task: (returnval){ [ 2026.085352] env[62824]: value = "task-2146029" [ 2026.085352] env[62824]: _type = "Task" [ 2026.085352] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2026.097218] env[62824]: DEBUG oslo_vmware.api [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Task: {'id': task-2146029, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2026.147653] env[62824]: DEBUG oslo_vmware.api [None req-58bcc65c-ba1f-4462-9f05-d0e06434f495 tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Task: {'id': task-2146028, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2026.165895] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Acquiring lock "refresh_cache-02405b17-7022-4e2c-a357-415de88f63d0" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2026.167991] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Acquired lock "refresh_cache-02405b17-7022-4e2c-a357-415de88f63d0" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2026.167991] env[62824]: DEBUG nova.network.neutron [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02405b17-7022-4e2c-a357-415de88f63d0] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2026.555147] env[62824]: DEBUG nova.compute.manager [req-6a2a83a7-2a6b-463f-8d38-11bb359056fd req-83951db6-a0b3-4d49-8461-e800938e83bb service nova] [instance: 02405b17-7022-4e2c-a357-415de88f63d0] Received event network-vif-plugged-cab3935d-90df-47ef-9198-10691fe660b4 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2026.555416] env[62824]: DEBUG oslo_concurrency.lockutils [req-6a2a83a7-2a6b-463f-8d38-11bb359056fd req-83951db6-a0b3-4d49-8461-e800938e83bb service nova] Acquiring lock "02405b17-7022-4e2c-a357-415de88f63d0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2026.555779] env[62824]: DEBUG oslo_concurrency.lockutils [req-6a2a83a7-2a6b-463f-8d38-11bb359056fd req-83951db6-a0b3-4d49-8461-e800938e83bb service nova] Lock "02405b17-7022-4e2c-a357-415de88f63d0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2026.555779] env[62824]: DEBUG oslo_concurrency.lockutils [req-6a2a83a7-2a6b-463f-8d38-11bb359056fd req-83951db6-a0b3-4d49-8461-e800938e83bb service nova] Lock "02405b17-7022-4e2c-a357-415de88f63d0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2026.555941] env[62824]: DEBUG nova.compute.manager [req-6a2a83a7-2a6b-463f-8d38-11bb359056fd req-83951db6-a0b3-4d49-8461-e800938e83bb service nova] [instance: 02405b17-7022-4e2c-a357-415de88f63d0] No waiting events found dispatching network-vif-plugged-cab3935d-90df-47ef-9198-10691fe660b4 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2026.556189] env[62824]: WARNING nova.compute.manager [req-6a2a83a7-2a6b-463f-8d38-11bb359056fd req-83951db6-a0b3-4d49-8461-e800938e83bb service nova] [instance: 02405b17-7022-4e2c-a357-415de88f63d0] Received unexpected event network-vif-plugged-cab3935d-90df-47ef-9198-10691fe660b4 for instance with vm_state building and task_state spawning. [ 2026.556366] env[62824]: DEBUG nova.compute.manager [req-6a2a83a7-2a6b-463f-8d38-11bb359056fd req-83951db6-a0b3-4d49-8461-e800938e83bb service nova] [instance: 02405b17-7022-4e2c-a357-415de88f63d0] Received event network-changed-cab3935d-90df-47ef-9198-10691fe660b4 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2026.556511] env[62824]: DEBUG nova.compute.manager [req-6a2a83a7-2a6b-463f-8d38-11bb359056fd req-83951db6-a0b3-4d49-8461-e800938e83bb service nova] [instance: 02405b17-7022-4e2c-a357-415de88f63d0] Refreshing instance network info cache due to event network-changed-cab3935d-90df-47ef-9198-10691fe660b4. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2026.556676] env[62824]: DEBUG oslo_concurrency.lockutils [req-6a2a83a7-2a6b-463f-8d38-11bb359056fd req-83951db6-a0b3-4d49-8461-e800938e83bb service nova] Acquiring lock "refresh_cache-02405b17-7022-4e2c-a357-415de88f63d0" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2026.598309] env[62824]: DEBUG oslo_vmware.api [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Task: {'id': task-2146029, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2026.631041] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca1d4431-6b65-456c-884f-282c85567c49 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2026.643131] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2afa968-ffba-4948-8cac-ac85a677bc0d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2026.650400] env[62824]: DEBUG oslo_vmware.api [None req-58bcc65c-ba1f-4462-9f05-d0e06434f495 tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Task: {'id': task-2146028, 'name': RemoveSnapshot_Task, 'duration_secs': 0.872985} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2026.651349] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-58bcc65c-ba1f-4462-9f05-d0e06434f495 tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Deleted Snapshot of the VM instance {{(pid=62824) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 2026.651805] env[62824]: INFO nova.compute.manager [None req-58bcc65c-ba1f-4462-9f05-d0e06434f495 tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Took 16.56 seconds to snapshot the instance on the hypervisor. [ 2026.689183] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12ce6759-0ee3-4999-bcd4-f7600cbf0d45 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2026.703199] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efc9828c-aacc-4ebb-9bbd-f8266eee9f95 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2026.719845] env[62824]: DEBUG nova.compute.provider_tree [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2026.745790] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1beb7be1-384e-40ae-965a-5630a0d7138d tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Acquiring lock "59b5b883-4188-471c-8862-444f3ce08cb0" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2026.746107] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1beb7be1-384e-40ae-965a-5630a0d7138d tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Lock "59b5b883-4188-471c-8862-444f3ce08cb0" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2026.746380] env[62824]: INFO nova.compute.manager [None req-1beb7be1-384e-40ae-965a-5630a0d7138d tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Attaching volume 5c4381e7-4dfa-4683-93fe-8eb8bef1df7e to /dev/sdb [ 2026.779226] env[62824]: DEBUG nova.network.neutron [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02405b17-7022-4e2c-a357-415de88f63d0] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2026.784274] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-916b2043-0625-4be1-95e4-642eee6a23a8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2026.791998] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a5365bf-e26e-4d75-bf9f-10460e19dac2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2026.806960] env[62824]: DEBUG nova.virt.block_device [None req-1beb7be1-384e-40ae-965a-5630a0d7138d tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Updating existing volume attachment record: 0b39b411-e651-4cf8-8867-23b891b15137 {{(pid=62824) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 2026.989169] env[62824]: DEBUG nova.compute.manager [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: c65ccd20-b682-420a-9c1a-47a45959197e] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2027.029662] env[62824]: DEBUG nova.virt.hardware [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2027.033031] env[62824]: DEBUG nova.virt.hardware [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2027.033031] env[62824]: DEBUG nova.virt.hardware [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2027.033031] env[62824]: DEBUG nova.virt.hardware [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2027.033031] env[62824]: DEBUG nova.virt.hardware [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2027.033031] env[62824]: DEBUG nova.virt.hardware [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2027.033031] env[62824]: DEBUG nova.virt.hardware [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2027.033031] env[62824]: DEBUG nova.virt.hardware [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2027.033031] env[62824]: DEBUG nova.virt.hardware [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2027.033031] env[62824]: DEBUG nova.virt.hardware [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2027.033031] env[62824]: DEBUG nova.virt.hardware [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2027.033031] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52e33c9b-4f4b-4f3d-9287-51a1326ec546 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2027.042537] env[62824]: DEBUG nova.network.neutron [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02405b17-7022-4e2c-a357-415de88f63d0] Updating instance_info_cache with network_info: [{"id": "cab3935d-90df-47ef-9198-10691fe660b4", "address": "fa:16:3e:09:37:39", "network": {"id": "fc381ca5-198f-4afc-9ebf-bc6e3de68c35", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-425455451-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "248af2bc3e7b41d2a74b0c6074b359e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4c9a12d2-469f-4199-bfaa-f791d765deac", "external-id": "nsx-vlan-transportzone-96", "segmentation_id": 96, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcab3935d-90", "ovs_interfaceid": "cab3935d-90df-47ef-9198-10691fe660b4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2027.045172] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b953dc0-bb8d-423c-ba35-a6095fee0c88 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2027.099503] env[62824]: DEBUG oslo_vmware.api [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Task: {'id': task-2146029, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.690192} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2027.099833] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] b9f30f87-3594-4468-9d29-70890d8761e3/b9f30f87-3594-4468-9d29-70890d8761e3.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2027.100087] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2027.100374] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7873528a-b5d9-4600-886e-21de3e14946c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2027.107091] env[62824]: DEBUG oslo_vmware.api [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Waiting for the task: (returnval){ [ 2027.107091] env[62824]: value = "task-2146031" [ 2027.107091] env[62824]: _type = "Task" [ 2027.107091] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2027.117108] env[62824]: DEBUG oslo_vmware.api [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Task: {'id': task-2146031, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2027.224440] env[62824]: DEBUG nova.scheduler.client.report [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2027.550246] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Releasing lock "refresh_cache-02405b17-7022-4e2c-a357-415de88f63d0" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2027.550640] env[62824]: DEBUG nova.compute.manager [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02405b17-7022-4e2c-a357-415de88f63d0] Instance network_info: |[{"id": "cab3935d-90df-47ef-9198-10691fe660b4", "address": "fa:16:3e:09:37:39", "network": {"id": "fc381ca5-198f-4afc-9ebf-bc6e3de68c35", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-425455451-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "248af2bc3e7b41d2a74b0c6074b359e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4c9a12d2-469f-4199-bfaa-f791d765deac", "external-id": "nsx-vlan-transportzone-96", "segmentation_id": 96, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcab3935d-90", "ovs_interfaceid": "cab3935d-90df-47ef-9198-10691fe660b4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2027.550951] env[62824]: DEBUG oslo_concurrency.lockutils [req-6a2a83a7-2a6b-463f-8d38-11bb359056fd req-83951db6-a0b3-4d49-8461-e800938e83bb service nova] Acquired lock "refresh_cache-02405b17-7022-4e2c-a357-415de88f63d0" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2027.551151] env[62824]: DEBUG nova.network.neutron [req-6a2a83a7-2a6b-463f-8d38-11bb359056fd req-83951db6-a0b3-4d49-8461-e800938e83bb service nova] [instance: 02405b17-7022-4e2c-a357-415de88f63d0] Refreshing network info cache for port cab3935d-90df-47ef-9198-10691fe660b4 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2027.552394] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02405b17-7022-4e2c-a357-415de88f63d0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:09:37:39', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4c9a12d2-469f-4199-bfaa-f791d765deac', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cab3935d-90df-47ef-9198-10691fe660b4', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2027.563085] env[62824]: DEBUG oslo.service.loopingcall [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2027.567423] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 02405b17-7022-4e2c-a357-415de88f63d0] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2027.568165] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a88461a2-d271-485d-9930-99478c334699 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2027.591936] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2027.591936] env[62824]: value = "task-2146034" [ 2027.591936] env[62824]: _type = "Task" [ 2027.591936] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2027.603132] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146034, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2027.618188] env[62824]: DEBUG oslo_vmware.api [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Task: {'id': task-2146031, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.291147} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2027.618685] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2027.620220] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e304f33e-bfb7-4b00-8c0d-9d8491d685ff {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2027.649020] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] Reconfiguring VM instance instance-0000002f to attach disk [datastore1] b9f30f87-3594-4468-9d29-70890d8761e3/b9f30f87-3594-4468-9d29-70890d8761e3.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2027.653542] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-534e4c6a-1929-4cd7-b327-ad031e3c64f7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2027.682320] env[62824]: DEBUG oslo_vmware.api [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Waiting for the task: (returnval){ [ 2027.682320] env[62824]: value = "task-2146035" [ 2027.682320] env[62824]: _type = "Task" [ 2027.682320] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2027.691541] env[62824]: DEBUG oslo_vmware.api [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Task: {'id': task-2146035, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2027.730773] env[62824]: DEBUG oslo_concurrency.lockutils [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.764s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2027.730773] env[62824]: DEBUG nova.compute.manager [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] [instance: 4423d172-acdd-4c69-b3b8-ff166e1b8548] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2027.733648] env[62824]: DEBUG oslo_concurrency.lockutils [None req-870c96bf-3657-45b9-96d5-c5182b858ac4 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 33.473s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2027.733791] env[62824]: DEBUG oslo_concurrency.lockutils [None req-870c96bf-3657-45b9-96d5-c5182b858ac4 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2027.737061] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.787s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2027.738716] env[62824]: INFO nova.compute.claims [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 88bad2dd-dce7-41df-b56c-93a5d054c11e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2027.772773] env[62824]: DEBUG nova.compute.manager [req-2ae7a1c3-8165-4afb-ab77-2f6cbfd7aac5 req-8fec336d-147b-4626-b9b3-0496b2626959 service nova] [instance: c65ccd20-b682-420a-9c1a-47a45959197e] Received event network-vif-plugged-344e044b-aab1-485b-bd89-5477bc394098 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2027.772892] env[62824]: DEBUG oslo_concurrency.lockutils [req-2ae7a1c3-8165-4afb-ab77-2f6cbfd7aac5 req-8fec336d-147b-4626-b9b3-0496b2626959 service nova] Acquiring lock "c65ccd20-b682-420a-9c1a-47a45959197e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2027.773815] env[62824]: DEBUG oslo_concurrency.lockutils [req-2ae7a1c3-8165-4afb-ab77-2f6cbfd7aac5 req-8fec336d-147b-4626-b9b3-0496b2626959 service nova] Lock "c65ccd20-b682-420a-9c1a-47a45959197e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2027.773815] env[62824]: DEBUG oslo_concurrency.lockutils [req-2ae7a1c3-8165-4afb-ab77-2f6cbfd7aac5 req-8fec336d-147b-4626-b9b3-0496b2626959 service nova] Lock "c65ccd20-b682-420a-9c1a-47a45959197e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2027.773815] env[62824]: DEBUG nova.compute.manager [req-2ae7a1c3-8165-4afb-ab77-2f6cbfd7aac5 req-8fec336d-147b-4626-b9b3-0496b2626959 service nova] [instance: c65ccd20-b682-420a-9c1a-47a45959197e] No waiting events found dispatching network-vif-plugged-344e044b-aab1-485b-bd89-5477bc394098 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2027.773815] env[62824]: WARNING nova.compute.manager [req-2ae7a1c3-8165-4afb-ab77-2f6cbfd7aac5 req-8fec336d-147b-4626-b9b3-0496b2626959 service nova] [instance: c65ccd20-b682-420a-9c1a-47a45959197e] Received unexpected event network-vif-plugged-344e044b-aab1-485b-bd89-5477bc394098 for instance with vm_state building and task_state spawning. [ 2027.797591] env[62824]: INFO nova.scheduler.client.report [None req-870c96bf-3657-45b9-96d5-c5182b858ac4 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Deleted allocations for instance dc3317e4-6958-4df6-8e39-c89b4a5d861c [ 2027.833306] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9d78c884-a693-423a-b835-e548c8af80b6 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Acquiring lock "b96d1351-f3a7-4bac-998c-a34ab2606041" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2027.833583] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9d78c884-a693-423a-b835-e548c8af80b6 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Lock "b96d1351-f3a7-4bac-998c-a34ab2606041" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2027.833858] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9d78c884-a693-423a-b835-e548c8af80b6 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Acquiring lock "b96d1351-f3a7-4bac-998c-a34ab2606041-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2027.833989] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9d78c884-a693-423a-b835-e548c8af80b6 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Lock "b96d1351-f3a7-4bac-998c-a34ab2606041-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2027.834208] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9d78c884-a693-423a-b835-e548c8af80b6 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Lock "b96d1351-f3a7-4bac-998c-a34ab2606041-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2027.836579] env[62824]: INFO nova.compute.manager [None req-9d78c884-a693-423a-b835-e548c8af80b6 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] [instance: b96d1351-f3a7-4bac-998c-a34ab2606041] Terminating instance [ 2028.062364] env[62824]: DEBUG nova.network.neutron [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: c65ccd20-b682-420a-9c1a-47a45959197e] Successfully updated port: 344e044b-aab1-485b-bd89-5477bc394098 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2028.105039] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146034, 'name': CreateVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2028.124372] env[62824]: DEBUG nova.network.neutron [req-6a2a83a7-2a6b-463f-8d38-11bb359056fd req-83951db6-a0b3-4d49-8461-e800938e83bb service nova] [instance: 02405b17-7022-4e2c-a357-415de88f63d0] Updated VIF entry in instance network info cache for port cab3935d-90df-47ef-9198-10691fe660b4. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2028.124372] env[62824]: DEBUG nova.network.neutron [req-6a2a83a7-2a6b-463f-8d38-11bb359056fd req-83951db6-a0b3-4d49-8461-e800938e83bb service nova] [instance: 02405b17-7022-4e2c-a357-415de88f63d0] Updating instance_info_cache with network_info: [{"id": "cab3935d-90df-47ef-9198-10691fe660b4", "address": "fa:16:3e:09:37:39", "network": {"id": "fc381ca5-198f-4afc-9ebf-bc6e3de68c35", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-425455451-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "248af2bc3e7b41d2a74b0c6074b359e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4c9a12d2-469f-4199-bfaa-f791d765deac", "external-id": "nsx-vlan-transportzone-96", "segmentation_id": 96, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcab3935d-90", "ovs_interfaceid": "cab3935d-90df-47ef-9198-10691fe660b4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2028.193361] env[62824]: DEBUG oslo_vmware.api [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Task: {'id': task-2146035, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2028.246704] env[62824]: DEBUG nova.compute.utils [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2028.255480] env[62824]: DEBUG nova.compute.manager [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] [instance: 4423d172-acdd-4c69-b3b8-ff166e1b8548] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2028.255480] env[62824]: DEBUG nova.network.neutron [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] [instance: 4423d172-acdd-4c69-b3b8-ff166e1b8548] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2028.305995] env[62824]: DEBUG oslo_concurrency.lockutils [None req-870c96bf-3657-45b9-96d5-c5182b858ac4 tempest-ServersTestMultiNic-2095051118 tempest-ServersTestMultiNic-2095051118-project-member] Lock "dc3317e4-6958-4df6-8e39-c89b4a5d861c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 38.200s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2028.344425] env[62824]: DEBUG nova.compute.manager [None req-9d78c884-a693-423a-b835-e548c8af80b6 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] [instance: b96d1351-f3a7-4bac-998c-a34ab2606041] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2028.345358] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d78c884-a693-423a-b835-e548c8af80b6 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] [instance: b96d1351-f3a7-4bac-998c-a34ab2606041] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2028.345738] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-251f474e-1fc6-4a9f-8999-f60f967386df {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2028.358190] env[62824]: DEBUG oslo_vmware.api [None req-9d78c884-a693-423a-b835-e548c8af80b6 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Waiting for the task: (returnval){ [ 2028.358190] env[62824]: value = "task-2146036" [ 2028.358190] env[62824]: _type = "Task" [ 2028.358190] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2028.364543] env[62824]: DEBUG nova.policy [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0059ec60e4884d22addac429b94794e9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '78831267dd354243b0f284437b835ef5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2028.381262] env[62824]: DEBUG oslo_vmware.api [None req-9d78c884-a693-423a-b835-e548c8af80b6 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Task: {'id': task-2146036, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2028.567417] env[62824]: DEBUG oslo_concurrency.lockutils [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Acquiring lock "refresh_cache-c65ccd20-b682-420a-9c1a-47a45959197e" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2028.567417] env[62824]: DEBUG oslo_concurrency.lockutils [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Acquired lock "refresh_cache-c65ccd20-b682-420a-9c1a-47a45959197e" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2028.567417] env[62824]: DEBUG nova.network.neutron [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: c65ccd20-b682-420a-9c1a-47a45959197e] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2028.608736] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146034, 'name': CreateVM_Task, 'duration_secs': 0.536828} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2028.608906] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 02405b17-7022-4e2c-a357-415de88f63d0] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2028.609953] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2028.609953] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2028.610133] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2028.610424] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cb8a6223-8d3c-412a-b02d-6b0895405222 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2028.618211] env[62824]: DEBUG oslo_vmware.api [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for the task: (returnval){ [ 2028.618211] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52b4d72e-2845-f503-4852-e8dbe3852234" [ 2028.618211] env[62824]: _type = "Task" [ 2028.618211] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2028.627575] env[62824]: DEBUG oslo_concurrency.lockutils [req-6a2a83a7-2a6b-463f-8d38-11bb359056fd req-83951db6-a0b3-4d49-8461-e800938e83bb service nova] Releasing lock "refresh_cache-02405b17-7022-4e2c-a357-415de88f63d0" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2028.632900] env[62824]: DEBUG oslo_vmware.api [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52b4d72e-2845-f503-4852-e8dbe3852234, 'name': SearchDatastore_Task, 'duration_secs': 0.010992} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2028.633324] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2028.633557] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02405b17-7022-4e2c-a357-415de88f63d0] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2028.633806] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2028.633984] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2028.634166] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2028.634456] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d6e13260-b291-4d18-b3b2-af8c3c6d0cb0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2028.643584] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2028.643584] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2028.644134] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d4b5d93f-1093-4169-87f3-0a3ca4e3137e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2028.652125] env[62824]: DEBUG oslo_vmware.api [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for the task: (returnval){ [ 2028.652125] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5244f297-9c7d-f144-f1c2-48f6cb7d2d5d" [ 2028.652125] env[62824]: _type = "Task" [ 2028.652125] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2028.666021] env[62824]: DEBUG oslo_vmware.api [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5244f297-9c7d-f144-f1c2-48f6cb7d2d5d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2028.692082] env[62824]: DEBUG oslo_vmware.api [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Task: {'id': task-2146035, 'name': ReconfigVM_Task, 'duration_secs': 0.614138} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2028.692387] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] Reconfigured VM instance instance-0000002f to attach disk [datastore1] b9f30f87-3594-4468-9d29-70890d8761e3/b9f30f87-3594-4468-9d29-70890d8761e3.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2028.693082] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dad57f8f-1c3f-45b6-bdee-1dd945f52259 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2028.699334] env[62824]: DEBUG oslo_vmware.api [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Waiting for the task: (returnval){ [ 2028.699334] env[62824]: value = "task-2146037" [ 2028.699334] env[62824]: _type = "Task" [ 2028.699334] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2028.707637] env[62824]: DEBUG oslo_vmware.api [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Task: {'id': task-2146037, 'name': Rename_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2028.759703] env[62824]: DEBUG nova.compute.manager [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] [instance: 4423d172-acdd-4c69-b3b8-ff166e1b8548] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2028.883021] env[62824]: DEBUG oslo_vmware.api [None req-9d78c884-a693-423a-b835-e548c8af80b6 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Task: {'id': task-2146036, 'name': PowerOffVM_Task, 'duration_secs': 0.241597} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2028.884599] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d78c884-a693-423a-b835-e548c8af80b6 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] [instance: b96d1351-f3a7-4bac-998c-a34ab2606041] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2028.885613] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d78c884-a693-423a-b835-e548c8af80b6 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] [instance: b96d1351-f3a7-4bac-998c-a34ab2606041] Volume detach. Driver type: vmdk {{(pid=62824) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 2028.885613] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d78c884-a693-423a-b835-e548c8af80b6 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] [instance: b96d1351-f3a7-4bac-998c-a34ab2606041] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-438510', 'volume_id': '64941ab3-102e-48a5-a3ad-433fb818850b', 'name': 'volume-64941ab3-102e-48a5-a3ad-433fb818850b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'b96d1351-f3a7-4bac-998c-a34ab2606041', 'attached_at': '', 'detached_at': '', 'volume_id': '64941ab3-102e-48a5-a3ad-433fb818850b', 'serial': '64941ab3-102e-48a5-a3ad-433fb818850b'} {{(pid=62824) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 2028.887550] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2ad1cd6-870a-4a91-81a1-e67a5cd19932 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2028.918313] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af9a53c3-9dfb-4c80-b29f-17b20e8afe38 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2028.927103] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e328099-1716-4df8-a9b9-ae01514662f6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2028.954384] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4090db99-42b1-46da-b928-e34b16445f87 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2028.974154] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d78c884-a693-423a-b835-e548c8af80b6 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] The volume has not been displaced from its original location: [datastore2] volume-64941ab3-102e-48a5-a3ad-433fb818850b/volume-64941ab3-102e-48a5-a3ad-433fb818850b.vmdk. No consolidation needed. {{(pid=62824) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 2028.980373] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d78c884-a693-423a-b835-e548c8af80b6 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] [instance: b96d1351-f3a7-4bac-998c-a34ab2606041] Reconfiguring VM instance instance-0000001f to detach disk 2000 {{(pid=62824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 2028.984242] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-97a73d34-5919-46ff-a22d-69ce416c9a60 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2029.004131] env[62824]: DEBUG oslo_vmware.api [None req-9d78c884-a693-423a-b835-e548c8af80b6 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Waiting for the task: (returnval){ [ 2029.004131] env[62824]: value = "task-2146038" [ 2029.004131] env[62824]: _type = "Task" [ 2029.004131] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2029.016496] env[62824]: DEBUG oslo_vmware.api [None req-9d78c884-a693-423a-b835-e548c8af80b6 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Task: {'id': task-2146038, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2029.116739] env[62824]: DEBUG nova.network.neutron [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: c65ccd20-b682-420a-9c1a-47a45959197e] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2029.170236] env[62824]: DEBUG oslo_vmware.api [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5244f297-9c7d-f144-f1c2-48f6cb7d2d5d, 'name': SearchDatastore_Task, 'duration_secs': 0.018772} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2029.170821] env[62824]: DEBUG nova.network.neutron [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] [instance: 4423d172-acdd-4c69-b3b8-ff166e1b8548] Successfully created port: 0f1c4fc4-1d12-4800-9861-68137c05d00f {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2029.177720] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5b814a08-a8f8-465d-8621-78afc53d6c87 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2029.187098] env[62824]: DEBUG oslo_vmware.api [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for the task: (returnval){ [ 2029.187098] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52bd6ba6-4659-060b-8c2c-b8f238461690" [ 2029.187098] env[62824]: _type = "Task" [ 2029.187098] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2029.196486] env[62824]: DEBUG oslo_vmware.api [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52bd6ba6-4659-060b-8c2c-b8f238461690, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2029.208361] env[62824]: DEBUG oslo_vmware.api [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Task: {'id': task-2146037, 'name': Rename_Task, 'duration_secs': 0.179153} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2029.208650] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2029.208939] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3724fe13-d77a-4851-9f05-b28ad6744341 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2029.218393] env[62824]: DEBUG oslo_vmware.api [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Waiting for the task: (returnval){ [ 2029.218393] env[62824]: value = "task-2146039" [ 2029.218393] env[62824]: _type = "Task" [ 2029.218393] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2029.229929] env[62824]: DEBUG oslo_vmware.api [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Task: {'id': task-2146039, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2029.300643] env[62824]: DEBUG nova.network.neutron [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: c65ccd20-b682-420a-9c1a-47a45959197e] Updating instance_info_cache with network_info: [{"id": "344e044b-aab1-485b-bd89-5477bc394098", "address": "fa:16:3e:4d:bf:ba", "network": {"id": "9e1885b6-0796-42b9-9cd8-8fa9ce9a3cfd", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-2118936609-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "64042a790d6c459186c68d73da32c019", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "afae6acd-1873-4228-9d5a-1cd5d4efe3e4", "external-id": "nsx-vlan-transportzone-183", "segmentation_id": 183, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap344e044b-aa", "ovs_interfaceid": "344e044b-aab1-485b-bd89-5477bc394098", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2029.472917] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9d7e564-b313-4c54-879c-2876731d7202 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2029.481827] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-151bb266-fe1c-4206-83f7-5ede0caac852 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2029.529573] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b57020a-f43d-4d28-bd99-c3ad6756762d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2029.538930] env[62824]: DEBUG oslo_vmware.api [None req-9d78c884-a693-423a-b835-e548c8af80b6 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Task: {'id': task-2146038, 'name': ReconfigVM_Task, 'duration_secs': 0.20233} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2029.542397] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d78c884-a693-423a-b835-e548c8af80b6 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] [instance: b96d1351-f3a7-4bac-998c-a34ab2606041] Reconfigured VM instance instance-0000001f to detach disk 2000 {{(pid=62824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 2029.545460] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ba134c72-e6e8-477e-b969-a4ffb2b232c2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2029.556760] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5579fc28-2318-4239-a4f4-9970e6d2f549 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2029.561982] env[62824]: DEBUG nova.compute.manager [None req-76aebf6c-e1c2-4914-b5ed-170d40366a5a tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2029.563287] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a19c7010-1e40-4d1a-a9da-e8108544436d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2029.567656] env[62824]: DEBUG oslo_vmware.api [None req-9d78c884-a693-423a-b835-e548c8af80b6 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Waiting for the task: (returnval){ [ 2029.567656] env[62824]: value = "task-2146041" [ 2029.567656] env[62824]: _type = "Task" [ 2029.567656] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2029.582634] env[62824]: DEBUG nova.compute.provider_tree [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2029.595024] env[62824]: DEBUG oslo_vmware.api [None req-9d78c884-a693-423a-b835-e548c8af80b6 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Task: {'id': task-2146041, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2029.700034] env[62824]: DEBUG oslo_vmware.api [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52bd6ba6-4659-060b-8c2c-b8f238461690, 'name': SearchDatastore_Task, 'duration_secs': 0.009112} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2029.700284] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2029.700548] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 02405b17-7022-4e2c-a357-415de88f63d0/02405b17-7022-4e2c-a357-415de88f63d0.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2029.700816] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-10d91cef-bfab-4441-9bdb-33c35cf4c075 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2029.708544] env[62824]: DEBUG oslo_vmware.api [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for the task: (returnval){ [ 2029.708544] env[62824]: value = "task-2146042" [ 2029.708544] env[62824]: _type = "Task" [ 2029.708544] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2029.717716] env[62824]: DEBUG oslo_vmware.api [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146042, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2029.726880] env[62824]: DEBUG oslo_vmware.api [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Task: {'id': task-2146039, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2029.770686] env[62824]: DEBUG nova.compute.manager [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] [instance: 4423d172-acdd-4c69-b3b8-ff166e1b8548] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2029.802895] env[62824]: DEBUG oslo_concurrency.lockutils [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Releasing lock "refresh_cache-c65ccd20-b682-420a-9c1a-47a45959197e" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2029.803352] env[62824]: DEBUG nova.compute.manager [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: c65ccd20-b682-420a-9c1a-47a45959197e] Instance network_info: |[{"id": "344e044b-aab1-485b-bd89-5477bc394098", "address": "fa:16:3e:4d:bf:ba", "network": {"id": "9e1885b6-0796-42b9-9cd8-8fa9ce9a3cfd", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-2118936609-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "64042a790d6c459186c68d73da32c019", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "afae6acd-1873-4228-9d5a-1cd5d4efe3e4", "external-id": "nsx-vlan-transportzone-183", "segmentation_id": 183, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap344e044b-aa", "ovs_interfaceid": "344e044b-aab1-485b-bd89-5477bc394098", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2029.803984] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: c65ccd20-b682-420a-9c1a-47a45959197e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4d:bf:ba', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'afae6acd-1873-4228-9d5a-1cd5d4efe3e4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '344e044b-aab1-485b-bd89-5477bc394098', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2029.811724] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Creating folder: Project (64042a790d6c459186c68d73da32c019). Parent ref: group-v438503. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 2029.816029] env[62824]: DEBUG nova.virt.hardware [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2029.816029] env[62824]: DEBUG nova.virt.hardware [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2029.816029] env[62824]: DEBUG nova.virt.hardware [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2029.816029] env[62824]: DEBUG nova.virt.hardware [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2029.816029] env[62824]: DEBUG nova.virt.hardware [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2029.816029] env[62824]: DEBUG nova.virt.hardware [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2029.816715] env[62824]: DEBUG nova.virt.hardware [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2029.816715] env[62824]: DEBUG nova.virt.hardware [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2029.816715] env[62824]: DEBUG nova.virt.hardware [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2029.816715] env[62824]: DEBUG nova.virt.hardware [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2029.817206] env[62824]: DEBUG nova.virt.hardware [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2029.817555] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f54261a7-6489-4b2d-aa3f-b010b79cb3ae {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2029.820279] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00058735-5776-48c3-947b-caf544d9b53d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2029.828735] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19d980eb-a0bc-457e-9a49-560933c6a95b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2029.834368] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Created folder: Project (64042a790d6c459186c68d73da32c019) in parent group-v438503. [ 2029.834565] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Creating folder: Instances. Parent ref: group-v438622. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 2029.835195] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-dcba612f-24bd-4ab2-9b8f-6b98f470fa78 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2029.853146] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Created folder: Instances in parent group-v438622. [ 2029.853146] env[62824]: DEBUG oslo.service.loopingcall [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2029.853146] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c65ccd20-b682-420a-9c1a-47a45959197e] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2029.853146] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f52e09d4-99fb-44f9-8d74-bd0147872cee {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2029.875528] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2029.875528] env[62824]: value = "task-2146045" [ 2029.875528] env[62824]: _type = "Task" [ 2029.875528] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2029.883846] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146045, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2030.086701] env[62824]: DEBUG oslo_vmware.api [None req-9d78c884-a693-423a-b835-e548c8af80b6 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Task: {'id': task-2146041, 'name': ReconfigVM_Task, 'duration_secs': 0.232451} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2030.086701] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d78c884-a693-423a-b835-e548c8af80b6 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] [instance: b96d1351-f3a7-4bac-998c-a34ab2606041] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-438510', 'volume_id': '64941ab3-102e-48a5-a3ad-433fb818850b', 'name': 'volume-64941ab3-102e-48a5-a3ad-433fb818850b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'b96d1351-f3a7-4bac-998c-a34ab2606041', 'attached_at': '', 'detached_at': '', 'volume_id': '64941ab3-102e-48a5-a3ad-433fb818850b', 'serial': '64941ab3-102e-48a5-a3ad-433fb818850b'} {{(pid=62824) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 2030.089633] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-9d78c884-a693-423a-b835-e548c8af80b6 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] [instance: b96d1351-f3a7-4bac-998c-a34ab2606041] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2030.092749] env[62824]: DEBUG nova.compute.manager [req-711c3fa2-143e-4c82-9594-7ff686d032cc req-beea736a-084d-42ae-a271-d69d4a092152 service nova] [instance: c65ccd20-b682-420a-9c1a-47a45959197e] Received event network-changed-344e044b-aab1-485b-bd89-5477bc394098 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2030.092749] env[62824]: DEBUG nova.compute.manager [req-711c3fa2-143e-4c82-9594-7ff686d032cc req-beea736a-084d-42ae-a271-d69d4a092152 service nova] [instance: c65ccd20-b682-420a-9c1a-47a45959197e] Refreshing instance network info cache due to event network-changed-344e044b-aab1-485b-bd89-5477bc394098. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2030.092749] env[62824]: DEBUG oslo_concurrency.lockutils [req-711c3fa2-143e-4c82-9594-7ff686d032cc req-beea736a-084d-42ae-a271-d69d4a092152 service nova] Acquiring lock "refresh_cache-c65ccd20-b682-420a-9c1a-47a45959197e" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2030.093232] env[62824]: DEBUG oslo_concurrency.lockutils [req-711c3fa2-143e-4c82-9594-7ff686d032cc req-beea736a-084d-42ae-a271-d69d4a092152 service nova] Acquired lock "refresh_cache-c65ccd20-b682-420a-9c1a-47a45959197e" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2030.093232] env[62824]: DEBUG nova.network.neutron [req-711c3fa2-143e-4c82-9594-7ff686d032cc req-beea736a-084d-42ae-a271-d69d4a092152 service nova] [instance: c65ccd20-b682-420a-9c1a-47a45959197e] Refreshing network info cache for port 344e044b-aab1-485b-bd89-5477bc394098 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2030.098039] env[62824]: DEBUG nova.scheduler.client.report [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2030.100158] env[62824]: INFO nova.compute.manager [None req-76aebf6c-e1c2-4914-b5ed-170d40366a5a tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] instance snapshotting [ 2030.102998] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0d5217b-1865-4426-9b78-db172d87bfd7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2030.110384] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5db9553c-69a8-4b43-b13d-55e9479cecc5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2030.118709] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-9d78c884-a693-423a-b835-e548c8af80b6 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] [instance: b96d1351-f3a7-4bac-998c-a34ab2606041] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2030.133652] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6ebf2cb9-7331-4bbf-87c7-62331c3a8f1b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2030.139119] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a590ba0f-47d9-45ca-9fe9-39e48113d297 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2030.232896] env[62824]: DEBUG oslo_vmware.api [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146042, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2030.240419] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-9d78c884-a693-423a-b835-e548c8af80b6 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] [instance: b96d1351-f3a7-4bac-998c-a34ab2606041] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2030.240419] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-9d78c884-a693-423a-b835-e548c8af80b6 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] [instance: b96d1351-f3a7-4bac-998c-a34ab2606041] Deleting contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2030.240419] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-9d78c884-a693-423a-b835-e548c8af80b6 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Deleting the datastore file [datastore2] b96d1351-f3a7-4bac-998c-a34ab2606041 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2030.240739] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-55d9d670-dbad-44d9-ade4-98cfcee4ce73 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2030.242611] env[62824]: DEBUG oslo_vmware.api [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Task: {'id': task-2146039, 'name': PowerOnVM_Task, 'duration_secs': 0.517124} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2030.242892] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2030.243153] env[62824]: INFO nova.compute.manager [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] Took 8.72 seconds to spawn the instance on the hypervisor. [ 2030.243350] env[62824]: DEBUG nova.compute.manager [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2030.244742] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7d99364-96d6-4206-a1fd-6c96a5ed126a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2030.252714] env[62824]: DEBUG oslo_vmware.api [None req-9d78c884-a693-423a-b835-e548c8af80b6 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Waiting for the task: (returnval){ [ 2030.252714] env[62824]: value = "task-2146047" [ 2030.252714] env[62824]: _type = "Task" [ 2030.252714] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2030.269959] env[62824]: DEBUG oslo_vmware.api [None req-9d78c884-a693-423a-b835-e548c8af80b6 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Task: {'id': task-2146047, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2030.386959] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146045, 'name': CreateVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2030.428363] env[62824]: DEBUG nova.network.neutron [req-711c3fa2-143e-4c82-9594-7ff686d032cc req-beea736a-084d-42ae-a271-d69d4a092152 service nova] [instance: c65ccd20-b682-420a-9c1a-47a45959197e] Updated VIF entry in instance network info cache for port 344e044b-aab1-485b-bd89-5477bc394098. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2030.429840] env[62824]: DEBUG nova.network.neutron [req-711c3fa2-143e-4c82-9594-7ff686d032cc req-beea736a-084d-42ae-a271-d69d4a092152 service nova] [instance: c65ccd20-b682-420a-9c1a-47a45959197e] Updating instance_info_cache with network_info: [{"id": "344e044b-aab1-485b-bd89-5477bc394098", "address": "fa:16:3e:4d:bf:ba", "network": {"id": "9e1885b6-0796-42b9-9cd8-8fa9ce9a3cfd", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-2118936609-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "64042a790d6c459186c68d73da32c019", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "afae6acd-1873-4228-9d5a-1cd5d4efe3e4", "external-id": "nsx-vlan-transportzone-183", "segmentation_id": 183, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap344e044b-aa", "ovs_interfaceid": "344e044b-aab1-485b-bd89-5477bc394098", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2030.609747] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.874s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2030.610289] env[62824]: DEBUG nova.compute.manager [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 88bad2dd-dce7-41df-b56c-93a5d054c11e] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2030.613726] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.049s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2030.615183] env[62824]: INFO nova.compute.claims [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] [instance: c63b05b7-1dbf-4637-8728-33673be0d3ea] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2030.653506] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-76aebf6c-e1c2-4914-b5ed-170d40366a5a tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Creating Snapshot of the VM instance {{(pid=62824) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 2030.653785] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-e20718a8-3e31-4c3b-90f3-8864defe5aa1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2030.662208] env[62824]: DEBUG oslo_vmware.api [None req-76aebf6c-e1c2-4914-b5ed-170d40366a5a tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Waiting for the task: (returnval){ [ 2030.662208] env[62824]: value = "task-2146048" [ 2030.662208] env[62824]: _type = "Task" [ 2030.662208] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2030.675683] env[62824]: DEBUG oslo_vmware.api [None req-76aebf6c-e1c2-4914-b5ed-170d40366a5a tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Task: {'id': task-2146048, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2030.724059] env[62824]: DEBUG oslo_vmware.api [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146042, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.629874} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2030.724291] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 02405b17-7022-4e2c-a357-415de88f63d0/02405b17-7022-4e2c-a357-415de88f63d0.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2030.724561] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02405b17-7022-4e2c-a357-415de88f63d0] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2030.724841] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9359c9e1-5396-4e98-bce1-6ff1920c0ced {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2030.732211] env[62824]: DEBUG oslo_vmware.api [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for the task: (returnval){ [ 2030.732211] env[62824]: value = "task-2146049" [ 2030.732211] env[62824]: _type = "Task" [ 2030.732211] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2030.742407] env[62824]: DEBUG oslo_vmware.api [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146049, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2030.763352] env[62824]: DEBUG oslo_vmware.api [None req-9d78c884-a693-423a-b835-e548c8af80b6 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Task: {'id': task-2146047, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.164591} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2030.763663] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-9d78c884-a693-423a-b835-e548c8af80b6 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2030.763842] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-9d78c884-a693-423a-b835-e548c8af80b6 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] [instance: b96d1351-f3a7-4bac-998c-a34ab2606041] Deleted contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2030.764049] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-9d78c884-a693-423a-b835-e548c8af80b6 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] [instance: b96d1351-f3a7-4bac-998c-a34ab2606041] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2030.764237] env[62824]: INFO nova.compute.manager [None req-9d78c884-a693-423a-b835-e548c8af80b6 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] [instance: b96d1351-f3a7-4bac-998c-a34ab2606041] Took 2.42 seconds to destroy the instance on the hypervisor. [ 2030.764513] env[62824]: DEBUG oslo.service.loopingcall [None req-9d78c884-a693-423a-b835-e548c8af80b6 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2030.764674] env[62824]: DEBUG nova.compute.manager [-] [instance: b96d1351-f3a7-4bac-998c-a34ab2606041] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2030.764769] env[62824]: DEBUG nova.network.neutron [-] [instance: b96d1351-f3a7-4bac-998c-a34ab2606041] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2030.776617] env[62824]: INFO nova.compute.manager [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] Took 45.15 seconds to build instance. [ 2030.777749] env[62824]: DEBUG oslo_concurrency.lockutils [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquiring lock "a6cd2032-de60-4f78-bf1e-79801d049df0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2030.777961] env[62824]: DEBUG oslo_concurrency.lockutils [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "a6cd2032-de60-4f78-bf1e-79801d049df0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2030.894522] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146045, 'name': CreateVM_Task, 'duration_secs': 0.589118} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2030.894893] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c65ccd20-b682-420a-9c1a-47a45959197e] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2030.896997] env[62824]: DEBUG oslo_concurrency.lockutils [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2030.899016] env[62824]: DEBUG oslo_concurrency.lockutils [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2030.899016] env[62824]: DEBUG oslo_concurrency.lockutils [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2030.899016] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fee6cb6e-3a43-4673-8006-ea7c16bab1fd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2030.903683] env[62824]: DEBUG oslo_vmware.api [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Waiting for the task: (returnval){ [ 2030.903683] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]521d0e13-85cc-99e1-07aa-4436b7a4ad54" [ 2030.903683] env[62824]: _type = "Task" [ 2030.903683] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2030.914798] env[62824]: DEBUG oslo_vmware.api [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]521d0e13-85cc-99e1-07aa-4436b7a4ad54, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2030.932480] env[62824]: DEBUG oslo_concurrency.lockutils [req-711c3fa2-143e-4c82-9594-7ff686d032cc req-beea736a-084d-42ae-a271-d69d4a092152 service nova] Releasing lock "refresh_cache-c65ccd20-b682-420a-9c1a-47a45959197e" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2031.121148] env[62824]: DEBUG nova.compute.utils [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2031.124616] env[62824]: DEBUG nova.compute.manager [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 88bad2dd-dce7-41df-b56c-93a5d054c11e] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2031.124812] env[62824]: DEBUG nova.network.neutron [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 88bad2dd-dce7-41df-b56c-93a5d054c11e] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2031.175952] env[62824]: DEBUG oslo_vmware.api [None req-76aebf6c-e1c2-4914-b5ed-170d40366a5a tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Task: {'id': task-2146048, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2031.244700] env[62824]: DEBUG oslo_vmware.api [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146049, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.08559} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2031.245043] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02405b17-7022-4e2c-a357-415de88f63d0] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2031.245885] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64591f78-36a6-4181-add7-bccb424ab85b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2031.279253] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02405b17-7022-4e2c-a357-415de88f63d0] Reconfiguring VM instance instance-00000030 to attach disk [datastore1] 02405b17-7022-4e2c-a357-415de88f63d0/02405b17-7022-4e2c-a357-415de88f63d0.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2031.281153] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1182089f-b88b-43da-bbac-1433ba108cf8 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Lock "b9f30f87-3594-4468-9d29-70890d8761e3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 115.105s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2031.281874] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-857ebf7a-9498-4d50-9853-6db4b364553f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2031.311510] env[62824]: DEBUG oslo_vmware.api [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for the task: (returnval){ [ 2031.311510] env[62824]: value = "task-2146050" [ 2031.311510] env[62824]: _type = "Task" [ 2031.311510] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2031.323272] env[62824]: DEBUG oslo_vmware.api [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146050, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2031.353236] env[62824]: DEBUG nova.policy [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4c914343569a462d9b8564f3376c2f54', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '10ee5fe93d584725ace90e76605a1b2e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2031.418426] env[62824]: DEBUG oslo_vmware.api [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]521d0e13-85cc-99e1-07aa-4436b7a4ad54, 'name': SearchDatastore_Task, 'duration_secs': 0.013682} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2031.418838] env[62824]: DEBUG oslo_concurrency.lockutils [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2031.419145] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: c65ccd20-b682-420a-9c1a-47a45959197e] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2031.419456] env[62824]: DEBUG oslo_concurrency.lockutils [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2031.419689] env[62824]: DEBUG oslo_concurrency.lockutils [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2031.419928] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2031.420430] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-811178ae-87af-4b5c-af77-635eed1bc208 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2031.432713] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2031.432974] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2031.434234] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-52fd8fe2-af15-49b1-8c99-049f9a7a59db {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2031.443025] env[62824]: DEBUG oslo_vmware.api [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Waiting for the task: (returnval){ [ 2031.443025] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52fd22b6-1264-3cbe-86e0-c2c3ec46d6bf" [ 2031.443025] env[62824]: _type = "Task" [ 2031.443025] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2031.453694] env[62824]: DEBUG oslo_vmware.api [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52fd22b6-1264-3cbe-86e0-c2c3ec46d6bf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2031.628517] env[62824]: DEBUG oslo_vmware.rw_handles [None req-b4894b47-a9c8-4b49-af68-93eace3f9a8c tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c54850-614d-97c0-8298-0d56ca98ffd3/disk-0.vmdk. {{(pid=62824) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 2031.629622] env[62824]: DEBUG nova.compute.manager [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 88bad2dd-dce7-41df-b56c-93a5d054c11e] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2031.639358] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4141cb3f-d4e5-480e-ae4d-b9b789d7771d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2031.648168] env[62824]: DEBUG oslo_vmware.rw_handles [None req-b4894b47-a9c8-4b49-af68-93eace3f9a8c tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c54850-614d-97c0-8298-0d56ca98ffd3/disk-0.vmdk is in state: ready. {{(pid=62824) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 2031.648168] env[62824]: ERROR oslo_vmware.rw_handles [None req-b4894b47-a9c8-4b49-af68-93eace3f9a8c tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c54850-614d-97c0-8298-0d56ca98ffd3/disk-0.vmdk due to incomplete transfer. [ 2031.648323] env[62824]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-b033e77b-3c66-4fa3-92cd-03202977e88b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2031.666725] env[62824]: DEBUG oslo_vmware.rw_handles [None req-b4894b47-a9c8-4b49-af68-93eace3f9a8c tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c54850-614d-97c0-8298-0d56ca98ffd3/disk-0.vmdk. {{(pid=62824) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 2031.667414] env[62824]: DEBUG nova.virt.vmwareapi.images [None req-b4894b47-a9c8-4b49-af68-93eace3f9a8c tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2] Uploaded image 47f7084b-090d-4f6a-b175-4c8a7ded78e5 to the Glance image server {{(pid=62824) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 2031.670032] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4894b47-a9c8-4b49-af68-93eace3f9a8c tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2] Destroying the VM {{(pid=62824) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 2031.673998] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-9e5d9a76-f026-4eb6-bc76-4836aebb3592 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2031.689014] env[62824]: DEBUG oslo_vmware.api [None req-76aebf6c-e1c2-4914-b5ed-170d40366a5a tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Task: {'id': task-2146048, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2031.690670] env[62824]: DEBUG oslo_vmware.api [None req-b4894b47-a9c8-4b49-af68-93eace3f9a8c tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Waiting for the task: (returnval){ [ 2031.690670] env[62824]: value = "task-2146051" [ 2031.690670] env[62824]: _type = "Task" [ 2031.690670] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2031.700974] env[62824]: DEBUG oslo_vmware.api [None req-b4894b47-a9c8-4b49-af68-93eace3f9a8c tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2146051, 'name': Destroy_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2031.800816] env[62824]: DEBUG nova.network.neutron [-] [instance: b96d1351-f3a7-4bac-998c-a34ab2606041] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2031.803315] env[62824]: DEBUG nova.compute.manager [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2031.828183] env[62824]: DEBUG oslo_vmware.api [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146050, 'name': ReconfigVM_Task, 'duration_secs': 0.334123} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2031.828183] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02405b17-7022-4e2c-a357-415de88f63d0] Reconfigured VM instance instance-00000030 to attach disk [datastore1] 02405b17-7022-4e2c-a357-415de88f63d0/02405b17-7022-4e2c-a357-415de88f63d0.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2031.828517] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1e3ecb48-a3f9-4ea9-b49a-7e489275feb7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2031.840489] env[62824]: DEBUG oslo_vmware.api [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for the task: (returnval){ [ 2031.840489] env[62824]: value = "task-2146052" [ 2031.840489] env[62824]: _type = "Task" [ 2031.840489] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2031.851515] env[62824]: DEBUG oslo_vmware.api [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146052, 'name': Rename_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2031.873557] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-1beb7be1-384e-40ae-965a-5630a0d7138d tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Volume attach. Driver type: vmdk {{(pid=62824) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 2031.873557] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-1beb7be1-384e-40ae-965a-5630a0d7138d tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-438620', 'volume_id': '5c4381e7-4dfa-4683-93fe-8eb8bef1df7e', 'name': 'volume-5c4381e7-4dfa-4683-93fe-8eb8bef1df7e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '59b5b883-4188-471c-8862-444f3ce08cb0', 'attached_at': '', 'detached_at': '', 'volume_id': '5c4381e7-4dfa-4683-93fe-8eb8bef1df7e', 'serial': '5c4381e7-4dfa-4683-93fe-8eb8bef1df7e'} {{(pid=62824) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 2031.874758] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d07d0fd3-e9f3-423e-a789-cf06080af773 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2031.903117] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-495c80e6-ea53-4f43-9c63-7ed8009d1703 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2031.935163] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-1beb7be1-384e-40ae-965a-5630a0d7138d tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Reconfiguring VM instance instance-00000022 to attach disk [datastore1] volume-5c4381e7-4dfa-4683-93fe-8eb8bef1df7e/volume-5c4381e7-4dfa-4683-93fe-8eb8bef1df7e.vmdk or device None with type thin {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2031.938856] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a6e06878-8a42-4aef-8d6e-485e9a3f6ebb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2031.963447] env[62824]: DEBUG oslo_vmware.api [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52fd22b6-1264-3cbe-86e0-c2c3ec46d6bf, 'name': SearchDatastore_Task, 'duration_secs': 0.017349} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2031.965804] env[62824]: DEBUG oslo_vmware.api [None req-1beb7be1-384e-40ae-965a-5630a0d7138d tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Waiting for the task: (returnval){ [ 2031.965804] env[62824]: value = "task-2146053" [ 2031.965804] env[62824]: _type = "Task" [ 2031.965804] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2031.965804] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-79d6d8ce-f320-4b0e-95e9-fffe9f4c47cd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2031.980302] env[62824]: DEBUG oslo_vmware.api [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Waiting for the task: (returnval){ [ 2031.980302] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5235e650-459a-26f6-49c0-a66e3c44f65c" [ 2031.980302] env[62824]: _type = "Task" [ 2031.980302] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2031.985030] env[62824]: DEBUG oslo_vmware.api [None req-1beb7be1-384e-40ae-965a-5630a0d7138d tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Task: {'id': task-2146053, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2031.994786] env[62824]: DEBUG oslo_vmware.api [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5235e650-459a-26f6-49c0-a66e3c44f65c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2032.052381] env[62824]: DEBUG nova.compute.manager [req-97f43c4a-2f34-4fe2-934b-12961548c212 req-1d20e2ad-f9ef-42cc-aac8-7a8653bcd2aa service nova] [instance: 4423d172-acdd-4c69-b3b8-ff166e1b8548] Received event network-vif-plugged-0f1c4fc4-1d12-4800-9861-68137c05d00f {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2032.052627] env[62824]: DEBUG oslo_concurrency.lockutils [req-97f43c4a-2f34-4fe2-934b-12961548c212 req-1d20e2ad-f9ef-42cc-aac8-7a8653bcd2aa service nova] Acquiring lock "4423d172-acdd-4c69-b3b8-ff166e1b8548-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2032.052847] env[62824]: DEBUG oslo_concurrency.lockutils [req-97f43c4a-2f34-4fe2-934b-12961548c212 req-1d20e2ad-f9ef-42cc-aac8-7a8653bcd2aa service nova] Lock "4423d172-acdd-4c69-b3b8-ff166e1b8548-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2032.053248] env[62824]: DEBUG oslo_concurrency.lockutils [req-97f43c4a-2f34-4fe2-934b-12961548c212 req-1d20e2ad-f9ef-42cc-aac8-7a8653bcd2aa service nova] Lock "4423d172-acdd-4c69-b3b8-ff166e1b8548-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2032.053248] env[62824]: DEBUG nova.compute.manager [req-97f43c4a-2f34-4fe2-934b-12961548c212 req-1d20e2ad-f9ef-42cc-aac8-7a8653bcd2aa service nova] [instance: 4423d172-acdd-4c69-b3b8-ff166e1b8548] No waiting events found dispatching network-vif-plugged-0f1c4fc4-1d12-4800-9861-68137c05d00f {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2032.053416] env[62824]: WARNING nova.compute.manager [req-97f43c4a-2f34-4fe2-934b-12961548c212 req-1d20e2ad-f9ef-42cc-aac8-7a8653bcd2aa service nova] [instance: 4423d172-acdd-4c69-b3b8-ff166e1b8548] Received unexpected event network-vif-plugged-0f1c4fc4-1d12-4800-9861-68137c05d00f for instance with vm_state building and task_state spawning. [ 2032.177392] env[62824]: DEBUG oslo_vmware.api [None req-76aebf6c-e1c2-4914-b5ed-170d40366a5a tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Task: {'id': task-2146048, 'name': CreateSnapshot_Task, 'duration_secs': 1.23013} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2032.180036] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-76aebf6c-e1c2-4914-b5ed-170d40366a5a tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Created Snapshot of the VM instance {{(pid=62824) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 2032.181648] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ff30200-7900-4863-a85d-fcaf77648356 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2032.207575] env[62824]: DEBUG oslo_vmware.api [None req-b4894b47-a9c8-4b49-af68-93eace3f9a8c tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2146051, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2032.295273] env[62824]: DEBUG nova.network.neutron [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] [instance: 4423d172-acdd-4c69-b3b8-ff166e1b8548] Successfully updated port: 0f1c4fc4-1d12-4800-9861-68137c05d00f {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2032.303335] env[62824]: INFO nova.compute.manager [-] [instance: b96d1351-f3a7-4bac-998c-a34ab2606041] Took 1.54 seconds to deallocate network for instance. [ 2032.332446] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2032.350607] env[62824]: DEBUG oslo_vmware.api [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146052, 'name': Rename_Task, 'duration_secs': 0.194133} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2032.350828] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02405b17-7022-4e2c-a357-415de88f63d0] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2032.351099] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dea2d81a-ecfd-42a0-8da8-043fbba38818 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2032.361283] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5560590d-9ebe-4f5e-b06e-59eb72fb9202 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2032.365678] env[62824]: DEBUG oslo_vmware.api [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for the task: (returnval){ [ 2032.365678] env[62824]: value = "task-2146054" [ 2032.365678] env[62824]: _type = "Task" [ 2032.365678] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2032.374663] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5c6a536-7743-42bc-955e-c2a810636830 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2032.377846] env[62824]: DEBUG oslo_vmware.api [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146054, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2032.407497] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-266d85de-5e50-48f6-a2df-443a4e0f3b58 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2032.416788] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04d4a055-4187-4abb-869c-d01efd1dc3fe {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2032.433491] env[62824]: DEBUG nova.compute.provider_tree [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2032.478997] env[62824]: DEBUG oslo_vmware.api [None req-1beb7be1-384e-40ae-965a-5630a0d7138d tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Task: {'id': task-2146053, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2032.494237] env[62824]: DEBUG oslo_vmware.api [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5235e650-459a-26f6-49c0-a66e3c44f65c, 'name': SearchDatastore_Task, 'duration_secs': 0.016109} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2032.494514] env[62824]: DEBUG oslo_concurrency.lockutils [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2032.494896] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] c65ccd20-b682-420a-9c1a-47a45959197e/c65ccd20-b682-420a-9c1a-47a45959197e.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2032.495101] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-525e0694-0893-4ddd-ba29-011f9fabdaa4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2032.504982] env[62824]: DEBUG oslo_vmware.api [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Waiting for the task: (returnval){ [ 2032.504982] env[62824]: value = "task-2146055" [ 2032.504982] env[62824]: _type = "Task" [ 2032.504982] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2032.514648] env[62824]: DEBUG oslo_vmware.api [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146055, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2032.572750] env[62824]: DEBUG nova.network.neutron [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 88bad2dd-dce7-41df-b56c-93a5d054c11e] Successfully created port: aaeb6e9d-ce3d-4e4c-b2a1-319b383fdf1d {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2032.590181] env[62824]: DEBUG nova.compute.manager [req-60fcf173-acd0-45b4-a9a3-328db2680882 req-eb421b8a-20f0-43a7-ba52-78ea59022cfe service nova] [instance: b96d1351-f3a7-4bac-998c-a34ab2606041] Received event network-vif-deleted-c7dfce05-2238-418a-b414-395b1f1fb4f6 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2032.590181] env[62824]: DEBUG nova.compute.manager [req-60fcf173-acd0-45b4-a9a3-328db2680882 req-eb421b8a-20f0-43a7-ba52-78ea59022cfe service nova] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] Received event network-changed-2d7ab792-8a41-4c8b-91a4-e4a677466a12 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2032.590913] env[62824]: DEBUG nova.compute.manager [req-60fcf173-acd0-45b4-a9a3-328db2680882 req-eb421b8a-20f0-43a7-ba52-78ea59022cfe service nova] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] Refreshing instance network info cache due to event network-changed-2d7ab792-8a41-4c8b-91a4-e4a677466a12. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2032.591353] env[62824]: DEBUG oslo_concurrency.lockutils [req-60fcf173-acd0-45b4-a9a3-328db2680882 req-eb421b8a-20f0-43a7-ba52-78ea59022cfe service nova] Acquiring lock "refresh_cache-b9f30f87-3594-4468-9d29-70890d8761e3" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2032.591740] env[62824]: DEBUG oslo_concurrency.lockutils [req-60fcf173-acd0-45b4-a9a3-328db2680882 req-eb421b8a-20f0-43a7-ba52-78ea59022cfe service nova] Acquired lock "refresh_cache-b9f30f87-3594-4468-9d29-70890d8761e3" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2032.592111] env[62824]: DEBUG nova.network.neutron [req-60fcf173-acd0-45b4-a9a3-328db2680882 req-eb421b8a-20f0-43a7-ba52-78ea59022cfe service nova] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] Refreshing network info cache for port 2d7ab792-8a41-4c8b-91a4-e4a677466a12 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2032.652584] env[62824]: DEBUG nova.compute.manager [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 88bad2dd-dce7-41df-b56c-93a5d054c11e] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2032.681590] env[62824]: DEBUG nova.virt.hardware [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2032.682118] env[62824]: DEBUG nova.virt.hardware [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2032.682319] env[62824]: DEBUG nova.virt.hardware [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2032.682517] env[62824]: DEBUG nova.virt.hardware [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2032.682666] env[62824]: DEBUG nova.virt.hardware [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2032.682811] env[62824]: DEBUG nova.virt.hardware [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2032.683058] env[62824]: DEBUG nova.virt.hardware [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2032.683278] env[62824]: DEBUG nova.virt.hardware [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2032.683814] env[62824]: DEBUG nova.virt.hardware [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2032.683814] env[62824]: DEBUG nova.virt.hardware [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2032.683814] env[62824]: DEBUG nova.virt.hardware [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2032.684697] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-038cb79b-e2b9-4d8b-9286-0f058b61a2a7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2032.693990] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5e941fa-f656-4847-9769-87255cf52528 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2032.705685] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-76aebf6c-e1c2-4914-b5ed-170d40366a5a tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Creating linked-clone VM from snapshot {{(pid=62824) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 2032.709789] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-1569a4b7-7c53-4dfc-b98d-7c5e740d47c5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2032.720260] env[62824]: DEBUG oslo_vmware.api [None req-b4894b47-a9c8-4b49-af68-93eace3f9a8c tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2146051, 'name': Destroy_Task, 'duration_secs': 0.723568} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2032.731682] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-b4894b47-a9c8-4b49-af68-93eace3f9a8c tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2] Destroyed the VM [ 2032.731805] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-b4894b47-a9c8-4b49-af68-93eace3f9a8c tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2] Deleting Snapshot of the VM instance {{(pid=62824) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 2032.732808] env[62824]: DEBUG oslo_vmware.api [None req-76aebf6c-e1c2-4914-b5ed-170d40366a5a tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Waiting for the task: (returnval){ [ 2032.732808] env[62824]: value = "task-2146056" [ 2032.732808] env[62824]: _type = "Task" [ 2032.732808] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2032.733087] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-edb5eeef-2bc8-4afa-8ace-c81ed96c4318 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2032.746009] env[62824]: DEBUG oslo_vmware.api [None req-76aebf6c-e1c2-4914-b5ed-170d40366a5a tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Task: {'id': task-2146056, 'name': CloneVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2032.747534] env[62824]: DEBUG oslo_vmware.api [None req-b4894b47-a9c8-4b49-af68-93eace3f9a8c tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Waiting for the task: (returnval){ [ 2032.747534] env[62824]: value = "task-2146057" [ 2032.747534] env[62824]: _type = "Task" [ 2032.747534] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2032.757225] env[62824]: DEBUG oslo_vmware.api [None req-b4894b47-a9c8-4b49-af68-93eace3f9a8c tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2146057, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2032.798463] env[62824]: DEBUG oslo_concurrency.lockutils [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Acquiring lock "refresh_cache-4423d172-acdd-4c69-b3b8-ff166e1b8548" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2032.798660] env[62824]: DEBUG oslo_concurrency.lockutils [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Acquired lock "refresh_cache-4423d172-acdd-4c69-b3b8-ff166e1b8548" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2032.798820] env[62824]: DEBUG nova.network.neutron [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] [instance: 4423d172-acdd-4c69-b3b8-ff166e1b8548] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2032.880677] env[62824]: DEBUG oslo_vmware.api [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146054, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2032.910030] env[62824]: INFO nova.compute.manager [None req-9d78c884-a693-423a-b835-e548c8af80b6 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] [instance: b96d1351-f3a7-4bac-998c-a34ab2606041] Took 0.60 seconds to detach 1 volumes for instance. [ 2032.914147] env[62824]: DEBUG nova.compute.manager [None req-9d78c884-a693-423a-b835-e548c8af80b6 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] [instance: b96d1351-f3a7-4bac-998c-a34ab2606041] Deleting volume: 64941ab3-102e-48a5-a3ad-433fb818850b {{(pid=62824) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3282}} [ 2032.943493] env[62824]: DEBUG nova.scheduler.client.report [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2032.985042] env[62824]: DEBUG oslo_vmware.api [None req-1beb7be1-384e-40ae-965a-5630a0d7138d tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Task: {'id': task-2146053, 'name': ReconfigVM_Task, 'duration_secs': 0.537724} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2032.985042] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-1beb7be1-384e-40ae-965a-5630a0d7138d tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Reconfigured VM instance instance-00000022 to attach disk [datastore1] volume-5c4381e7-4dfa-4683-93fe-8eb8bef1df7e/volume-5c4381e7-4dfa-4683-93fe-8eb8bef1df7e.vmdk or device None with type thin {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2032.992352] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-01d60492-b8ac-4928-b6cb-671fc307ea0c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2033.039354] env[62824]: DEBUG oslo_vmware.api [None req-1beb7be1-384e-40ae-965a-5630a0d7138d tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Waiting for the task: (returnval){ [ 2033.039354] env[62824]: value = "task-2146058" [ 2033.039354] env[62824]: _type = "Task" [ 2033.039354] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2033.043115] env[62824]: DEBUG oslo_vmware.api [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146055, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2033.052580] env[62824]: DEBUG oslo_vmware.api [None req-1beb7be1-384e-40ae-965a-5630a0d7138d tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Task: {'id': task-2146058, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2033.247597] env[62824]: DEBUG oslo_vmware.api [None req-76aebf6c-e1c2-4914-b5ed-170d40366a5a tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Task: {'id': task-2146056, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2033.258271] env[62824]: DEBUG oslo_vmware.api [None req-b4894b47-a9c8-4b49-af68-93eace3f9a8c tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2146057, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2033.346504] env[62824]: DEBUG nova.network.neutron [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] [instance: 4423d172-acdd-4c69-b3b8-ff166e1b8548] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2033.380853] env[62824]: DEBUG oslo_vmware.api [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146054, 'name': PowerOnVM_Task, 'duration_secs': 0.665032} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2033.381171] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02405b17-7022-4e2c-a357-415de88f63d0] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2033.381379] env[62824]: INFO nova.compute.manager [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02405b17-7022-4e2c-a357-415de88f63d0] Took 9.18 seconds to spawn the instance on the hypervisor. [ 2033.381562] env[62824]: DEBUG nova.compute.manager [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02405b17-7022-4e2c-a357-415de88f63d0] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2033.382371] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f61b686-2930-42b5-94e0-a3058b1cb349 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2033.452238] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.838s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2033.452492] env[62824]: DEBUG nova.compute.manager [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] [instance: c63b05b7-1dbf-4637-8728-33673be0d3ea] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2033.457938] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.236s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2033.460214] env[62824]: INFO nova.compute.claims [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 6a724332-a165-4a2b-9dd8-85f27e7b7637] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2033.519801] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9d78c884-a693-423a-b835-e548c8af80b6 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2033.531604] env[62824]: DEBUG oslo_vmware.api [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146055, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.667887} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2033.532363] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] c65ccd20-b682-420a-9c1a-47a45959197e/c65ccd20-b682-420a-9c1a-47a45959197e.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2033.532636] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: c65ccd20-b682-420a-9c1a-47a45959197e] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2033.535292] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f0e8b5c4-4167-4744-99af-c8c0782ff6f4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2033.561255] env[62824]: DEBUG oslo_vmware.api [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Waiting for the task: (returnval){ [ 2033.561255] env[62824]: value = "task-2146060" [ 2033.561255] env[62824]: _type = "Task" [ 2033.561255] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2033.570451] env[62824]: DEBUG oslo_vmware.api [None req-1beb7be1-384e-40ae-965a-5630a0d7138d tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Task: {'id': task-2146058, 'name': ReconfigVM_Task, 'duration_secs': 0.221303} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2033.571104] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-1beb7be1-384e-40ae-965a-5630a0d7138d tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-438620', 'volume_id': '5c4381e7-4dfa-4683-93fe-8eb8bef1df7e', 'name': 'volume-5c4381e7-4dfa-4683-93fe-8eb8bef1df7e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '59b5b883-4188-471c-8862-444f3ce08cb0', 'attached_at': '', 'detached_at': '', 'volume_id': '5c4381e7-4dfa-4683-93fe-8eb8bef1df7e', 'serial': '5c4381e7-4dfa-4683-93fe-8eb8bef1df7e'} {{(pid=62824) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 2033.577593] env[62824]: DEBUG oslo_vmware.api [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146060, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2033.625362] env[62824]: DEBUG nova.network.neutron [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] [instance: 4423d172-acdd-4c69-b3b8-ff166e1b8548] Updating instance_info_cache with network_info: [{"id": "0f1c4fc4-1d12-4800-9861-68137c05d00f", "address": "fa:16:3e:22:b3:9a", "network": {"id": "b6403dc3-30ef-4c76-b521-380c179f345d", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.218", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "22c5c40c2e004babaa15896c89307303", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1002b79b-224e-41e3-a484-4245a767147a", "external-id": "nsx-vlan-transportzone-353", "segmentation_id": 353, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0f1c4fc4-1d", "ovs_interfaceid": "0f1c4fc4-1d12-4800-9861-68137c05d00f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2033.726529] env[62824]: DEBUG nova.network.neutron [req-60fcf173-acd0-45b4-a9a3-328db2680882 req-eb421b8a-20f0-43a7-ba52-78ea59022cfe service nova] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] Updated VIF entry in instance network info cache for port 2d7ab792-8a41-4c8b-91a4-e4a677466a12. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2033.726910] env[62824]: DEBUG nova.network.neutron [req-60fcf173-acd0-45b4-a9a3-328db2680882 req-eb421b8a-20f0-43a7-ba52-78ea59022cfe service nova] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] Updating instance_info_cache with network_info: [{"id": "2d7ab792-8a41-4c8b-91a4-e4a677466a12", "address": "fa:16:3e:20:e4:39", "network": {"id": "d299eecf-1628-4fce-bc67-f9f038d261cd", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1701982809-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.189", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d180bbdc8e72414081363fe696ead6fb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae4e3171-21cd-4094-b6cf-81bf366c75bd", "external-id": "nsx-vlan-transportzone-193", "segmentation_id": 193, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2d7ab792-8a", "ovs_interfaceid": "2d7ab792-8a41-4c8b-91a4-e4a677466a12", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2033.748023] env[62824]: DEBUG oslo_vmware.api [None req-76aebf6c-e1c2-4914-b5ed-170d40366a5a tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Task: {'id': task-2146056, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2033.764342] env[62824]: DEBUG oslo_vmware.api [None req-b4894b47-a9c8-4b49-af68-93eace3f9a8c tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2146057, 'name': RemoveSnapshot_Task, 'duration_secs': 0.603883} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2033.764635] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-b4894b47-a9c8-4b49-af68-93eace3f9a8c tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2] Deleted Snapshot of the VM instance {{(pid=62824) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 2033.764922] env[62824]: INFO nova.compute.manager [None req-b4894b47-a9c8-4b49-af68-93eace3f9a8c tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2] Took 14.96 seconds to snapshot the instance on the hypervisor. [ 2033.901814] env[62824]: INFO nova.compute.manager [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02405b17-7022-4e2c-a357-415de88f63d0] Took 47.20 seconds to build instance. [ 2033.973812] env[62824]: DEBUG nova.compute.utils [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2033.978077] env[62824]: DEBUG nova.compute.manager [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] [instance: c63b05b7-1dbf-4637-8728-33673be0d3ea] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2033.978077] env[62824]: DEBUG nova.network.neutron [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] [instance: c63b05b7-1dbf-4637-8728-33673be0d3ea] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2034.043158] env[62824]: DEBUG nova.policy [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd4c2f6e1e8ea42a8bce5d4096e3868a9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8face4a61e49498eae3a9920f9a91a1f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2034.078121] env[62824]: DEBUG oslo_vmware.api [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146060, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.095162} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2034.079209] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: c65ccd20-b682-420a-9c1a-47a45959197e] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2034.079899] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cf3cbe7-e704-48f5-b5bd-eea14b6c06a0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2034.113968] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: c65ccd20-b682-420a-9c1a-47a45959197e] Reconfiguring VM instance instance-00000031 to attach disk [datastore1] c65ccd20-b682-420a-9c1a-47a45959197e/c65ccd20-b682-420a-9c1a-47a45959197e.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2034.113968] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d716b78d-be1c-4887-8043-25c97c1b01c6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2034.133936] env[62824]: DEBUG oslo_concurrency.lockutils [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Releasing lock "refresh_cache-4423d172-acdd-4c69-b3b8-ff166e1b8548" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2034.134258] env[62824]: DEBUG nova.compute.manager [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] [instance: 4423d172-acdd-4c69-b3b8-ff166e1b8548] Instance network_info: |[{"id": "0f1c4fc4-1d12-4800-9861-68137c05d00f", "address": "fa:16:3e:22:b3:9a", "network": {"id": "b6403dc3-30ef-4c76-b521-380c179f345d", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.218", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "22c5c40c2e004babaa15896c89307303", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1002b79b-224e-41e3-a484-4245a767147a", "external-id": "nsx-vlan-transportzone-353", "segmentation_id": 353, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0f1c4fc4-1d", "ovs_interfaceid": "0f1c4fc4-1d12-4800-9861-68137c05d00f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2034.135621] env[62824]: DEBUG nova.compute.manager [req-3ce47ba8-177e-4ada-9668-7239f4867a15 req-e97f4756-baa5-425b-ab72-3d7b8c409f4c service nova] [instance: 4423d172-acdd-4c69-b3b8-ff166e1b8548] Received event network-changed-0f1c4fc4-1d12-4800-9861-68137c05d00f {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2034.135797] env[62824]: DEBUG nova.compute.manager [req-3ce47ba8-177e-4ada-9668-7239f4867a15 req-e97f4756-baa5-425b-ab72-3d7b8c409f4c service nova] [instance: 4423d172-acdd-4c69-b3b8-ff166e1b8548] Refreshing instance network info cache due to event network-changed-0f1c4fc4-1d12-4800-9861-68137c05d00f. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2034.136017] env[62824]: DEBUG oslo_concurrency.lockutils [req-3ce47ba8-177e-4ada-9668-7239f4867a15 req-e97f4756-baa5-425b-ab72-3d7b8c409f4c service nova] Acquiring lock "refresh_cache-4423d172-acdd-4c69-b3b8-ff166e1b8548" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2034.136162] env[62824]: DEBUG oslo_concurrency.lockutils [req-3ce47ba8-177e-4ada-9668-7239f4867a15 req-e97f4756-baa5-425b-ab72-3d7b8c409f4c service nova] Acquired lock "refresh_cache-4423d172-acdd-4c69-b3b8-ff166e1b8548" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2034.136368] env[62824]: DEBUG nova.network.neutron [req-3ce47ba8-177e-4ada-9668-7239f4867a15 req-e97f4756-baa5-425b-ab72-3d7b8c409f4c service nova] [instance: 4423d172-acdd-4c69-b3b8-ff166e1b8548] Refreshing network info cache for port 0f1c4fc4-1d12-4800-9861-68137c05d00f {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2034.138300] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] [instance: 4423d172-acdd-4c69-b3b8-ff166e1b8548] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:22:b3:9a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1002b79b-224e-41e3-a484-4245a767147a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0f1c4fc4-1d12-4800-9861-68137c05d00f', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2034.146327] env[62824]: DEBUG oslo.service.loopingcall [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2034.149420] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4423d172-acdd-4c69-b3b8-ff166e1b8548] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2034.150049] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-53bb630a-cb0b-4458-a9fe-0af33dcb7f26 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2034.167467] env[62824]: DEBUG oslo_vmware.api [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Waiting for the task: (returnval){ [ 2034.167467] env[62824]: value = "task-2146061" [ 2034.167467] env[62824]: _type = "Task" [ 2034.167467] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2034.176492] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2034.176492] env[62824]: value = "task-2146062" [ 2034.176492] env[62824]: _type = "Task" [ 2034.176492] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2034.180193] env[62824]: DEBUG oslo_vmware.api [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146061, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2034.189275] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146062, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2034.230240] env[62824]: DEBUG oslo_concurrency.lockutils [req-60fcf173-acd0-45b4-a9a3-328db2680882 req-eb421b8a-20f0-43a7-ba52-78ea59022cfe service nova] Releasing lock "refresh_cache-b9f30f87-3594-4468-9d29-70890d8761e3" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2034.247893] env[62824]: DEBUG oslo_vmware.api [None req-76aebf6c-e1c2-4914-b5ed-170d40366a5a tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Task: {'id': task-2146056, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2034.408575] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b4efc757-b5d5-4ab5-b783-5cae18a9feca tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Lock "02405b17-7022-4e2c-a357-415de88f63d0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 108.622s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2034.478703] env[62824]: DEBUG nova.compute.manager [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] [instance: c63b05b7-1dbf-4637-8728-33673be0d3ea] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2034.654968] env[62824]: DEBUG nova.objects.instance [None req-1beb7be1-384e-40ae-965a-5630a0d7138d tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Lazy-loading 'flavor' on Instance uuid 59b5b883-4188-471c-8862-444f3ce08cb0 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2034.691189] env[62824]: DEBUG oslo_vmware.api [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146061, 'name': ReconfigVM_Task, 'duration_secs': 0.331219} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2034.692214] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: c65ccd20-b682-420a-9c1a-47a45959197e] Reconfigured VM instance instance-00000031 to attach disk [datastore1] c65ccd20-b682-420a-9c1a-47a45959197e/c65ccd20-b682-420a-9c1a-47a45959197e.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2034.693034] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-05524fb1-e93a-453a-8473-833d3f56b28d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2034.699036] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146062, 'name': CreateVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2034.708691] env[62824]: DEBUG oslo_vmware.api [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Waiting for the task: (returnval){ [ 2034.708691] env[62824]: value = "task-2146063" [ 2034.708691] env[62824]: _type = "Task" [ 2034.708691] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2034.730600] env[62824]: DEBUG oslo_vmware.api [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146063, 'name': Rename_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2034.749547] env[62824]: DEBUG oslo_vmware.api [None req-76aebf6c-e1c2-4914-b5ed-170d40366a5a tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Task: {'id': task-2146056, 'name': CloneVM_Task} progress is 95%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2034.844308] env[62824]: DEBUG nova.network.neutron [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] [instance: c63b05b7-1dbf-4637-8728-33673be0d3ea] Successfully created port: a0edea0b-d2cc-4b03-9d46-9cc5e6cd1794 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2034.910475] env[62824]: DEBUG nova.compute.manager [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2034.947211] env[62824]: DEBUG nova.network.neutron [req-3ce47ba8-177e-4ada-9668-7239f4867a15 req-e97f4756-baa5-425b-ab72-3d7b8c409f4c service nova] [instance: 4423d172-acdd-4c69-b3b8-ff166e1b8548] Updated VIF entry in instance network info cache for port 0f1c4fc4-1d12-4800-9861-68137c05d00f. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2034.947730] env[62824]: DEBUG nova.network.neutron [req-3ce47ba8-177e-4ada-9668-7239f4867a15 req-e97f4756-baa5-425b-ab72-3d7b8c409f4c service nova] [instance: 4423d172-acdd-4c69-b3b8-ff166e1b8548] Updating instance_info_cache with network_info: [{"id": "0f1c4fc4-1d12-4800-9861-68137c05d00f", "address": "fa:16:3e:22:b3:9a", "network": {"id": "b6403dc3-30ef-4c76-b521-380c179f345d", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.218", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "22c5c40c2e004babaa15896c89307303", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1002b79b-224e-41e3-a484-4245a767147a", "external-id": "nsx-vlan-transportzone-353", "segmentation_id": 353, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0f1c4fc4-1d", "ovs_interfaceid": "0f1c4fc4-1d12-4800-9861-68137c05d00f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2035.160872] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1beb7be1-384e-40ae-965a-5630a0d7138d tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Lock "59b5b883-4188-471c-8862-444f3ce08cb0" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.415s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2035.195688] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146062, 'name': CreateVM_Task, 'duration_secs': 0.599813} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2035.196930] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4423d172-acdd-4c69-b3b8-ff166e1b8548] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2035.197735] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-882c9689-f778-454a-9e29-fcc1e0abcf74 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2035.201877] env[62824]: DEBUG oslo_concurrency.lockutils [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2035.202090] env[62824]: DEBUG oslo_concurrency.lockutils [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2035.202436] env[62824]: DEBUG oslo_concurrency.lockutils [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2035.202930] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f51f1a2f-17af-4c54-9018-97bc9db6e6db {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2035.208926] env[62824]: DEBUG oslo_vmware.api [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Waiting for the task: (returnval){ [ 2035.208926] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5287a2ee-da6c-f3ca-bdea-f6593fa10152" [ 2035.208926] env[62824]: _type = "Task" [ 2035.208926] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2035.215201] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5892e1b8-2544-4dd2-9ef2-78f7e0b5e137 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2035.228739] env[62824]: DEBUG oslo_vmware.api [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146063, 'name': Rename_Task, 'duration_secs': 0.174288} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2035.256247] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: c65ccd20-b682-420a-9c1a-47a45959197e] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2035.256744] env[62824]: DEBUG oslo_vmware.api [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5287a2ee-da6c-f3ca-bdea-f6593fa10152, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2035.257159] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-918734ad-67c5-4071-adf8-ec4161abee01 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2035.262065] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4d9e127-acb8-4698-b511-4032a021406b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2035.271242] env[62824]: DEBUG oslo_vmware.api [None req-76aebf6c-e1c2-4914-b5ed-170d40366a5a tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Task: {'id': task-2146056, 'name': CloneVM_Task, 'duration_secs': 2.398581} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2035.273959] env[62824]: INFO nova.virt.vmwareapi.vmops [None req-76aebf6c-e1c2-4914-b5ed-170d40366a5a tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Created linked-clone VM from snapshot [ 2035.274274] env[62824]: DEBUG oslo_vmware.api [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Waiting for the task: (returnval){ [ 2035.274274] env[62824]: value = "task-2146064" [ 2035.274274] env[62824]: _type = "Task" [ 2035.274274] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2035.274989] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44f4cbf3-6145-406a-936e-fa24e99048f5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2035.278931] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-624dbef7-22c2-43ab-9012-f102761f2668 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2035.292607] env[62824]: DEBUG nova.virt.vmwareapi.images [None req-76aebf6c-e1c2-4914-b5ed-170d40366a5a tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Uploading image 45e8c4ae-efb5-498b-869d-2abeee747d8b {{(pid=62824) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 2035.305910] env[62824]: DEBUG nova.compute.provider_tree [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2035.307319] env[62824]: DEBUG oslo_vmware.api [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146064, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2035.339303] env[62824]: DEBUG oslo_vmware.rw_handles [None req-76aebf6c-e1c2-4914-b5ed-170d40366a5a tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 2035.339303] env[62824]: value = "vm-438626" [ 2035.339303] env[62824]: _type = "VirtualMachine" [ 2035.339303] env[62824]: }. {{(pid=62824) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 2035.339534] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-9a332ea4-b0f9-4cee-ad28-19392f857cd3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2035.347968] env[62824]: DEBUG oslo_vmware.rw_handles [None req-76aebf6c-e1c2-4914-b5ed-170d40366a5a tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Lease: (returnval){ [ 2035.347968] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52e235bb-2224-3578-5058-b63e0a82f762" [ 2035.347968] env[62824]: _type = "HttpNfcLease" [ 2035.347968] env[62824]: } obtained for exporting VM: (result){ [ 2035.347968] env[62824]: value = "vm-438626" [ 2035.347968] env[62824]: _type = "VirtualMachine" [ 2035.347968] env[62824]: }. {{(pid=62824) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 2035.348442] env[62824]: DEBUG oslo_vmware.api [None req-76aebf6c-e1c2-4914-b5ed-170d40366a5a tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Waiting for the lease: (returnval){ [ 2035.348442] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52e235bb-2224-3578-5058-b63e0a82f762" [ 2035.348442] env[62824]: _type = "HttpNfcLease" [ 2035.348442] env[62824]: } to be ready. {{(pid=62824) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 2035.356100] env[62824]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 2035.356100] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52e235bb-2224-3578-5058-b63e0a82f762" [ 2035.356100] env[62824]: _type = "HttpNfcLease" [ 2035.356100] env[62824]: } is initializing. {{(pid=62824) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 2035.441705] env[62824]: DEBUG oslo_concurrency.lockutils [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2035.445573] env[62824]: DEBUG nova.network.neutron [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 88bad2dd-dce7-41df-b56c-93a5d054c11e] Successfully updated port: aaeb6e9d-ce3d-4e4c-b2a1-319b383fdf1d {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2035.450052] env[62824]: DEBUG oslo_concurrency.lockutils [req-3ce47ba8-177e-4ada-9668-7239f4867a15 req-e97f4756-baa5-425b-ab72-3d7b8c409f4c service nova] Releasing lock "refresh_cache-4423d172-acdd-4c69-b3b8-ff166e1b8548" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2035.480147] env[62824]: DEBUG nova.compute.manager [None req-457a3e46-ac18-4c02-91aa-31023cbb56f4 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2035.481301] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1848eeb4-16b9-4eff-a330-ff29a9908de8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2035.492491] env[62824]: DEBUG nova.compute.manager [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] [instance: c63b05b7-1dbf-4637-8728-33673be0d3ea] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2035.533097] env[62824]: DEBUG nova.virt.hardware [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2035.533373] env[62824]: DEBUG nova.virt.hardware [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2035.533529] env[62824]: DEBUG nova.virt.hardware [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2035.533708] env[62824]: DEBUG nova.virt.hardware [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2035.533853] env[62824]: DEBUG nova.virt.hardware [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2035.533998] env[62824]: DEBUG nova.virt.hardware [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2035.534576] env[62824]: DEBUG nova.virt.hardware [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2035.534757] env[62824]: DEBUG nova.virt.hardware [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2035.534934] env[62824]: DEBUG nova.virt.hardware [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2035.535141] env[62824]: DEBUG nova.virt.hardware [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2035.535389] env[62824]: DEBUG nova.virt.hardware [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2035.536669] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0da0625-4287-4a2e-a79d-5e5f9d354810 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2035.553342] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2801b11-fb67-41a0-843e-e1ff2ccfc9cf {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2035.727259] env[62824]: DEBUG oslo_vmware.api [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5287a2ee-da6c-f3ca-bdea-f6593fa10152, 'name': SearchDatastore_Task, 'duration_secs': 0.026579} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2035.727259] env[62824]: DEBUG oslo_concurrency.lockutils [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2035.727259] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] [instance: 4423d172-acdd-4c69-b3b8-ff166e1b8548] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2035.727259] env[62824]: DEBUG oslo_concurrency.lockutils [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2035.727259] env[62824]: DEBUG oslo_concurrency.lockutils [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2035.727494] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2035.727733] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5ee6208c-000a-49e2-970d-c1875a3a4acc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2035.738873] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2035.739082] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2035.739939] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-093f05d4-d43e-4a58-b968-d1be299c1f3d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2035.750059] env[62824]: DEBUG oslo_vmware.api [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Waiting for the task: (returnval){ [ 2035.750059] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52a50649-1157-a337-9f5d-d1ffb078d027" [ 2035.750059] env[62824]: _type = "Task" [ 2035.750059] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2035.759402] env[62824]: DEBUG oslo_vmware.api [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52a50649-1157-a337-9f5d-d1ffb078d027, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2035.794265] env[62824]: DEBUG oslo_vmware.api [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146064, 'name': PowerOnVM_Task, 'duration_secs': 0.474534} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2035.794604] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: c65ccd20-b682-420a-9c1a-47a45959197e] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2035.794782] env[62824]: INFO nova.compute.manager [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: c65ccd20-b682-420a-9c1a-47a45959197e] Took 8.81 seconds to spawn the instance on the hypervisor. [ 2035.795060] env[62824]: DEBUG nova.compute.manager [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: c65ccd20-b682-420a-9c1a-47a45959197e] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2035.795912] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d31616a0-d8e8-47a7-977d-fd38d02ce74d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2035.808647] env[62824]: DEBUG nova.scheduler.client.report [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2035.859567] env[62824]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 2035.859567] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52e235bb-2224-3578-5058-b63e0a82f762" [ 2035.859567] env[62824]: _type = "HttpNfcLease" [ 2035.859567] env[62824]: } is ready. {{(pid=62824) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 2035.859567] env[62824]: DEBUG oslo_vmware.rw_handles [None req-76aebf6c-e1c2-4914-b5ed-170d40366a5a tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 2035.859567] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52e235bb-2224-3578-5058-b63e0a82f762" [ 2035.859567] env[62824]: _type = "HttpNfcLease" [ 2035.859567] env[62824]: }. {{(pid=62824) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 2035.860977] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6daa9909-5f24-4dd8-86f3-ddb4fe9c8a1e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2035.870252] env[62824]: DEBUG oslo_vmware.rw_handles [None req-76aebf6c-e1c2-4914-b5ed-170d40366a5a tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52645093-a57f-8239-89a9-271e4e104280/disk-0.vmdk from lease info. {{(pid=62824) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 2035.870474] env[62824]: DEBUG oslo_vmware.rw_handles [None req-76aebf6c-e1c2-4914-b5ed-170d40366a5a tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52645093-a57f-8239-89a9-271e4e104280/disk-0.vmdk for reading. {{(pid=62824) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 2035.948286] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Acquiring lock "refresh_cache-88bad2dd-dce7-41df-b56c-93a5d054c11e" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2035.948727] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Acquired lock "refresh_cache-88bad2dd-dce7-41df-b56c-93a5d054c11e" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2035.948964] env[62824]: DEBUG nova.network.neutron [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 88bad2dd-dce7-41df-b56c-93a5d054c11e] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2035.954257] env[62824]: DEBUG oslo_concurrency.lockutils [None req-aad7d4de-1683-40a8-81f1-d9d3d7dbf6e4 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Acquiring lock "59b5b883-4188-471c-8862-444f3ce08cb0" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2035.954257] env[62824]: DEBUG oslo_concurrency.lockutils [None req-aad7d4de-1683-40a8-81f1-d9d3d7dbf6e4 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Lock "59b5b883-4188-471c-8862-444f3ce08cb0" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2035.954257] env[62824]: DEBUG nova.compute.manager [None req-aad7d4de-1683-40a8-81f1-d9d3d7dbf6e4 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2035.955054] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3feb1c2e-400b-46da-9214-2cf9a8bed653 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2035.962781] env[62824]: DEBUG nova.compute.manager [None req-aad7d4de-1683-40a8-81f1-d9d3d7dbf6e4 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62824) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3404}} [ 2035.964051] env[62824]: DEBUG nova.objects.instance [None req-aad7d4de-1683-40a8-81f1-d9d3d7dbf6e4 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Lazy-loading 'flavor' on Instance uuid 59b5b883-4188-471c-8862-444f3ce08cb0 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2036.005028] env[62824]: INFO nova.compute.manager [None req-457a3e46-ac18-4c02-91aa-31023cbb56f4 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] instance snapshotting [ 2036.007446] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-266160cb-c01b-4c2d-b10a-54f7ab29e267 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2036.032793] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d9e1a87-80a5-44b1-8dd8-59dc2b83fd33 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2036.036521] env[62824]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-fd360ff2-7a5f-41fb-80c5-4aa28881fa8a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2036.166032] env[62824]: DEBUG oslo_concurrency.lockutils [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Acquiring lock "02f34565-6675-4c79-ac47-b131ceba9df8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2036.166253] env[62824]: DEBUG oslo_concurrency.lockutils [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Lock "02f34565-6675-4c79-ac47-b131ceba9df8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2036.262023] env[62824]: DEBUG oslo_vmware.api [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52a50649-1157-a337-9f5d-d1ffb078d027, 'name': SearchDatastore_Task, 'duration_secs': 0.022048} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2036.265260] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f7af6a2b-c0a9-47d2-b9b5-b25dacfb6892 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2036.269497] env[62824]: DEBUG nova.compute.manager [req-ad01f8c5-3f62-48a8-8d14-c7b7b35fa754 req-ee9ec9c6-b88b-47b6-a481-882beba44a1d service nova] [instance: 88bad2dd-dce7-41df-b56c-93a5d054c11e] Received event network-vif-plugged-aaeb6e9d-ce3d-4e4c-b2a1-319b383fdf1d {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2036.269887] env[62824]: DEBUG oslo_concurrency.lockutils [req-ad01f8c5-3f62-48a8-8d14-c7b7b35fa754 req-ee9ec9c6-b88b-47b6-a481-882beba44a1d service nova] Acquiring lock "88bad2dd-dce7-41df-b56c-93a5d054c11e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2036.270214] env[62824]: DEBUG oslo_concurrency.lockutils [req-ad01f8c5-3f62-48a8-8d14-c7b7b35fa754 req-ee9ec9c6-b88b-47b6-a481-882beba44a1d service nova] Lock "88bad2dd-dce7-41df-b56c-93a5d054c11e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2036.270502] env[62824]: DEBUG oslo_concurrency.lockutils [req-ad01f8c5-3f62-48a8-8d14-c7b7b35fa754 req-ee9ec9c6-b88b-47b6-a481-882beba44a1d service nova] Lock "88bad2dd-dce7-41df-b56c-93a5d054c11e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2036.270773] env[62824]: DEBUG nova.compute.manager [req-ad01f8c5-3f62-48a8-8d14-c7b7b35fa754 req-ee9ec9c6-b88b-47b6-a481-882beba44a1d service nova] [instance: 88bad2dd-dce7-41df-b56c-93a5d054c11e] No waiting events found dispatching network-vif-plugged-aaeb6e9d-ce3d-4e4c-b2a1-319b383fdf1d {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2036.271055] env[62824]: WARNING nova.compute.manager [req-ad01f8c5-3f62-48a8-8d14-c7b7b35fa754 req-ee9ec9c6-b88b-47b6-a481-882beba44a1d service nova] [instance: 88bad2dd-dce7-41df-b56c-93a5d054c11e] Received unexpected event network-vif-plugged-aaeb6e9d-ce3d-4e4c-b2a1-319b383fdf1d for instance with vm_state building and task_state spawning. [ 2036.271300] env[62824]: DEBUG nova.compute.manager [req-ad01f8c5-3f62-48a8-8d14-c7b7b35fa754 req-ee9ec9c6-b88b-47b6-a481-882beba44a1d service nova] [instance: 88bad2dd-dce7-41df-b56c-93a5d054c11e] Received event network-changed-aaeb6e9d-ce3d-4e4c-b2a1-319b383fdf1d {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2036.271497] env[62824]: DEBUG nova.compute.manager [req-ad01f8c5-3f62-48a8-8d14-c7b7b35fa754 req-ee9ec9c6-b88b-47b6-a481-882beba44a1d service nova] [instance: 88bad2dd-dce7-41df-b56c-93a5d054c11e] Refreshing instance network info cache due to event network-changed-aaeb6e9d-ce3d-4e4c-b2a1-319b383fdf1d. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2036.271751] env[62824]: DEBUG oslo_concurrency.lockutils [req-ad01f8c5-3f62-48a8-8d14-c7b7b35fa754 req-ee9ec9c6-b88b-47b6-a481-882beba44a1d service nova] Acquiring lock "refresh_cache-88bad2dd-dce7-41df-b56c-93a5d054c11e" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2036.277439] env[62824]: DEBUG oslo_vmware.api [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Waiting for the task: (returnval){ [ 2036.277439] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52f0d6b6-6981-5afd-7792-8da0f41b7f13" [ 2036.277439] env[62824]: _type = "Task" [ 2036.277439] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2036.287495] env[62824]: DEBUG oslo_vmware.api [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52f0d6b6-6981-5afd-7792-8da0f41b7f13, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2036.316756] env[62824]: INFO nova.compute.manager [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: c65ccd20-b682-420a-9c1a-47a45959197e] Took 46.23 seconds to build instance. [ 2036.318700] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.861s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2036.320131] env[62824]: DEBUG nova.compute.manager [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 6a724332-a165-4a2b-9dd8-85f27e7b7637] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2036.324280] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.501s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2036.326613] env[62824]: INFO nova.compute.claims [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 21c05725-3dad-4965-98d0-0622ebcc8ece] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2036.521047] env[62824]: DEBUG nova.network.neutron [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 88bad2dd-dce7-41df-b56c-93a5d054c11e] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2036.552206] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-457a3e46-ac18-4c02-91aa-31023cbb56f4 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Creating Snapshot of the VM instance {{(pid=62824) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 2036.552619] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-08f01665-9cd9-4a22-9460-ed46e80c6f55 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2036.567145] env[62824]: DEBUG oslo_vmware.api [None req-457a3e46-ac18-4c02-91aa-31023cbb56f4 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Waiting for the task: (returnval){ [ 2036.567145] env[62824]: value = "task-2146066" [ 2036.567145] env[62824]: _type = "Task" [ 2036.567145] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2036.580432] env[62824]: DEBUG oslo_vmware.api [None req-457a3e46-ac18-4c02-91aa-31023cbb56f4 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2146066, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2036.792261] env[62824]: DEBUG oslo_vmware.api [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52f0d6b6-6981-5afd-7792-8da0f41b7f13, 'name': SearchDatastore_Task, 'duration_secs': 0.015267} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2036.792918] env[62824]: DEBUG oslo_concurrency.lockutils [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2036.797234] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 4423d172-acdd-4c69-b3b8-ff166e1b8548/4423d172-acdd-4c69-b3b8-ff166e1b8548.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2036.797925] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-60088e43-f946-43c0-8c9d-0bc1eb26db7d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2036.809582] env[62824]: DEBUG oslo_vmware.api [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Waiting for the task: (returnval){ [ 2036.809582] env[62824]: value = "task-2146067" [ 2036.809582] env[62824]: _type = "Task" [ 2036.809582] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2036.822651] env[62824]: DEBUG oslo_vmware.api [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Task: {'id': task-2146067, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2036.825597] env[62824]: DEBUG oslo_concurrency.lockutils [None req-08d83dc8-12f7-4a27-b764-933409174c53 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Lock "c65ccd20-b682-420a-9c1a-47a45959197e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 101.134s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2036.838752] env[62824]: DEBUG nova.compute.utils [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2036.842732] env[62824]: DEBUG nova.compute.manager [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 6a724332-a165-4a2b-9dd8-85f27e7b7637] Not allocating networking since 'none' was specified. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1984}} [ 2036.950049] env[62824]: DEBUG nova.network.neutron [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 88bad2dd-dce7-41df-b56c-93a5d054c11e] Updating instance_info_cache with network_info: [{"id": "aaeb6e9d-ce3d-4e4c-b2a1-319b383fdf1d", "address": "fa:16:3e:64:9c:cc", "network": {"id": "aeacc588-ae73-49ec-9aa1-68f7d92055be", "bridge": "br-int", "label": "tempest-ImagesTestJSON-314094063-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10ee5fe93d584725ace90e76605a1b2e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1e7a4976-597e-4636-990e-6062b5faadee", "external-id": "nsx-vlan-transportzone-847", "segmentation_id": 847, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaaeb6e9d-ce", "ovs_interfaceid": "aaeb6e9d-ce3d-4e4c-b2a1-319b383fdf1d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2036.974847] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-aad7d4de-1683-40a8-81f1-d9d3d7dbf6e4 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2036.975654] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2af3d394-9cc2-4e2a-b44f-a1e4e27fa703 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2036.989231] env[62824]: DEBUG oslo_vmware.api [None req-aad7d4de-1683-40a8-81f1-d9d3d7dbf6e4 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Waiting for the task: (returnval){ [ 2036.989231] env[62824]: value = "task-2146068" [ 2036.989231] env[62824]: _type = "Task" [ 2036.989231] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2037.001199] env[62824]: DEBUG oslo_vmware.api [None req-aad7d4de-1683-40a8-81f1-d9d3d7dbf6e4 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Task: {'id': task-2146068, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2037.079712] env[62824]: DEBUG oslo_vmware.api [None req-457a3e46-ac18-4c02-91aa-31023cbb56f4 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2146066, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2037.327932] env[62824]: DEBUG nova.network.neutron [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] [instance: c63b05b7-1dbf-4637-8728-33673be0d3ea] Successfully updated port: a0edea0b-d2cc-4b03-9d46-9cc5e6cd1794 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2037.335019] env[62824]: DEBUG oslo_vmware.api [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Task: {'id': task-2146067, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2037.344698] env[62824]: DEBUG nova.compute.manager [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 919e8a45-7810-4a8d-a1aa-5046b5ab059c] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2037.351298] env[62824]: DEBUG nova.compute.manager [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 6a724332-a165-4a2b-9dd8-85f27e7b7637] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2037.457690] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Releasing lock "refresh_cache-88bad2dd-dce7-41df-b56c-93a5d054c11e" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2037.458186] env[62824]: DEBUG nova.compute.manager [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 88bad2dd-dce7-41df-b56c-93a5d054c11e] Instance network_info: |[{"id": "aaeb6e9d-ce3d-4e4c-b2a1-319b383fdf1d", "address": "fa:16:3e:64:9c:cc", "network": {"id": "aeacc588-ae73-49ec-9aa1-68f7d92055be", "bridge": "br-int", "label": "tempest-ImagesTestJSON-314094063-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10ee5fe93d584725ace90e76605a1b2e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1e7a4976-597e-4636-990e-6062b5faadee", "external-id": "nsx-vlan-transportzone-847", "segmentation_id": 847, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaaeb6e9d-ce", "ovs_interfaceid": "aaeb6e9d-ce3d-4e4c-b2a1-319b383fdf1d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2037.461676] env[62824]: DEBUG oslo_concurrency.lockutils [req-ad01f8c5-3f62-48a8-8d14-c7b7b35fa754 req-ee9ec9c6-b88b-47b6-a481-882beba44a1d service nova] Acquired lock "refresh_cache-88bad2dd-dce7-41df-b56c-93a5d054c11e" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2037.463080] env[62824]: DEBUG nova.network.neutron [req-ad01f8c5-3f62-48a8-8d14-c7b7b35fa754 req-ee9ec9c6-b88b-47b6-a481-882beba44a1d service nova] [instance: 88bad2dd-dce7-41df-b56c-93a5d054c11e] Refreshing network info cache for port aaeb6e9d-ce3d-4e4c-b2a1-319b383fdf1d {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2037.464168] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 88bad2dd-dce7-41df-b56c-93a5d054c11e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:64:9c:cc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1e7a4976-597e-4636-990e-6062b5faadee', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'aaeb6e9d-ce3d-4e4c-b2a1-319b383fdf1d', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2037.474675] env[62824]: DEBUG oslo.service.loopingcall [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2037.478515] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 88bad2dd-dce7-41df-b56c-93a5d054c11e] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2037.479865] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-de42b71f-23cf-4557-bbf9-66f51737bbe0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2037.515876] env[62824]: DEBUG oslo_vmware.api [None req-aad7d4de-1683-40a8-81f1-d9d3d7dbf6e4 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Task: {'id': task-2146068, 'name': PowerOffVM_Task, 'duration_secs': 0.426879} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2037.520482] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-aad7d4de-1683-40a8-81f1-d9d3d7dbf6e4 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2037.520717] env[62824]: DEBUG nova.compute.manager [None req-aad7d4de-1683-40a8-81f1-d9d3d7dbf6e4 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2037.520995] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2037.520995] env[62824]: value = "task-2146069" [ 2037.520995] env[62824]: _type = "Task" [ 2037.520995] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2037.522035] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-413b22bb-7597-400b-a66b-ea869d2b7ccb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2037.540983] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146069, 'name': CreateVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2037.590338] env[62824]: DEBUG oslo_vmware.api [None req-457a3e46-ac18-4c02-91aa-31023cbb56f4 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2146066, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2037.830738] env[62824]: DEBUG oslo_vmware.api [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Task: {'id': task-2146067, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.695956} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2037.830738] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 4423d172-acdd-4c69-b3b8-ff166e1b8548/4423d172-acdd-4c69-b3b8-ff166e1b8548.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2037.830738] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] [instance: 4423d172-acdd-4c69-b3b8-ff166e1b8548] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2037.830738] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8e8f33b2-e905-4518-a457-16bb65b2d49e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2037.836464] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Acquiring lock "refresh_cache-c63b05b7-1dbf-4637-8728-33673be0d3ea" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2037.837026] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Acquired lock "refresh_cache-c63b05b7-1dbf-4637-8728-33673be0d3ea" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2037.837351] env[62824]: DEBUG nova.network.neutron [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] [instance: c63b05b7-1dbf-4637-8728-33673be0d3ea] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2037.845251] env[62824]: DEBUG oslo_vmware.api [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Waiting for the task: (returnval){ [ 2037.845251] env[62824]: value = "task-2146070" [ 2037.845251] env[62824]: _type = "Task" [ 2037.845251] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2037.869152] env[62824]: DEBUG oslo_vmware.api [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Task: {'id': task-2146070, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2037.899071] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2037.965089] env[62824]: DEBUG nova.network.neutron [req-ad01f8c5-3f62-48a8-8d14-c7b7b35fa754 req-ee9ec9c6-b88b-47b6-a481-882beba44a1d service nova] [instance: 88bad2dd-dce7-41df-b56c-93a5d054c11e] Updated VIF entry in instance network info cache for port aaeb6e9d-ce3d-4e4c-b2a1-319b383fdf1d. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2037.965977] env[62824]: DEBUG nova.network.neutron [req-ad01f8c5-3f62-48a8-8d14-c7b7b35fa754 req-ee9ec9c6-b88b-47b6-a481-882beba44a1d service nova] [instance: 88bad2dd-dce7-41df-b56c-93a5d054c11e] Updating instance_info_cache with network_info: [{"id": "aaeb6e9d-ce3d-4e4c-b2a1-319b383fdf1d", "address": "fa:16:3e:64:9c:cc", "network": {"id": "aeacc588-ae73-49ec-9aa1-68f7d92055be", "bridge": "br-int", "label": "tempest-ImagesTestJSON-314094063-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10ee5fe93d584725ace90e76605a1b2e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1e7a4976-597e-4636-990e-6062b5faadee", "external-id": "nsx-vlan-transportzone-847", "segmentation_id": 847, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaaeb6e9d-ce", "ovs_interfaceid": "aaeb6e9d-ce3d-4e4c-b2a1-319b383fdf1d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2038.041892] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146069, 'name': CreateVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2038.048652] env[62824]: DEBUG oslo_concurrency.lockutils [None req-aad7d4de-1683-40a8-81f1-d9d3d7dbf6e4 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Lock "59b5b883-4188-471c-8862-444f3ce08cb0" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.093s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2038.080073] env[62824]: DEBUG oslo_vmware.api [None req-457a3e46-ac18-4c02-91aa-31023cbb56f4 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2146066, 'name': CreateSnapshot_Task, 'duration_secs': 1.156706} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2038.081034] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-457a3e46-ac18-4c02-91aa-31023cbb56f4 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Created Snapshot of the VM instance {{(pid=62824) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 2038.081554] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7d3943a-ab8f-4212-a650-9cb2a0fddfde {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2038.156204] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca61350c-8d90-42d9-954f-80701e1b0c6f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2038.172855] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4d30989-768e-45d4-8460-4a8e46e2ffb0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2038.246334] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7559d86-8717-44b1-b455-3dbcb8635013 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2038.256720] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8fd9141-5c9b-4701-9d04-27f0c3a79c51 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2038.275183] env[62824]: DEBUG nova.compute.provider_tree [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2038.360641] env[62824]: DEBUG oslo_vmware.api [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Task: {'id': task-2146070, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.137761} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2038.360641] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] [instance: 4423d172-acdd-4c69-b3b8-ff166e1b8548] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2038.361676] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-827b1a96-db4d-473b-9691-f77c047e28eb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2038.380058] env[62824]: DEBUG nova.compute.manager [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 6a724332-a165-4a2b-9dd8-85f27e7b7637] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2038.391081] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] [instance: 4423d172-acdd-4c69-b3b8-ff166e1b8548] Reconfiguring VM instance instance-00000032 to attach disk [datastore1] 4423d172-acdd-4c69-b3b8-ff166e1b8548/4423d172-acdd-4c69-b3b8-ff166e1b8548.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2038.391747] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-71ba7d2b-cf31-4d91-b620-bfae31f39d83 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2038.412191] env[62824]: DEBUG nova.network.neutron [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] [instance: c63b05b7-1dbf-4637-8728-33673be0d3ea] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2038.417564] env[62824]: DEBUG nova.virt.hardware [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2038.417869] env[62824]: DEBUG nova.virt.hardware [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2038.418107] env[62824]: DEBUG nova.virt.hardware [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2038.418318] env[62824]: DEBUG nova.virt.hardware [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2038.418469] env[62824]: DEBUG nova.virt.hardware [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2038.418621] env[62824]: DEBUG nova.virt.hardware [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2038.418829] env[62824]: DEBUG nova.virt.hardware [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2038.419049] env[62824]: DEBUG nova.virt.hardware [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2038.419268] env[62824]: DEBUG nova.virt.hardware [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2038.419448] env[62824]: DEBUG nova.virt.hardware [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2038.419672] env[62824]: DEBUG nova.virt.hardware [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2038.420976] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a20b81b-32f1-48f5-826c-2c4d5e9c77dc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2038.427039] env[62824]: DEBUG oslo_vmware.api [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Waiting for the task: (returnval){ [ 2038.427039] env[62824]: value = "task-2146071" [ 2038.427039] env[62824]: _type = "Task" [ 2038.427039] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2038.427787] env[62824]: DEBUG nova.compute.manager [req-a80d6b32-d524-4df5-8357-82b6065483f6 req-c3a1bb39-fbc8-4ee5-870f-3f51dc995b49 service nova] [instance: c63b05b7-1dbf-4637-8728-33673be0d3ea] Received event network-vif-plugged-a0edea0b-d2cc-4b03-9d46-9cc5e6cd1794 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2038.427991] env[62824]: DEBUG oslo_concurrency.lockutils [req-a80d6b32-d524-4df5-8357-82b6065483f6 req-c3a1bb39-fbc8-4ee5-870f-3f51dc995b49 service nova] Acquiring lock "c63b05b7-1dbf-4637-8728-33673be0d3ea-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2038.428256] env[62824]: DEBUG oslo_concurrency.lockutils [req-a80d6b32-d524-4df5-8357-82b6065483f6 req-c3a1bb39-fbc8-4ee5-870f-3f51dc995b49 service nova] Lock "c63b05b7-1dbf-4637-8728-33673be0d3ea-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2038.428466] env[62824]: DEBUG oslo_concurrency.lockutils [req-a80d6b32-d524-4df5-8357-82b6065483f6 req-c3a1bb39-fbc8-4ee5-870f-3f51dc995b49 service nova] Lock "c63b05b7-1dbf-4637-8728-33673be0d3ea-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2038.428649] env[62824]: DEBUG nova.compute.manager [req-a80d6b32-d524-4df5-8357-82b6065483f6 req-c3a1bb39-fbc8-4ee5-870f-3f51dc995b49 service nova] [instance: c63b05b7-1dbf-4637-8728-33673be0d3ea] No waiting events found dispatching network-vif-plugged-a0edea0b-d2cc-4b03-9d46-9cc5e6cd1794 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2038.428811] env[62824]: WARNING nova.compute.manager [req-a80d6b32-d524-4df5-8357-82b6065483f6 req-c3a1bb39-fbc8-4ee5-870f-3f51dc995b49 service nova] [instance: c63b05b7-1dbf-4637-8728-33673be0d3ea] Received unexpected event network-vif-plugged-a0edea0b-d2cc-4b03-9d46-9cc5e6cd1794 for instance with vm_state building and task_state spawning. [ 2038.428970] env[62824]: DEBUG nova.compute.manager [req-a80d6b32-d524-4df5-8357-82b6065483f6 req-c3a1bb39-fbc8-4ee5-870f-3f51dc995b49 service nova] [instance: c63b05b7-1dbf-4637-8728-33673be0d3ea] Received event network-changed-a0edea0b-d2cc-4b03-9d46-9cc5e6cd1794 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2038.429140] env[62824]: DEBUG nova.compute.manager [req-a80d6b32-d524-4df5-8357-82b6065483f6 req-c3a1bb39-fbc8-4ee5-870f-3f51dc995b49 service nova] [instance: c63b05b7-1dbf-4637-8728-33673be0d3ea] Refreshing instance network info cache due to event network-changed-a0edea0b-d2cc-4b03-9d46-9cc5e6cd1794. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2038.429316] env[62824]: DEBUG oslo_concurrency.lockutils [req-a80d6b32-d524-4df5-8357-82b6065483f6 req-c3a1bb39-fbc8-4ee5-870f-3f51dc995b49 service nova] Acquiring lock "refresh_cache-c63b05b7-1dbf-4637-8728-33673be0d3ea" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2038.442353] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7bacd25-73b1-49b3-86a9-6a794cca589e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2038.447915] env[62824]: DEBUG oslo_vmware.api [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Task: {'id': task-2146071, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2038.463362] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 6a724332-a165-4a2b-9dd8-85f27e7b7637] Instance VIF info [] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2038.469942] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Creating folder: Project (f5a12196000e4487997b00d6013a960f). Parent ref: group-v438503. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 2038.470697] env[62824]: DEBUG oslo_concurrency.lockutils [req-ad01f8c5-3f62-48a8-8d14-c7b7b35fa754 req-ee9ec9c6-b88b-47b6-a481-882beba44a1d service nova] Releasing lock "refresh_cache-88bad2dd-dce7-41df-b56c-93a5d054c11e" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2038.471161] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4b4c0cb1-04ea-4597-8b29-9d7e511a5250 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2038.485922] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Created folder: Project (f5a12196000e4487997b00d6013a960f) in parent group-v438503. [ 2038.486134] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Creating folder: Instances. Parent ref: group-v438630. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 2038.486438] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-03251979-248c-4697-9ddb-f4c35a9193f9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2038.499969] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Created folder: Instances in parent group-v438630. [ 2038.500426] env[62824]: DEBUG oslo.service.loopingcall [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2038.501072] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6a724332-a165-4a2b-9dd8-85f27e7b7637] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2038.501072] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4517150d-c2f2-43be-a576-319e276c5020 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2038.527118] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2038.527118] env[62824]: value = "task-2146074" [ 2038.527118] env[62824]: _type = "Task" [ 2038.527118] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2038.541197] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146074, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2038.548074] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146069, 'name': CreateVM_Task, 'duration_secs': 0.59575} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2038.548237] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 88bad2dd-dce7-41df-b56c-93a5d054c11e] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2038.549185] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2038.549438] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2038.550173] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2038.550616] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5e8bd834-3657-4952-92d5-7c2fab9e5f99 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2038.557434] env[62824]: DEBUG oslo_vmware.api [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the task: (returnval){ [ 2038.557434] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52e4b523-5108-f680-e550-e8b19b890305" [ 2038.557434] env[62824]: _type = "Task" [ 2038.557434] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2038.567354] env[62824]: DEBUG oslo_vmware.api [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52e4b523-5108-f680-e550-e8b19b890305, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2038.608603] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-457a3e46-ac18-4c02-91aa-31023cbb56f4 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Creating linked-clone VM from snapshot {{(pid=62824) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 2038.609033] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-3bf8b4d9-835d-43b4-9665-33821f173fb3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2038.620336] env[62824]: DEBUG oslo_vmware.api [None req-457a3e46-ac18-4c02-91aa-31023cbb56f4 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Waiting for the task: (returnval){ [ 2038.620336] env[62824]: value = "task-2146075" [ 2038.620336] env[62824]: _type = "Task" [ 2038.620336] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2038.631487] env[62824]: DEBUG oslo_vmware.api [None req-457a3e46-ac18-4c02-91aa-31023cbb56f4 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2146075, 'name': CloneVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2038.778587] env[62824]: DEBUG nova.scheduler.client.report [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2038.791666] env[62824]: DEBUG nova.network.neutron [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] [instance: c63b05b7-1dbf-4637-8728-33673be0d3ea] Updating instance_info_cache with network_info: [{"id": "a0edea0b-d2cc-4b03-9d46-9cc5e6cd1794", "address": "fa:16:3e:17:95:fb", "network": {"id": "eb8486ef-73d8-4ef6-a534-fbdae6bc7fad", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-628019626-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8face4a61e49498eae3a9920f9a91a1f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd96b39f-bd2e-48d1-85c3-577cf97f08c8", "external-id": "cl2-zone-84", "segmentation_id": 84, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa0edea0b-d2", "ovs_interfaceid": "a0edea0b-d2cc-4b03-9d46-9cc5e6cd1794", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2038.945804] env[62824]: DEBUG oslo_vmware.api [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Task: {'id': task-2146071, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2039.042247] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146074, 'name': CreateVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2039.069606] env[62824]: DEBUG oslo_vmware.api [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52e4b523-5108-f680-e550-e8b19b890305, 'name': SearchDatastore_Task, 'duration_secs': 0.02261} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2039.070125] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2039.070463] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 88bad2dd-dce7-41df-b56c-93a5d054c11e] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2039.070759] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2039.071011] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2039.071311] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2039.071711] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-be3ffbb4-98ec-474e-93ba-b1a9b729e801 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2039.083216] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2039.083460] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2039.084286] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bb763230-e651-401b-97ef-54b87756de95 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2039.091599] env[62824]: DEBUG oslo_vmware.api [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the task: (returnval){ [ 2039.091599] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]529b4bce-d3c3-3a31-def4-1b1a0111cdb2" [ 2039.091599] env[62824]: _type = "Task" [ 2039.091599] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2039.101594] env[62824]: DEBUG oslo_vmware.api [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]529b4bce-d3c3-3a31-def4-1b1a0111cdb2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2039.132602] env[62824]: DEBUG oslo_vmware.api [None req-457a3e46-ac18-4c02-91aa-31023cbb56f4 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2146075, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2039.285068] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.961s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2039.285698] env[62824]: DEBUG nova.compute.manager [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 21c05725-3dad-4965-98d0-0622ebcc8ece] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2039.288349] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f8aad8ab-6936-4c8d-8d26-3f40750d45f7 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.175s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2039.288573] env[62824]: DEBUG nova.objects.instance [None req-f8aad8ab-6936-4c8d-8d26-3f40750d45f7 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Lazy-loading 'resources' on Instance uuid 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2039.294210] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Releasing lock "refresh_cache-c63b05b7-1dbf-4637-8728-33673be0d3ea" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2039.294852] env[62824]: DEBUG nova.compute.manager [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] [instance: c63b05b7-1dbf-4637-8728-33673be0d3ea] Instance network_info: |[{"id": "a0edea0b-d2cc-4b03-9d46-9cc5e6cd1794", "address": "fa:16:3e:17:95:fb", "network": {"id": "eb8486ef-73d8-4ef6-a534-fbdae6bc7fad", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-628019626-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8face4a61e49498eae3a9920f9a91a1f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd96b39f-bd2e-48d1-85c3-577cf97f08c8", "external-id": "cl2-zone-84", "segmentation_id": 84, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa0edea0b-d2", "ovs_interfaceid": "a0edea0b-d2cc-4b03-9d46-9cc5e6cd1794", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2039.294852] env[62824]: DEBUG oslo_concurrency.lockutils [req-a80d6b32-d524-4df5-8357-82b6065483f6 req-c3a1bb39-fbc8-4ee5-870f-3f51dc995b49 service nova] Acquired lock "refresh_cache-c63b05b7-1dbf-4637-8728-33673be0d3ea" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2039.294987] env[62824]: DEBUG nova.network.neutron [req-a80d6b32-d524-4df5-8357-82b6065483f6 req-c3a1bb39-fbc8-4ee5-870f-3f51dc995b49 service nova] [instance: c63b05b7-1dbf-4637-8728-33673be0d3ea] Refreshing network info cache for port a0edea0b-d2cc-4b03-9d46-9cc5e6cd1794 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2039.295942] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] [instance: c63b05b7-1dbf-4637-8728-33673be0d3ea] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:17:95:fb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dd96b39f-bd2e-48d1-85c3-577cf97f08c8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a0edea0b-d2cc-4b03-9d46-9cc5e6cd1794', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2039.303374] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Creating folder: Project (8face4a61e49498eae3a9920f9a91a1f). Parent ref: group-v438503. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 2039.304399] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d9590f9d-e9e9-47ca-bcfb-e0e17ff2a33f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2039.322025] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Created folder: Project (8face4a61e49498eae3a9920f9a91a1f) in parent group-v438503. [ 2039.322025] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Creating folder: Instances. Parent ref: group-v438634. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 2039.322025] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e20e1b44-073a-401f-8418-d7b6518fbaf8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2039.331334] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Created folder: Instances in parent group-v438634. [ 2039.331607] env[62824]: DEBUG oslo.service.loopingcall [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2039.331808] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c63b05b7-1dbf-4637-8728-33673be0d3ea] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2039.332040] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cf18ec8c-a614-4efb-af65-552ca52a527c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2039.351008] env[62824]: DEBUG nova.objects.instance [None req-a95c14b1-46f7-482d-8cd9-d9d9543d1991 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Lazy-loading 'flavor' on Instance uuid 59b5b883-4188-471c-8862-444f3ce08cb0 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2039.356449] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2039.356449] env[62824]: value = "task-2146078" [ 2039.356449] env[62824]: _type = "Task" [ 2039.356449] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2039.368551] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146078, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2039.445243] env[62824]: DEBUG oslo_vmware.api [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Task: {'id': task-2146071, 'name': ReconfigVM_Task, 'duration_secs': 0.672908} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2039.445243] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] [instance: 4423d172-acdd-4c69-b3b8-ff166e1b8548] Reconfigured VM instance instance-00000032 to attach disk [datastore1] 4423d172-acdd-4c69-b3b8-ff166e1b8548/4423d172-acdd-4c69-b3b8-ff166e1b8548.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2039.445243] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d822c907-ea60-45db-9a7c-eb93b15f7b48 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2039.452713] env[62824]: DEBUG oslo_vmware.api [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Waiting for the task: (returnval){ [ 2039.452713] env[62824]: value = "task-2146079" [ 2039.452713] env[62824]: _type = "Task" [ 2039.452713] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2039.462966] env[62824]: DEBUG oslo_vmware.api [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Task: {'id': task-2146079, 'name': Rename_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2039.542849] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146074, 'name': CreateVM_Task, 'duration_secs': 0.548551} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2039.543058] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6a724332-a165-4a2b-9dd8-85f27e7b7637] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2039.543535] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2039.543707] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2039.544039] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2039.544315] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7442cf6b-6611-4d68-bac5-2c1669d6b42d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2039.551209] env[62824]: DEBUG oslo_vmware.api [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Waiting for the task: (returnval){ [ 2039.551209] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52235e5b-d384-181e-7a3b-21dd0d1d6a22" [ 2039.551209] env[62824]: _type = "Task" [ 2039.551209] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2039.560750] env[62824]: DEBUG oslo_vmware.api [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52235e5b-d384-181e-7a3b-21dd0d1d6a22, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2039.603813] env[62824]: DEBUG oslo_vmware.api [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]529b4bce-d3c3-3a31-def4-1b1a0111cdb2, 'name': SearchDatastore_Task, 'duration_secs': 0.014789} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2039.604629] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f81f276e-4d20-4b66-ae2a-4d6278cbcae3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2039.612407] env[62824]: DEBUG oslo_vmware.api [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the task: (returnval){ [ 2039.612407] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52fb0f76-d68b-b9bd-699d-3eddd4098078" [ 2039.612407] env[62824]: _type = "Task" [ 2039.612407] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2039.622698] env[62824]: DEBUG oslo_vmware.api [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52fb0f76-d68b-b9bd-699d-3eddd4098078, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2039.632807] env[62824]: DEBUG oslo_vmware.api [None req-457a3e46-ac18-4c02-91aa-31023cbb56f4 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2146075, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2039.793872] env[62824]: DEBUG nova.compute.utils [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2039.798706] env[62824]: DEBUG nova.compute.manager [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 21c05725-3dad-4965-98d0-0622ebcc8ece] Not allocating networking since 'none' was specified. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1984}} [ 2039.859070] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a95c14b1-46f7-482d-8cd9-d9d9543d1991 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Acquiring lock "refresh_cache-59b5b883-4188-471c-8862-444f3ce08cb0" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2039.859070] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a95c14b1-46f7-482d-8cd9-d9d9543d1991 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Acquired lock "refresh_cache-59b5b883-4188-471c-8862-444f3ce08cb0" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2039.859070] env[62824]: DEBUG nova.network.neutron [None req-a95c14b1-46f7-482d-8cd9-d9d9543d1991 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2039.859070] env[62824]: DEBUG nova.objects.instance [None req-a95c14b1-46f7-482d-8cd9-d9d9543d1991 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Lazy-loading 'info_cache' on Instance uuid 59b5b883-4188-471c-8862-444f3ce08cb0 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2039.876134] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146078, 'name': CreateVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2039.969478] env[62824]: DEBUG oslo_vmware.api [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Task: {'id': task-2146079, 'name': Rename_Task, 'duration_secs': 0.29016} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2039.969974] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] [instance: 4423d172-acdd-4c69-b3b8-ff166e1b8548] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2039.970070] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-56dcda13-d60d-4556-8cd8-16f7d06849db {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2039.979296] env[62824]: DEBUG oslo_vmware.api [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Waiting for the task: (returnval){ [ 2039.979296] env[62824]: value = "task-2146080" [ 2039.979296] env[62824]: _type = "Task" [ 2039.979296] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2039.991290] env[62824]: DEBUG oslo_vmware.api [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Task: {'id': task-2146080, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2040.066542] env[62824]: DEBUG oslo_vmware.api [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52235e5b-d384-181e-7a3b-21dd0d1d6a22, 'name': SearchDatastore_Task, 'duration_secs': 0.014603} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2040.067580] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2040.067580] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 6a724332-a165-4a2b-9dd8-85f27e7b7637] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2040.067580] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2040.067730] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2040.067887] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2040.068252] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-63c014e6-5007-4a43-91ac-6d128ae8090c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2040.087095] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2040.087095] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2040.090517] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-41152c88-bb49-48cf-8dc1-5c6738e1906e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2040.098438] env[62824]: DEBUG oslo_vmware.api [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Waiting for the task: (returnval){ [ 2040.098438] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5237833b-8e09-393a-4f53-b41f7577ab0f" [ 2040.098438] env[62824]: _type = "Task" [ 2040.098438] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2040.110198] env[62824]: DEBUG oslo_vmware.api [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5237833b-8e09-393a-4f53-b41f7577ab0f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2040.127249] env[62824]: DEBUG oslo_vmware.api [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52fb0f76-d68b-b9bd-699d-3eddd4098078, 'name': SearchDatastore_Task, 'duration_secs': 0.01952} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2040.136128] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2040.136481] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 88bad2dd-dce7-41df-b56c-93a5d054c11e/88bad2dd-dce7-41df-b56c-93a5d054c11e.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2040.137048] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6cccb984-212a-4773-8b08-58ccaf6b27eb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2040.148810] env[62824]: DEBUG oslo_vmware.api [None req-457a3e46-ac18-4c02-91aa-31023cbb56f4 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2146075, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2040.153108] env[62824]: DEBUG oslo_vmware.api [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the task: (returnval){ [ 2040.153108] env[62824]: value = "task-2146081" [ 2040.153108] env[62824]: _type = "Task" [ 2040.153108] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2040.162602] env[62824]: DEBUG oslo_vmware.api [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146081, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2040.180657] env[62824]: DEBUG nova.network.neutron [req-a80d6b32-d524-4df5-8357-82b6065483f6 req-c3a1bb39-fbc8-4ee5-870f-3f51dc995b49 service nova] [instance: c63b05b7-1dbf-4637-8728-33673be0d3ea] Updated VIF entry in instance network info cache for port a0edea0b-d2cc-4b03-9d46-9cc5e6cd1794. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2040.181051] env[62824]: DEBUG nova.network.neutron [req-a80d6b32-d524-4df5-8357-82b6065483f6 req-c3a1bb39-fbc8-4ee5-870f-3f51dc995b49 service nova] [instance: c63b05b7-1dbf-4637-8728-33673be0d3ea] Updating instance_info_cache with network_info: [{"id": "a0edea0b-d2cc-4b03-9d46-9cc5e6cd1794", "address": "fa:16:3e:17:95:fb", "network": {"id": "eb8486ef-73d8-4ef6-a534-fbdae6bc7fad", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-628019626-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8face4a61e49498eae3a9920f9a91a1f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd96b39f-bd2e-48d1-85c3-577cf97f08c8", "external-id": "cl2-zone-84", "segmentation_id": 84, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa0edea0b-d2", "ovs_interfaceid": "a0edea0b-d2cc-4b03-9d46-9cc5e6cd1794", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2040.299169] env[62824]: DEBUG nova.compute.manager [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 21c05725-3dad-4965-98d0-0622ebcc8ece] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2040.365652] env[62824]: DEBUG nova.objects.base [None req-a95c14b1-46f7-482d-8cd9-d9d9543d1991 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Object Instance<59b5b883-4188-471c-8862-444f3ce08cb0> lazy-loaded attributes: flavor,info_cache {{(pid=62824) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 2040.370296] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146078, 'name': CreateVM_Task, 'duration_secs': 0.62756} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2040.373654] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c63b05b7-1dbf-4637-8728-33673be0d3ea] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2040.376742] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2040.376742] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2040.376742] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2040.378114] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cfb18fab-a7ce-4d85-ac9f-5a51077b4723 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2040.389159] env[62824]: DEBUG oslo_vmware.api [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Waiting for the task: (returnval){ [ 2040.389159] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]528a9bea-2b79-f246-22e5-3bce1c180467" [ 2040.389159] env[62824]: _type = "Task" [ 2040.389159] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2040.408032] env[62824]: DEBUG oslo_vmware.api [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]528a9bea-2b79-f246-22e5-3bce1c180467, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2040.444832] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-508c98c1-4ef3-45bb-b99e-03cc361935bb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2040.454809] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b2a5cae-dc8c-46b3-9f1c-bfc40dac2e43 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2040.497181] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66a46445-a3f9-4723-a8d3-364738683de4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2040.509081] env[62824]: DEBUG oslo_vmware.api [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Task: {'id': task-2146080, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2040.513424] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5c81093-00f8-414d-b0fc-d3c0e2d5e0bc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2040.531054] env[62824]: DEBUG nova.compute.provider_tree [None req-f8aad8ab-6936-4c8d-8d26-3f40750d45f7 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2040.611515] env[62824]: DEBUG oslo_vmware.api [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5237833b-8e09-393a-4f53-b41f7577ab0f, 'name': SearchDatastore_Task, 'duration_secs': 0.023211} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2040.614848] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-61c04816-7855-4c4a-a15a-a67f1e5b56c3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2040.620584] env[62824]: DEBUG oslo_vmware.api [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Waiting for the task: (returnval){ [ 2040.620584] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52597f09-e282-096f-34ed-95e8b3977b85" [ 2040.620584] env[62824]: _type = "Task" [ 2040.620584] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2040.634589] env[62824]: DEBUG oslo_vmware.api [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52597f09-e282-096f-34ed-95e8b3977b85, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2040.641651] env[62824]: DEBUG oslo_vmware.api [None req-457a3e46-ac18-4c02-91aa-31023cbb56f4 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2146075, 'name': CloneVM_Task, 'duration_secs': 1.975365} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2040.641940] env[62824]: INFO nova.virt.vmwareapi.vmops [None req-457a3e46-ac18-4c02-91aa-31023cbb56f4 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Created linked-clone VM from snapshot [ 2040.642762] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1773af4b-d5e9-46be-b969-852b16dbde1c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2040.653045] env[62824]: DEBUG nova.virt.vmwareapi.images [None req-457a3e46-ac18-4c02-91aa-31023cbb56f4 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Uploading image 3ad686d1-b5e5-42bb-af64-cda5329ea3f2 {{(pid=62824) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 2040.664478] env[62824]: DEBUG oslo_vmware.api [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146081, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2040.681539] env[62824]: DEBUG oslo_vmware.rw_handles [None req-457a3e46-ac18-4c02-91aa-31023cbb56f4 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 2040.681539] env[62824]: value = "vm-438633" [ 2040.681539] env[62824]: _type = "VirtualMachine" [ 2040.681539] env[62824]: }. {{(pid=62824) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 2040.681859] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-e1eb4808-d168-4313-a195-d9aecf563c5a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2040.684212] env[62824]: DEBUG oslo_concurrency.lockutils [req-a80d6b32-d524-4df5-8357-82b6065483f6 req-c3a1bb39-fbc8-4ee5-870f-3f51dc995b49 service nova] Releasing lock "refresh_cache-c63b05b7-1dbf-4637-8728-33673be0d3ea" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2040.692996] env[62824]: DEBUG oslo_vmware.rw_handles [None req-457a3e46-ac18-4c02-91aa-31023cbb56f4 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Lease: (returnval){ [ 2040.692996] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]520a1ce5-2e9b-feb7-752b-c2132eb10153" [ 2040.692996] env[62824]: _type = "HttpNfcLease" [ 2040.692996] env[62824]: } obtained for exporting VM: (result){ [ 2040.692996] env[62824]: value = "vm-438633" [ 2040.692996] env[62824]: _type = "VirtualMachine" [ 2040.692996] env[62824]: }. {{(pid=62824) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 2040.693483] env[62824]: DEBUG oslo_vmware.api [None req-457a3e46-ac18-4c02-91aa-31023cbb56f4 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Waiting for the lease: (returnval){ [ 2040.693483] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]520a1ce5-2e9b-feb7-752b-c2132eb10153" [ 2040.693483] env[62824]: _type = "HttpNfcLease" [ 2040.693483] env[62824]: } to be ready. {{(pid=62824) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 2040.702804] env[62824]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 2040.702804] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]520a1ce5-2e9b-feb7-752b-c2132eb10153" [ 2040.702804] env[62824]: _type = "HttpNfcLease" [ 2040.702804] env[62824]: } is initializing. {{(pid=62824) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 2040.901444] env[62824]: DEBUG oslo_vmware.api [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]528a9bea-2b79-f246-22e5-3bce1c180467, 'name': SearchDatastore_Task, 'duration_secs': 0.016317} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2040.901733] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2040.902062] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] [instance: c63b05b7-1dbf-4637-8728-33673be0d3ea] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2040.902253] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2041.003977] env[62824]: DEBUG oslo_vmware.api [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Task: {'id': task-2146080, 'name': PowerOnVM_Task, 'duration_secs': 0.769199} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2041.006511] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] [instance: 4423d172-acdd-4c69-b3b8-ff166e1b8548] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2041.006749] env[62824]: INFO nova.compute.manager [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] [instance: 4423d172-acdd-4c69-b3b8-ff166e1b8548] Took 11.24 seconds to spawn the instance on the hypervisor. [ 2041.007545] env[62824]: DEBUG nova.compute.manager [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] [instance: 4423d172-acdd-4c69-b3b8-ff166e1b8548] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2041.007792] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6692460-3ddf-4596-8a52-3e342d00b35e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2041.034161] env[62824]: DEBUG nova.scheduler.client.report [None req-f8aad8ab-6936-4c8d-8d26-3f40750d45f7 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2041.134914] env[62824]: DEBUG oslo_vmware.api [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52597f09-e282-096f-34ed-95e8b3977b85, 'name': SearchDatastore_Task, 'duration_secs': 0.029271} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2041.135376] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2041.135637] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 6a724332-a165-4a2b-9dd8-85f27e7b7637/6a724332-a165-4a2b-9dd8-85f27e7b7637.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2041.135951] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2041.136353] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2041.136408] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e3e92a1f-063f-47b2-a6da-eaedc34c872e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2041.138738] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3bf3b390-0e46-4a86-b9fa-13fbd66ea286 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2041.150717] env[62824]: DEBUG oslo_vmware.api [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Waiting for the task: (returnval){ [ 2041.150717] env[62824]: value = "task-2146083" [ 2041.150717] env[62824]: _type = "Task" [ 2041.150717] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2041.152218] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2041.152500] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2041.156741] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fcebbecd-0730-458c-9de0-0cd263456731 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2041.169924] env[62824]: DEBUG oslo_vmware.api [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Task: {'id': task-2146083, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2041.174226] env[62824]: DEBUG oslo_vmware.api [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146081, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.659539} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2041.174532] env[62824]: DEBUG oslo_vmware.api [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Waiting for the task: (returnval){ [ 2041.174532] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5249cc63-68e4-da75-1c7d-e428c46418a1" [ 2041.174532] env[62824]: _type = "Task" [ 2041.174532] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2041.175230] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 88bad2dd-dce7-41df-b56c-93a5d054c11e/88bad2dd-dce7-41df-b56c-93a5d054c11e.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2041.175230] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 88bad2dd-dce7-41df-b56c-93a5d054c11e] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2041.175405] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-22d2b444-a228-4b34-a367-4a089af9ea0c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2041.186658] env[62824]: DEBUG oslo_vmware.api [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5249cc63-68e4-da75-1c7d-e428c46418a1, 'name': SearchDatastore_Task, 'duration_secs': 0.015142} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2041.190160] env[62824]: DEBUG oslo_vmware.api [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the task: (returnval){ [ 2041.190160] env[62824]: value = "task-2146084" [ 2041.190160] env[62824]: _type = "Task" [ 2041.190160] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2041.190160] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-97f43aab-dc35-44d6-b1c1-f12d0d91053a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2041.200189] env[62824]: DEBUG oslo_vmware.api [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Waiting for the task: (returnval){ [ 2041.200189] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52eab841-f395-6730-3ca8-d403fe38de0a" [ 2041.200189] env[62824]: _type = "Task" [ 2041.200189] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2041.211258] env[62824]: DEBUG oslo_vmware.api [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146084, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2041.211258] env[62824]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 2041.211258] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]520a1ce5-2e9b-feb7-752b-c2132eb10153" [ 2041.211258] env[62824]: _type = "HttpNfcLease" [ 2041.211258] env[62824]: } is ready. {{(pid=62824) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 2041.211258] env[62824]: DEBUG oslo_vmware.rw_handles [None req-457a3e46-ac18-4c02-91aa-31023cbb56f4 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 2041.211258] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]520a1ce5-2e9b-feb7-752b-c2132eb10153" [ 2041.211258] env[62824]: _type = "HttpNfcLease" [ 2041.211258] env[62824]: }. {{(pid=62824) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 2041.211258] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9d28fed-fc53-42b6-a40b-55959601d864 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2041.216580] env[62824]: DEBUG oslo_vmware.api [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52eab841-f395-6730-3ca8-d403fe38de0a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2041.222761] env[62824]: DEBUG oslo_vmware.rw_handles [None req-457a3e46-ac18-4c02-91aa-31023cbb56f4 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e5ec0c-f672-1439-f6cb-723be3e13a42/disk-0.vmdk from lease info. {{(pid=62824) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 2041.223010] env[62824]: DEBUG oslo_vmware.rw_handles [None req-457a3e46-ac18-4c02-91aa-31023cbb56f4 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e5ec0c-f672-1439-f6cb-723be3e13a42/disk-0.vmdk for reading. {{(pid=62824) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 2041.309061] env[62824]: DEBUG nova.compute.manager [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 21c05725-3dad-4965-98d0-0622ebcc8ece] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2041.334239] env[62824]: DEBUG nova.virt.hardware [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2041.334499] env[62824]: DEBUG nova.virt.hardware [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2041.334657] env[62824]: DEBUG nova.virt.hardware [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2041.334844] env[62824]: DEBUG nova.virt.hardware [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2041.334984] env[62824]: DEBUG nova.virt.hardware [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2041.335191] env[62824]: DEBUG nova.virt.hardware [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2041.335429] env[62824]: DEBUG nova.virt.hardware [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2041.335635] env[62824]: DEBUG nova.virt.hardware [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2041.335818] env[62824]: DEBUG nova.virt.hardware [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2041.335991] env[62824]: DEBUG nova.virt.hardware [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2041.336174] env[62824]: DEBUG nova.virt.hardware [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2041.337134] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ac09130-99e3-48fe-bdc1-138868a43a79 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2041.346443] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dff321fc-6e72-4232-b968-d21dfaccfd09 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2041.353425] env[62824]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-dc1b7b92-f931-4213-8756-8b082e3c08b1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2041.364547] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 21c05725-3dad-4965-98d0-0622ebcc8ece] Instance VIF info [] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2041.370421] env[62824]: DEBUG oslo.service.loopingcall [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2041.371597] env[62824]: DEBUG nova.network.neutron [None req-a95c14b1-46f7-482d-8cd9-d9d9543d1991 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Updating instance_info_cache with network_info: [{"id": "1e74959f-64c2-4c37-85f8-fa8f81a30d60", "address": "fa:16:3e:6d:26:78", "network": {"id": "ae1dfabd-223a-4d4c-a8c4-fe65a83dd9c0", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-797102932-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.148", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "102d2f886fd44e1da6780e48e36f1b62", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "517421c3-bea0-419c-ab0b-987815e5d160", "external-id": "nsx-vlan-transportzone-68", "segmentation_id": 68, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1e74959f-64", "ovs_interfaceid": "1e74959f-64c2-4c37-85f8-fa8f81a30d60", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2041.375897] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 21c05725-3dad-4965-98d0-0622ebcc8ece] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2041.376651] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0f0949d3-9451-46e8-be0e-306ce81a131c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2041.392679] env[62824]: DEBUG nova.compute.manager [None req-28126ed4-3627-4b81-ab9c-0a91bbeea6cf tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: c65ccd20-b682-420a-9c1a-47a45959197e] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2041.394054] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7049e88b-86ce-4309-8ea1-24b7af9314f2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2041.404555] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2041.404555] env[62824]: value = "task-2146085" [ 2041.404555] env[62824]: _type = "Task" [ 2041.404555] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2041.417112] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146085, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2041.533020] env[62824]: INFO nova.compute.manager [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] [instance: 4423d172-acdd-4c69-b3b8-ff166e1b8548] Took 49.80 seconds to build instance. [ 2041.540918] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f8aad8ab-6936-4c8d-8d26-3f40750d45f7 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.252s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2041.549384] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 32.031s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2041.553937] env[62824]: INFO nova.compute.claims [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] [instance: ed7d3e95-f26e-40fd-a370-e17922bbff8e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2041.585798] env[62824]: INFO nova.scheduler.client.report [None req-f8aad8ab-6936-4c8d-8d26-3f40750d45f7 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Deleted allocations for instance 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c [ 2041.669831] env[62824]: DEBUG oslo_vmware.api [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Task: {'id': task-2146083, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2041.708324] env[62824]: DEBUG oslo_vmware.api [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146084, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.089789} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2041.709130] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 88bad2dd-dce7-41df-b56c-93a5d054c11e] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2041.710075] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-151c1e1f-6d5a-497e-8f81-238607ddf980 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2041.716495] env[62824]: DEBUG oslo_vmware.api [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52eab841-f395-6730-3ca8-d403fe38de0a, 'name': SearchDatastore_Task, 'duration_secs': 0.016632} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2041.717280] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2041.717652] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] c63b05b7-1dbf-4637-8728-33673be0d3ea/c63b05b7-1dbf-4637-8728-33673be0d3ea.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2041.717992] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-726a2030-127c-4268-a670-7baa7fc878de {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2041.742308] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 88bad2dd-dce7-41df-b56c-93a5d054c11e] Reconfiguring VM instance instance-00000033 to attach disk [datastore1] 88bad2dd-dce7-41df-b56c-93a5d054c11e/88bad2dd-dce7-41df-b56c-93a5d054c11e.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2041.743374] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-232a0bef-3067-4f44-badb-2ebd2df3dbc3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2041.760201] env[62824]: DEBUG oslo_vmware.api [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Waiting for the task: (returnval){ [ 2041.760201] env[62824]: value = "task-2146086" [ 2041.760201] env[62824]: _type = "Task" [ 2041.760201] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2041.767136] env[62824]: DEBUG oslo_vmware.api [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the task: (returnval){ [ 2041.767136] env[62824]: value = "task-2146087" [ 2041.767136] env[62824]: _type = "Task" [ 2041.767136] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2041.774023] env[62824]: DEBUG oslo_vmware.api [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Task: {'id': task-2146086, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2041.783740] env[62824]: DEBUG oslo_vmware.api [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146087, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2041.866213] env[62824]: DEBUG nova.objects.instance [None req-234f8d25-619a-49f9-b628-6c9162e49d13 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Lazy-loading 'flavor' on Instance uuid 7febb8b4-f984-4d79-a888-b2829f2a9df6 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2041.877642] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a95c14b1-46f7-482d-8cd9-d9d9543d1991 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Releasing lock "refresh_cache-59b5b883-4188-471c-8862-444f3ce08cb0" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2041.918979] env[62824]: INFO nova.compute.manager [None req-28126ed4-3627-4b81-ab9c-0a91bbeea6cf tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: c65ccd20-b682-420a-9c1a-47a45959197e] instance snapshotting [ 2041.920724] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146085, 'name': CreateVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2041.922179] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1aae62f-a7c5-42df-9d47-ed8b53c4b064 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2041.943585] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ee08643-b37b-4d75-b28d-35eb905dff6e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2042.037330] env[62824]: DEBUG oslo_concurrency.lockutils [None req-73313ea7-5317-4594-908c-13d3dd01cce2 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Lock "4423d172-acdd-4c69-b3b8-ff166e1b8548" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 103.078s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2042.101958] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f8aad8ab-6936-4c8d-8d26-3f40750d45f7 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Lock "9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 38.304s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2042.168499] env[62824]: DEBUG oslo_vmware.api [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Task: {'id': task-2146083, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.589659} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2042.168891] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 6a724332-a165-4a2b-9dd8-85f27e7b7637/6a724332-a165-4a2b-9dd8-85f27e7b7637.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2042.169424] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 6a724332-a165-4a2b-9dd8-85f27e7b7637] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2042.169629] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-556acf89-bd60-4e31-a0a2-3da139582bed {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2042.181994] env[62824]: DEBUG oslo_vmware.api [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Waiting for the task: (returnval){ [ 2042.181994] env[62824]: value = "task-2146088" [ 2042.181994] env[62824]: _type = "Task" [ 2042.181994] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2042.194648] env[62824]: DEBUG oslo_vmware.api [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Task: {'id': task-2146088, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2042.276870] env[62824]: DEBUG oslo_vmware.api [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Task: {'id': task-2146086, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2042.284999] env[62824]: DEBUG oslo_vmware.api [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146087, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2042.370758] env[62824]: DEBUG oslo_concurrency.lockutils [None req-234f8d25-619a-49f9-b628-6c9162e49d13 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Acquiring lock "refresh_cache-7febb8b4-f984-4d79-a888-b2829f2a9df6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2042.371148] env[62824]: DEBUG oslo_concurrency.lockutils [None req-234f8d25-619a-49f9-b628-6c9162e49d13 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Acquired lock "refresh_cache-7febb8b4-f984-4d79-a888-b2829f2a9df6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2042.427315] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146085, 'name': CreateVM_Task, 'duration_secs': 0.544476} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2042.427763] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 21c05725-3dad-4965-98d0-0622ebcc8ece] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2042.428421] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2042.428637] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2042.429226] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2042.429636] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5f63e4e2-f620-48c5-a7ce-dd9a3f62a49a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2042.436760] env[62824]: DEBUG oslo_vmware.api [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Waiting for the task: (returnval){ [ 2042.436760] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52d221b4-7be3-6314-9083-aed0682096cb" [ 2042.436760] env[62824]: _type = "Task" [ 2042.436760] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2042.453505] env[62824]: DEBUG oslo_vmware.api [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52d221b4-7be3-6314-9083-aed0682096cb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2042.457423] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-28126ed4-3627-4b81-ab9c-0a91bbeea6cf tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: c65ccd20-b682-420a-9c1a-47a45959197e] Creating Snapshot of the VM instance {{(pid=62824) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 2042.457970] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-ac29eb72-b621-4a41-adf6-9b824875c0c6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2042.469943] env[62824]: DEBUG oslo_vmware.api [None req-28126ed4-3627-4b81-ab9c-0a91bbeea6cf tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Waiting for the task: (returnval){ [ 2042.469943] env[62824]: value = "task-2146089" [ 2042.469943] env[62824]: _type = "Task" [ 2042.469943] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2042.484318] env[62824]: DEBUG oslo_vmware.api [None req-28126ed4-3627-4b81-ab9c-0a91bbeea6cf tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146089, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2042.540411] env[62824]: DEBUG nova.compute.manager [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 15479062-af75-4925-99b3-77d6a49751ad] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2042.697084] env[62824]: DEBUG oslo_vmware.api [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Task: {'id': task-2146088, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.234637} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2042.699291] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 6a724332-a165-4a2b-9dd8-85f27e7b7637] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2042.700975] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0b4af1a-903d-484c-98cf-aa7e2ddbedde {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2042.727276] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 6a724332-a165-4a2b-9dd8-85f27e7b7637] Reconfiguring VM instance instance-00000035 to attach disk [datastore2] 6a724332-a165-4a2b-9dd8-85f27e7b7637/6a724332-a165-4a2b-9dd8-85f27e7b7637.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2042.733269] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7b399b7a-da61-430b-896e-019f684372d6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2042.750567] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8eacba3d-7056-4a91-bc0c-0f40d3c3cacc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2042.759020] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-4d352032-1bf1-4db4-bb25-8e513ed3b754 tempest-ServersAdminNegativeTestJSON-1450984671 tempest-ServersAdminNegativeTestJSON-1450984671-project-admin] [instance: 4423d172-acdd-4c69-b3b8-ff166e1b8548] Suspending the VM {{(pid=62824) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 2042.759506] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-419d9126-881a-4c62-bb26-1778a24cc682 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2042.763567] env[62824]: DEBUG oslo_vmware.api [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Waiting for the task: (returnval){ [ 2042.763567] env[62824]: value = "task-2146090" [ 2042.763567] env[62824]: _type = "Task" [ 2042.763567] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2042.779156] env[62824]: DEBUG oslo_vmware.api [None req-4d352032-1bf1-4db4-bb25-8e513ed3b754 tempest-ServersAdminNegativeTestJSON-1450984671 tempest-ServersAdminNegativeTestJSON-1450984671-project-admin] Waiting for the task: (returnval){ [ 2042.779156] env[62824]: value = "task-2146091" [ 2042.779156] env[62824]: _type = "Task" [ 2042.779156] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2042.795548] env[62824]: DEBUG oslo_vmware.api [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146087, 'name': ReconfigVM_Task, 'duration_secs': 0.669149} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2042.795967] env[62824]: DEBUG oslo_vmware.api [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Task: {'id': task-2146086, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.70316} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2042.796231] env[62824]: DEBUG oslo_vmware.api [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Task: {'id': task-2146090, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2042.797312] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 88bad2dd-dce7-41df-b56c-93a5d054c11e] Reconfigured VM instance instance-00000033 to attach disk [datastore1] 88bad2dd-dce7-41df-b56c-93a5d054c11e/88bad2dd-dce7-41df-b56c-93a5d054c11e.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2042.798319] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] c63b05b7-1dbf-4637-8728-33673be0d3ea/c63b05b7-1dbf-4637-8728-33673be0d3ea.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2042.798627] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] [instance: c63b05b7-1dbf-4637-8728-33673be0d3ea] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2042.799125] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fd939c97-edc8-4065-8dc6-6a575eeb1a4c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2042.801223] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-edc369f7-9535-4535-80f7-487b9d39fdd0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2042.807379] env[62824]: DEBUG oslo_vmware.api [None req-4d352032-1bf1-4db4-bb25-8e513ed3b754 tempest-ServersAdminNegativeTestJSON-1450984671 tempest-ServersAdminNegativeTestJSON-1450984671-project-admin] Task: {'id': task-2146091, 'name': SuspendVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2042.817622] env[62824]: DEBUG oslo_vmware.api [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Waiting for the task: (returnval){ [ 2042.817622] env[62824]: value = "task-2146093" [ 2042.817622] env[62824]: _type = "Task" [ 2042.817622] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2042.820196] env[62824]: DEBUG oslo_vmware.api [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the task: (returnval){ [ 2042.820196] env[62824]: value = "task-2146092" [ 2042.820196] env[62824]: _type = "Task" [ 2042.820196] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2042.837923] env[62824]: DEBUG oslo_vmware.api [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146092, 'name': Rename_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2042.844099] env[62824]: DEBUG oslo_vmware.api [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Task: {'id': task-2146093, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2042.883872] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-a95c14b1-46f7-482d-8cd9-d9d9543d1991 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2042.883988] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bbb895ba-f2dc-49e4-92e3-e8fafba0a8d3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2042.898035] env[62824]: DEBUG oslo_vmware.api [None req-a95c14b1-46f7-482d-8cd9-d9d9543d1991 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Waiting for the task: (returnval){ [ 2042.898035] env[62824]: value = "task-2146094" [ 2042.898035] env[62824]: _type = "Task" [ 2042.898035] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2042.910702] env[62824]: DEBUG oslo_vmware.api [None req-a95c14b1-46f7-482d-8cd9-d9d9543d1991 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Task: {'id': task-2146094, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2042.952789] env[62824]: DEBUG oslo_vmware.api [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52d221b4-7be3-6314-9083-aed0682096cb, 'name': SearchDatastore_Task, 'duration_secs': 0.031207} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2042.953346] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2042.953676] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 21c05725-3dad-4965-98d0-0622ebcc8ece] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2042.954018] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2042.954210] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2042.954570] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2042.954955] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2f464b81-6707-4e8e-af71-b5f4e60065f4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2042.969286] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2042.969286] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2042.972036] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2054d5c9-b1c8-4051-85dd-aed17a1bd0ec {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2042.986231] env[62824]: DEBUG oslo_vmware.api [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Waiting for the task: (returnval){ [ 2042.986231] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]525067b8-5e96-eb8d-2b00-32637458b546" [ 2042.986231] env[62824]: _type = "Task" [ 2042.986231] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2042.989818] env[62824]: DEBUG oslo_vmware.api [None req-28126ed4-3627-4b81-ab9c-0a91bbeea6cf tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146089, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2043.007027] env[62824]: DEBUG nova.network.neutron [None req-234f8d25-619a-49f9-b628-6c9162e49d13 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2043.024323] env[62824]: DEBUG oslo_vmware.api [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]525067b8-5e96-eb8d-2b00-32637458b546, 'name': SearchDatastore_Task, 'duration_secs': 0.015917} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2043.030800] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-03b98015-86f6-44ab-bcf0-d7b1145f8cdc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2043.039149] env[62824]: DEBUG oslo_vmware.api [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Waiting for the task: (returnval){ [ 2043.039149] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5259a639-96fc-05da-ec8a-34ec12f04ab1" [ 2043.039149] env[62824]: _type = "Task" [ 2043.039149] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2043.057191] env[62824]: DEBUG oslo_vmware.api [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5259a639-96fc-05da-ec8a-34ec12f04ab1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2043.065316] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2043.235896] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40a1f954-5175-46c5-8ee8-831667de6335 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2043.244745] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b335505-293e-40a9-87f7-15f95f79852b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2043.293850] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-696d2701-5742-4562-ab88-7b8b821b4f39 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2043.302745] env[62824]: DEBUG oslo_vmware.api [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Task: {'id': task-2146090, 'name': ReconfigVM_Task, 'duration_secs': 0.529891} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2043.308193] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 6a724332-a165-4a2b-9dd8-85f27e7b7637] Reconfigured VM instance instance-00000035 to attach disk [datastore2] 6a724332-a165-4a2b-9dd8-85f27e7b7637/6a724332-a165-4a2b-9dd8-85f27e7b7637.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2043.309995] env[62824]: DEBUG oslo_vmware.api [None req-4d352032-1bf1-4db4-bb25-8e513ed3b754 tempest-ServersAdminNegativeTestJSON-1450984671 tempest-ServersAdminNegativeTestJSON-1450984671-project-admin] Task: {'id': task-2146091, 'name': SuspendVM_Task} progress is 62%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2043.309995] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-95234524-4bc2-42f3-9a8d-e252af6fdeff {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2043.312455] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd18a1ca-afb9-4473-9c46-46f1b601c488 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2043.336493] env[62824]: DEBUG nova.compute.provider_tree [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Updating inventory in ProviderTree for provider bf95157b-a274-42de-9ccf-9851128a44a1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 173, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 2043.338167] env[62824]: DEBUG oslo_vmware.api [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Waiting for the task: (returnval){ [ 2043.338167] env[62824]: value = "task-2146095" [ 2043.338167] env[62824]: _type = "Task" [ 2043.338167] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2043.348329] env[62824]: DEBUG oslo_vmware.api [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Task: {'id': task-2146093, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079238} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2043.349946] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] [instance: c63b05b7-1dbf-4637-8728-33673be0d3ea] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2043.351358] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6b6a17d-093e-42ef-b484-a97ac4781d38 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2043.361351] env[62824]: DEBUG oslo_vmware.api [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146092, 'name': Rename_Task, 'duration_secs': 0.241078} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2043.361648] env[62824]: DEBUG oslo_vmware.api [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Task: {'id': task-2146095, 'name': Rename_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2043.362415] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 88bad2dd-dce7-41df-b56c-93a5d054c11e] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2043.362732] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d11b523f-cd8a-4892-80d6-4707d27dc90f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2043.383635] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] [instance: c63b05b7-1dbf-4637-8728-33673be0d3ea] Reconfiguring VM instance instance-00000034 to attach disk [datastore2] c63b05b7-1dbf-4637-8728-33673be0d3ea/c63b05b7-1dbf-4637-8728-33673be0d3ea.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2043.385767] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-582065ff-ebbd-4bc7-b77d-c5baef846284 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2043.404322] env[62824]: DEBUG oslo_vmware.api [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the task: (returnval){ [ 2043.404322] env[62824]: value = "task-2146096" [ 2043.404322] env[62824]: _type = "Task" [ 2043.404322] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2043.415706] env[62824]: DEBUG oslo_vmware.api [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Waiting for the task: (returnval){ [ 2043.415706] env[62824]: value = "task-2146097" [ 2043.415706] env[62824]: _type = "Task" [ 2043.415706] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2043.422998] env[62824]: DEBUG oslo_vmware.api [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146096, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2043.423389] env[62824]: DEBUG oslo_vmware.api [None req-a95c14b1-46f7-482d-8cd9-d9d9543d1991 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Task: {'id': task-2146094, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2043.434995] env[62824]: DEBUG oslo_vmware.api [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Task: {'id': task-2146097, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2043.487913] env[62824]: DEBUG oslo_vmware.api [None req-28126ed4-3627-4b81-ab9c-0a91bbeea6cf tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146089, 'name': CreateSnapshot_Task, 'duration_secs': 0.981199} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2043.488309] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-28126ed4-3627-4b81-ab9c-0a91bbeea6cf tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: c65ccd20-b682-420a-9c1a-47a45959197e] Created Snapshot of the VM instance {{(pid=62824) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 2043.489293] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd5f38b1-c3f1-421f-a742-13cf684f01f7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2043.552832] env[62824]: DEBUG oslo_vmware.api [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5259a639-96fc-05da-ec8a-34ec12f04ab1, 'name': SearchDatastore_Task, 'duration_secs': 0.016328} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2043.555675] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2043.556019] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 21c05725-3dad-4965-98d0-0622ebcc8ece/21c05725-3dad-4965-98d0-0622ebcc8ece.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2043.556404] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7d968f40-4b29-48fb-8786-7e54f72ad656 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2043.565924] env[62824]: DEBUG oslo_vmware.api [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Waiting for the task: (returnval){ [ 2043.565924] env[62824]: value = "task-2146098" [ 2043.565924] env[62824]: _type = "Task" [ 2043.565924] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2043.576326] env[62824]: DEBUG oslo_vmware.api [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Task: {'id': task-2146098, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2043.797674] env[62824]: DEBUG oslo_vmware.api [None req-4d352032-1bf1-4db4-bb25-8e513ed3b754 tempest-ServersAdminNegativeTestJSON-1450984671 tempest-ServersAdminNegativeTestJSON-1450984671-project-admin] Task: {'id': task-2146091, 'name': SuspendVM_Task, 'duration_secs': 0.995582} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2043.798288] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-4d352032-1bf1-4db4-bb25-8e513ed3b754 tempest-ServersAdminNegativeTestJSON-1450984671 tempest-ServersAdminNegativeTestJSON-1450984671-project-admin] [instance: 4423d172-acdd-4c69-b3b8-ff166e1b8548] Suspended the VM {{(pid=62824) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 2043.798644] env[62824]: DEBUG nova.compute.manager [None req-4d352032-1bf1-4db4-bb25-8e513ed3b754 tempest-ServersAdminNegativeTestJSON-1450984671 tempest-ServersAdminNegativeTestJSON-1450984671-project-admin] [instance: 4423d172-acdd-4c69-b3b8-ff166e1b8548] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2043.799574] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a21d10b0-afca-4cb8-b5a6-84ea9fa8cf9c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2043.803479] env[62824]: DEBUG nova.network.neutron [None req-234f8d25-619a-49f9-b628-6c9162e49d13 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Updating instance_info_cache with network_info: [{"id": "b27bda53-5f70-40ff-bb04-9b22f2241093", "address": "fa:16:3e:4d:09:75", "network": {"id": "3d2f0f85-4d51-4459-ad99-443b0821fd9d", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1138734647-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}, {"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "adf40520c1ef4d97b1ffe2c8cc9649b9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "10b81051-1eb1-406b-888c-4548c470c77e", "external-id": "nsx-vlan-transportzone-207", "segmentation_id": 207, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb27bda53-5f", "ovs_interfaceid": "b27bda53-5f70-40ff-bb04-9b22f2241093", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2043.857045] env[62824]: DEBUG oslo_vmware.api [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Task: {'id': task-2146095, 'name': Rename_Task, 'duration_secs': 0.227267} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2043.857364] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 6a724332-a165-4a2b-9dd8-85f27e7b7637] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2043.857636] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-30b546fa-291d-4b4e-8018-de06b572878d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2043.863430] env[62824]: ERROR nova.scheduler.client.report [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] [req-2732e71c-f829-4622-aeb4-7d7d2290bbb6] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 173, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID bf95157b-a274-42de-9ccf-9851128a44a1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-2732e71c-f829-4622-aeb4-7d7d2290bbb6"}]} [ 2043.868381] env[62824]: DEBUG oslo_vmware.api [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Waiting for the task: (returnval){ [ 2043.868381] env[62824]: value = "task-2146099" [ 2043.868381] env[62824]: _type = "Task" [ 2043.868381] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2043.887591] env[62824]: DEBUG oslo_vmware.api [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Task: {'id': task-2146099, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2043.888936] env[62824]: DEBUG nova.scheduler.client.report [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Refreshing inventories for resource provider bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 2043.906318] env[62824]: DEBUG nova.scheduler.client.report [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Updating ProviderTree inventory for provider bf95157b-a274-42de-9ccf-9851128a44a1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 2043.906642] env[62824]: DEBUG nova.compute.provider_tree [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Updating inventory in ProviderTree for provider bf95157b-a274-42de-9ccf-9851128a44a1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 2043.922838] env[62824]: DEBUG nova.scheduler.client.report [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Refreshing aggregate associations for resource provider bf95157b-a274-42de-9ccf-9851128a44a1, aggregates: None {{(pid=62824) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 2043.937429] env[62824]: DEBUG oslo_vmware.api [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146096, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2043.937753] env[62824]: DEBUG oslo_vmware.api [None req-a95c14b1-46f7-482d-8cd9-d9d9543d1991 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Task: {'id': task-2146094, 'name': PowerOnVM_Task, 'duration_secs': 0.64876} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2043.938036] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-a95c14b1-46f7-482d-8cd9-d9d9543d1991 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2043.938264] env[62824]: DEBUG nova.compute.manager [None req-a95c14b1-46f7-482d-8cd9-d9d9543d1991 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2043.939740] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5037338e-657c-4076-b166-c166690fc587 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2043.945952] env[62824]: DEBUG oslo_vmware.api [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Task: {'id': task-2146097, 'name': ReconfigVM_Task, 'duration_secs': 0.512341} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2043.946797] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] [instance: c63b05b7-1dbf-4637-8728-33673be0d3ea] Reconfigured VM instance instance-00000034 to attach disk [datastore2] c63b05b7-1dbf-4637-8728-33673be0d3ea/c63b05b7-1dbf-4637-8728-33673be0d3ea.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2043.947522] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-700010b3-23cd-44c2-b9c8-5cd18d58c613 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2043.955250] env[62824]: DEBUG nova.scheduler.client.report [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Refreshing trait associations for resource provider bf95157b-a274-42de-9ccf-9851128a44a1, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_ISO {{(pid=62824) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 2043.966959] env[62824]: DEBUG oslo_vmware.api [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Waiting for the task: (returnval){ [ 2043.966959] env[62824]: value = "task-2146100" [ 2043.966959] env[62824]: _type = "Task" [ 2043.966959] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2043.982130] env[62824]: DEBUG oslo_vmware.api [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Task: {'id': task-2146100, 'name': Rename_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2044.011384] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-28126ed4-3627-4b81-ab9c-0a91bbeea6cf tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: c65ccd20-b682-420a-9c1a-47a45959197e] Creating linked-clone VM from snapshot {{(pid=62824) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 2044.015233] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-1ea8c952-0f50-4444-bb84-2f419f2d5dfa {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2044.029026] env[62824]: DEBUG oslo_vmware.api [None req-28126ed4-3627-4b81-ab9c-0a91bbeea6cf tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Waiting for the task: (returnval){ [ 2044.029026] env[62824]: value = "task-2146101" [ 2044.029026] env[62824]: _type = "Task" [ 2044.029026] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2044.047834] env[62824]: DEBUG oslo_vmware.api [None req-28126ed4-3627-4b81-ab9c-0a91bbeea6cf tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146101, 'name': CloneVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2044.085362] env[62824]: DEBUG oslo_vmware.api [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Task: {'id': task-2146098, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2044.307031] env[62824]: DEBUG oslo_concurrency.lockutils [None req-234f8d25-619a-49f9-b628-6c9162e49d13 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Releasing lock "refresh_cache-7febb8b4-f984-4d79-a888-b2829f2a9df6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2044.307467] env[62824]: DEBUG nova.compute.manager [None req-234f8d25-619a-49f9-b628-6c9162e49d13 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Inject network info {{(pid=62824) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7737}} [ 2044.307839] env[62824]: DEBUG nova.compute.manager [None req-234f8d25-619a-49f9-b628-6c9162e49d13 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] network_info to inject: |[{"id": "b27bda53-5f70-40ff-bb04-9b22f2241093", "address": "fa:16:3e:4d:09:75", "network": {"id": "3d2f0f85-4d51-4459-ad99-443b0821fd9d", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1138734647-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}, {"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "adf40520c1ef4d97b1ffe2c8cc9649b9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "10b81051-1eb1-406b-888c-4548c470c77e", "external-id": "nsx-vlan-transportzone-207", "segmentation_id": 207, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb27bda53-5f", "ovs_interfaceid": "b27bda53-5f70-40ff-bb04-9b22f2241093", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7738}} [ 2044.313818] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-234f8d25-619a-49f9-b628-6c9162e49d13 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Reconfiguring VM instance to set the machine id {{(pid=62824) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1796}} [ 2044.319604] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-78d7bf47-f8f5-479d-bf2d-5cf418f02c3f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2044.338807] env[62824]: DEBUG oslo_vmware.api [None req-234f8d25-619a-49f9-b628-6c9162e49d13 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Waiting for the task: (returnval){ [ 2044.338807] env[62824]: value = "task-2146102" [ 2044.338807] env[62824]: _type = "Task" [ 2044.338807] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2044.351751] env[62824]: DEBUG oslo_vmware.api [None req-234f8d25-619a-49f9-b628-6c9162e49d13 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Task: {'id': task-2146102, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2044.383267] env[62824]: DEBUG oslo_vmware.api [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Task: {'id': task-2146099, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2044.425335] env[62824]: DEBUG oslo_vmware.api [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146096, 'name': PowerOnVM_Task, 'duration_secs': 0.695275} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2044.425786] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 88bad2dd-dce7-41df-b56c-93a5d054c11e] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2044.425837] env[62824]: INFO nova.compute.manager [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 88bad2dd-dce7-41df-b56c-93a5d054c11e] Took 11.77 seconds to spawn the instance on the hypervisor. [ 2044.426166] env[62824]: DEBUG nova.compute.manager [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 88bad2dd-dce7-41df-b56c-93a5d054c11e] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2044.427201] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fad76c9-de8e-4607-9b4c-4369f8668c5f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2044.484042] env[62824]: DEBUG oslo_vmware.api [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Task: {'id': task-2146100, 'name': Rename_Task, 'duration_secs': 0.463767} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2044.484371] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] [instance: c63b05b7-1dbf-4637-8728-33673be0d3ea] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2044.485548] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5b7822bd-0a61-48b2-a0fc-668593328c01 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2044.495933] env[62824]: DEBUG oslo_vmware.api [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Waiting for the task: (returnval){ [ 2044.495933] env[62824]: value = "task-2146103" [ 2044.495933] env[62824]: _type = "Task" [ 2044.495933] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2044.512949] env[62824]: DEBUG oslo_vmware.api [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Task: {'id': task-2146103, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2044.542309] env[62824]: DEBUG oslo_vmware.api [None req-28126ed4-3627-4b81-ab9c-0a91bbeea6cf tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146101, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2044.573957] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46aa9acf-27b0-4599-931f-fd3dcb9ab9ff {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2044.582425] env[62824]: DEBUG oslo_vmware.api [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Task: {'id': task-2146098, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.644183} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2044.583350] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 21c05725-3dad-4965-98d0-0622ebcc8ece/21c05725-3dad-4965-98d0-0622ebcc8ece.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2044.584226] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 21c05725-3dad-4965-98d0-0622ebcc8ece] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2044.584226] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3cf125d8-a4f1-4d47-b774-10a50a7b1322 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2044.589857] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b008fd4a-8aa5-4a10-bd44-a5f4bb291285 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2044.595459] env[62824]: DEBUG oslo_vmware.api [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Waiting for the task: (returnval){ [ 2044.595459] env[62824]: value = "task-2146104" [ 2044.595459] env[62824]: _type = "Task" [ 2044.595459] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2044.636771] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2b48b24-30a4-4a8a-b914-2776186dc83b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2044.640117] env[62824]: DEBUG oslo_vmware.api [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Task: {'id': task-2146104, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2044.646964] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b85b934-42cf-4fbb-ab02-3627b9af95c3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2044.662659] env[62824]: DEBUG nova.compute.provider_tree [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Updating inventory in ProviderTree for provider bf95157b-a274-42de-9ccf-9851128a44a1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 173, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 2044.849896] env[62824]: DEBUG oslo_vmware.api [None req-234f8d25-619a-49f9-b628-6c9162e49d13 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Task: {'id': task-2146102, 'name': ReconfigVM_Task, 'duration_secs': 0.263585} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2044.850196] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-234f8d25-619a-49f9-b628-6c9162e49d13 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Reconfigured VM instance to set the machine id {{(pid=62824) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1799}} [ 2044.881038] env[62824]: DEBUG oslo_vmware.api [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Task: {'id': task-2146099, 'name': PowerOnVM_Task, 'duration_secs': 0.815014} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2044.881038] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 6a724332-a165-4a2b-9dd8-85f27e7b7637] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2044.881038] env[62824]: INFO nova.compute.manager [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 6a724332-a165-4a2b-9dd8-85f27e7b7637] Took 6.50 seconds to spawn the instance on the hypervisor. [ 2044.881323] env[62824]: DEBUG nova.compute.manager [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 6a724332-a165-4a2b-9dd8-85f27e7b7637] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2044.882120] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7251a2a7-494d-46d6-b190-4374e139c02f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2044.948063] env[62824]: INFO nova.compute.manager [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 88bad2dd-dce7-41df-b56c-93a5d054c11e] Took 48.02 seconds to build instance. [ 2045.024411] env[62824]: DEBUG oslo_vmware.api [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Task: {'id': task-2146103, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2045.043121] env[62824]: DEBUG oslo_vmware.api [None req-28126ed4-3627-4b81-ab9c-0a91bbeea6cf tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146101, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2045.107440] env[62824]: DEBUG oslo_vmware.api [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Task: {'id': task-2146104, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.11578} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2045.107836] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 21c05725-3dad-4965-98d0-0622ebcc8ece] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2045.108678] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78c8bbb5-f757-4bf4-855b-989045c41a3c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2045.134097] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 21c05725-3dad-4965-98d0-0622ebcc8ece] Reconfiguring VM instance instance-00000036 to attach disk [datastore2] 21c05725-3dad-4965-98d0-0622ebcc8ece/21c05725-3dad-4965-98d0-0622ebcc8ece.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2045.134464] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3f0b02e2-4bef-49d7-b732-d54fd2286b71 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2045.157449] env[62824]: DEBUG oslo_vmware.api [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Waiting for the task: (returnval){ [ 2045.157449] env[62824]: value = "task-2146105" [ 2045.157449] env[62824]: _type = "Task" [ 2045.157449] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2045.170126] env[62824]: DEBUG oslo_vmware.api [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Task: {'id': task-2146105, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2045.204189] env[62824]: DEBUG nova.scheduler.client.report [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Updated inventory for provider bf95157b-a274-42de-9ccf-9851128a44a1 with generation 76 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 173, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 2045.204528] env[62824]: DEBUG nova.compute.provider_tree [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Updating resource provider bf95157b-a274-42de-9ccf-9851128a44a1 generation from 76 to 77 during operation: update_inventory {{(pid=62824) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 2045.204712] env[62824]: DEBUG nova.compute.provider_tree [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Updating inventory in ProviderTree for provider bf95157b-a274-42de-9ccf-9851128a44a1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 173, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 2045.404446] env[62824]: INFO nova.compute.manager [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 6a724332-a165-4a2b-9dd8-85f27e7b7637] Took 41.20 seconds to build instance. [ 2045.450126] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a959719e-7b58-4ffd-9011-e00a1b8e1748 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Lock "88bad2dd-dce7-41df-b56c-93a5d054c11e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 98.129s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2045.507135] env[62824]: DEBUG oslo_vmware.api [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Task: {'id': task-2146103, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2045.541821] env[62824]: DEBUG oslo_vmware.api [None req-28126ed4-3627-4b81-ab9c-0a91bbeea6cf tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146101, 'name': CloneVM_Task} progress is 100%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2045.668309] env[62824]: DEBUG oslo_vmware.api [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Task: {'id': task-2146105, 'name': ReconfigVM_Task, 'duration_secs': 0.417741} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2045.668680] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 21c05725-3dad-4965-98d0-0622ebcc8ece] Reconfigured VM instance instance-00000036 to attach disk [datastore2] 21c05725-3dad-4965-98d0-0622ebcc8ece/21c05725-3dad-4965-98d0-0622ebcc8ece.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2045.669393] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-00e8d237-3df9-4b4d-b5d3-3241ed1a67dd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2045.678421] env[62824]: DEBUG oslo_vmware.api [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Waiting for the task: (returnval){ [ 2045.678421] env[62824]: value = "task-2146106" [ 2045.678421] env[62824]: _type = "Task" [ 2045.678421] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2045.688496] env[62824]: DEBUG oslo_vmware.api [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Task: {'id': task-2146106, 'name': Rename_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2045.711055] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 4.162s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2045.711631] env[62824]: DEBUG nova.compute.manager [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] [instance: ed7d3e95-f26e-40fd-a370-e17922bbff8e] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2045.714578] env[62824]: DEBUG oslo_concurrency.lockutils [None req-6008bdc7-ed30-4e87-8975-1eeef6a5d44f tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 32.609s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2045.714805] env[62824]: DEBUG nova.objects.instance [None req-6008bdc7-ed30-4e87-8975-1eeef6a5d44f tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Lazy-loading 'resources' on Instance uuid e158b5d4-c120-4e6c-89c4-7668e097926d {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2045.907922] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c6da376a-9c3b-447f-bfec-79b7605c322a tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Lock "6a724332-a165-4a2b-9dd8-85f27e7b7637" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 88.510s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2045.953251] env[62824]: DEBUG nova.compute.manager [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: a6cd2032-de60-4f78-bf1e-79801d049df0] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2046.022246] env[62824]: DEBUG oslo_vmware.api [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Task: {'id': task-2146103, 'name': PowerOnVM_Task, 'duration_secs': 1.330127} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2046.022547] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] [instance: c63b05b7-1dbf-4637-8728-33673be0d3ea] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2046.022753] env[62824]: INFO nova.compute.manager [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] [instance: c63b05b7-1dbf-4637-8728-33673be0d3ea] Took 10.53 seconds to spawn the instance on the hypervisor. [ 2046.022929] env[62824]: DEBUG nova.compute.manager [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] [instance: c63b05b7-1dbf-4637-8728-33673be0d3ea] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2046.023829] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19526d42-8554-46f3-b709-ccc632bc5a83 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2046.043022] env[62824]: DEBUG oslo_vmware.api [None req-28126ed4-3627-4b81-ab9c-0a91bbeea6cf tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146101, 'name': CloneVM_Task, 'duration_secs': 1.551429} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2046.043360] env[62824]: INFO nova.virt.vmwareapi.vmops [None req-28126ed4-3627-4b81-ab9c-0a91bbeea6cf tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: c65ccd20-b682-420a-9c1a-47a45959197e] Created linked-clone VM from snapshot [ 2046.044147] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f33d6dc1-b788-4c6a-a555-2045343970c3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2046.052777] env[62824]: DEBUG nova.virt.vmwareapi.images [None req-28126ed4-3627-4b81-ab9c-0a91bbeea6cf tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: c65ccd20-b682-420a-9c1a-47a45959197e] Uploading image 6c8b574b-aa50-4be7-896f-fe8af1b20d01 {{(pid=62824) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 2046.191679] env[62824]: DEBUG oslo_vmware.api [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Task: {'id': task-2146106, 'name': Rename_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2046.217893] env[62824]: DEBUG nova.compute.utils [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2046.222392] env[62824]: DEBUG nova.compute.manager [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] [instance: ed7d3e95-f26e-40fd-a370-e17922bbff8e] Not allocating networking since 'none' was specified. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1984}} [ 2046.330445] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-28126ed4-3627-4b81-ab9c-0a91bbeea6cf tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: c65ccd20-b682-420a-9c1a-47a45959197e] Destroying the VM {{(pid=62824) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 2046.330942] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-028f36a0-1b71-4172-97ed-3c7c9d20dfc1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2046.341320] env[62824]: DEBUG oslo_vmware.api [None req-28126ed4-3627-4b81-ab9c-0a91bbeea6cf tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Waiting for the task: (returnval){ [ 2046.341320] env[62824]: value = "task-2146107" [ 2046.341320] env[62824]: _type = "Task" [ 2046.341320] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2046.351981] env[62824]: DEBUG oslo_vmware.rw_handles [None req-76aebf6c-e1c2-4914-b5ed-170d40366a5a tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52645093-a57f-8239-89a9-271e4e104280/disk-0.vmdk. {{(pid=62824) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 2046.353149] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70f1f6a9-d14d-4c2f-a874-5e55f0169ea8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2046.362199] env[62824]: DEBUG oslo_vmware.api [None req-28126ed4-3627-4b81-ab9c-0a91bbeea6cf tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146107, 'name': Destroy_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2046.367563] env[62824]: DEBUG oslo_vmware.rw_handles [None req-76aebf6c-e1c2-4914-b5ed-170d40366a5a tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52645093-a57f-8239-89a9-271e4e104280/disk-0.vmdk is in state: ready. {{(pid=62824) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 2046.367563] env[62824]: ERROR oslo_vmware.rw_handles [None req-76aebf6c-e1c2-4914-b5ed-170d40366a5a tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52645093-a57f-8239-89a9-271e4e104280/disk-0.vmdk due to incomplete transfer. [ 2046.367830] env[62824]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-c37ad432-9c22-434e-94b0-3d7b3d1ad8f0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2046.377394] env[62824]: DEBUG oslo_vmware.rw_handles [None req-76aebf6c-e1c2-4914-b5ed-170d40366a5a tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52645093-a57f-8239-89a9-271e4e104280/disk-0.vmdk. {{(pid=62824) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 2046.377673] env[62824]: DEBUG nova.virt.vmwareapi.images [None req-76aebf6c-e1c2-4914-b5ed-170d40366a5a tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Uploaded image 45e8c4ae-efb5-498b-869d-2abeee747d8b to the Glance image server {{(pid=62824) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 2046.379926] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-76aebf6c-e1c2-4914-b5ed-170d40366a5a tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Destroying the VM {{(pid=62824) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 2046.382789] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-9a3c54b3-1e11-4b05-9752-c72a0992f6e3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2046.393543] env[62824]: DEBUG oslo_vmware.api [None req-76aebf6c-e1c2-4914-b5ed-170d40366a5a tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Waiting for the task: (returnval){ [ 2046.393543] env[62824]: value = "task-2146108" [ 2046.393543] env[62824]: _type = "Task" [ 2046.393543] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2046.415777] env[62824]: DEBUG oslo_vmware.api [None req-76aebf6c-e1c2-4914-b5ed-170d40366a5a tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Task: {'id': task-2146108, 'name': Destroy_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2046.416899] env[62824]: DEBUG nova.compute.manager [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02f34565-6675-4c79-ac47-b131ceba9df8] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2046.476621] env[62824]: DEBUG oslo_concurrency.lockutils [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2046.556416] env[62824]: INFO nova.compute.manager [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] [instance: c63b05b7-1dbf-4637-8728-33673be0d3ea] Took 45.01 seconds to build instance. [ 2046.613640] env[62824]: DEBUG oslo_concurrency.lockutils [None req-895a9f6d-c24d-4e89-b988-6024da95cb4d tempest-VolumesAssistedSnapshotsTest-163918615 tempest-VolumesAssistedSnapshotsTest-163918615-project-admin] Acquiring lock "29b8dd5f-1855-490c-a01b-54840073a753" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2046.614509] env[62824]: DEBUG oslo_concurrency.lockutils [None req-895a9f6d-c24d-4e89-b988-6024da95cb4d tempest-VolumesAssistedSnapshotsTest-163918615 tempest-VolumesAssistedSnapshotsTest-163918615-project-admin] Lock "29b8dd5f-1855-490c-a01b-54840073a753" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2046.693354] env[62824]: DEBUG oslo_vmware.api [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Task: {'id': task-2146106, 'name': Rename_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2046.722975] env[62824]: DEBUG nova.compute.manager [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] [instance: ed7d3e95-f26e-40fd-a370-e17922bbff8e] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2046.807386] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5bc692f-d16e-4ad9-b528-88fad5d28bb5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2046.816758] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad8db094-120c-4020-8c4e-93cee523ebbe {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2046.856485] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-006c39e5-2d7a-4bb9-8365-d3f9933a95bb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2046.869905] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c498732-e03c-4990-98f5-9b384595f4c1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2046.874883] env[62824]: DEBUG oslo_vmware.api [None req-28126ed4-3627-4b81-ab9c-0a91bbeea6cf tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146107, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2046.889577] env[62824]: DEBUG nova.compute.provider_tree [None req-6008bdc7-ed30-4e87-8975-1eeef6a5d44f tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2046.904104] env[62824]: DEBUG oslo_vmware.api [None req-76aebf6c-e1c2-4914-b5ed-170d40366a5a tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Task: {'id': task-2146108, 'name': Destroy_Task} progress is 100%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2046.943593] env[62824]: DEBUG oslo_concurrency.lockutils [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2047.058379] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e3d3c9ac-4c0f-4037-8595-0a6e8a175fdf tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Lock "c63b05b7-1dbf-4637-8728-33673be0d3ea" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 95.624s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2047.120981] env[62824]: DEBUG nova.compute.utils [None req-895a9f6d-c24d-4e89-b988-6024da95cb4d tempest-VolumesAssistedSnapshotsTest-163918615 tempest-VolumesAssistedSnapshotsTest-163918615-project-admin] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2047.191691] env[62824]: DEBUG oslo_vmware.api [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Task: {'id': task-2146106, 'name': Rename_Task, 'duration_secs': 1.189999} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2047.192008] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 21c05725-3dad-4965-98d0-0622ebcc8ece] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2047.192905] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6cf10bac-8bb2-4529-a2e1-0f267fc818dd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2047.201204] env[62824]: DEBUG oslo_vmware.api [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Waiting for the task: (returnval){ [ 2047.201204] env[62824]: value = "task-2146109" [ 2047.201204] env[62824]: _type = "Task" [ 2047.201204] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2047.204028] env[62824]: DEBUG nova.compute.manager [req-b7b4d360-8efa-4dea-bd28-311a3af3b63b req-00e29436-a0dc-4d84-b80f-dfd371a5c650 service nova] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Received event network-changed-b27bda53-5f70-40ff-bb04-9b22f2241093 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2047.204245] env[62824]: DEBUG nova.compute.manager [req-b7b4d360-8efa-4dea-bd28-311a3af3b63b req-00e29436-a0dc-4d84-b80f-dfd371a5c650 service nova] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Refreshing instance network info cache due to event network-changed-b27bda53-5f70-40ff-bb04-9b22f2241093. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2047.204488] env[62824]: DEBUG oslo_concurrency.lockutils [req-b7b4d360-8efa-4dea-bd28-311a3af3b63b req-00e29436-a0dc-4d84-b80f-dfd371a5c650 service nova] Acquiring lock "refresh_cache-7febb8b4-f984-4d79-a888-b2829f2a9df6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2047.204658] env[62824]: DEBUG oslo_concurrency.lockutils [req-b7b4d360-8efa-4dea-bd28-311a3af3b63b req-00e29436-a0dc-4d84-b80f-dfd371a5c650 service nova] Acquired lock "refresh_cache-7febb8b4-f984-4d79-a888-b2829f2a9df6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2047.204838] env[62824]: DEBUG nova.network.neutron [req-b7b4d360-8efa-4dea-bd28-311a3af3b63b req-00e29436-a0dc-4d84-b80f-dfd371a5c650 service nova] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Refreshing network info cache for port b27bda53-5f70-40ff-bb04-9b22f2241093 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2047.216702] env[62824]: DEBUG oslo_vmware.api [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Task: {'id': task-2146109, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2047.290524] env[62824]: DEBUG oslo_concurrency.lockutils [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Acquiring lock "b8cc8cd3-ea03-40bf-b867-7ad193365552" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2047.290833] env[62824]: DEBUG oslo_concurrency.lockutils [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Lock "b8cc8cd3-ea03-40bf-b867-7ad193365552" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2047.291099] env[62824]: INFO nova.compute.manager [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Shelving [ 2047.363653] env[62824]: DEBUG oslo_vmware.api [None req-28126ed4-3627-4b81-ab9c-0a91bbeea6cf tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146107, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2047.393912] env[62824]: DEBUG nova.scheduler.client.report [None req-6008bdc7-ed30-4e87-8975-1eeef6a5d44f tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 173, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2047.411671] env[62824]: DEBUG oslo_vmware.api [None req-76aebf6c-e1c2-4914-b5ed-170d40366a5a tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Task: {'id': task-2146108, 'name': Destroy_Task, 'duration_secs': 0.519775} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2047.412346] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-76aebf6c-e1c2-4914-b5ed-170d40366a5a tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Destroyed the VM [ 2047.412346] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-76aebf6c-e1c2-4914-b5ed-170d40366a5a tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Deleting Snapshot of the VM instance {{(pid=62824) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 2047.412527] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-5bd0b1a7-56c0-4365-917e-49243ea2e20e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2047.424698] env[62824]: DEBUG oslo_vmware.api [None req-76aebf6c-e1c2-4914-b5ed-170d40366a5a tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Waiting for the task: (returnval){ [ 2047.424698] env[62824]: value = "task-2146110" [ 2047.424698] env[62824]: _type = "Task" [ 2047.424698] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2047.436772] env[62824]: DEBUG oslo_vmware.api [None req-76aebf6c-e1c2-4914-b5ed-170d40366a5a tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Task: {'id': task-2146110, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2047.624705] env[62824]: DEBUG oslo_concurrency.lockutils [None req-895a9f6d-c24d-4e89-b988-6024da95cb4d tempest-VolumesAssistedSnapshotsTest-163918615 tempest-VolumesAssistedSnapshotsTest-163918615-project-admin] Lock "29b8dd5f-1855-490c-a01b-54840073a753" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.010s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2047.717294] env[62824]: DEBUG oslo_vmware.api [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Task: {'id': task-2146109, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2047.734867] env[62824]: DEBUG nova.compute.manager [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] [instance: ed7d3e95-f26e-40fd-a370-e17922bbff8e] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2047.765664] env[62824]: DEBUG nova.virt.hardware [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2047.766026] env[62824]: DEBUG nova.virt.hardware [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2047.766253] env[62824]: DEBUG nova.virt.hardware [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2047.766529] env[62824]: DEBUG nova.virt.hardware [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2047.766729] env[62824]: DEBUG nova.virt.hardware [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2047.766927] env[62824]: DEBUG nova.virt.hardware [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2047.767246] env[62824]: DEBUG nova.virt.hardware [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2047.767459] env[62824]: DEBUG nova.virt.hardware [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2047.767716] env[62824]: DEBUG nova.virt.hardware [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2047.767946] env[62824]: DEBUG nova.virt.hardware [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2047.768202] env[62824]: DEBUG nova.virt.hardware [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2047.769473] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f2d5f04-981c-40f6-aba1-c7b5a0b2543e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2047.784643] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41a39d22-35d0-440c-98ff-b5418fcdc7da {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2047.811258] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] [instance: ed7d3e95-f26e-40fd-a370-e17922bbff8e] Instance VIF info [] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2047.820416] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Creating folder: Project (9971fa077f954d919b130f96263cc2b3). Parent ref: group-v438503. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 2047.823972] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-da75010f-cdea-4f11-b1b2-7a43ab5572f5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2047.844290] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Created folder: Project (9971fa077f954d919b130f96263cc2b3) in parent group-v438503. [ 2047.844695] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Creating folder: Instances. Parent ref: group-v438640. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 2047.845018] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8150103c-cc84-4c65-9c11-b0ecdf89bd85 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2047.861577] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Created folder: Instances in parent group-v438640. [ 2047.861995] env[62824]: DEBUG oslo.service.loopingcall [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2047.862756] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ed7d3e95-f26e-40fd-a370-e17922bbff8e] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2047.863149] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-593b99e9-37b2-4f43-9dee-5cfcca327220 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2047.892297] env[62824]: DEBUG oslo_vmware.api [None req-28126ed4-3627-4b81-ab9c-0a91bbeea6cf tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146107, 'name': Destroy_Task} progress is 100%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2047.904263] env[62824]: DEBUG oslo_concurrency.lockutils [None req-6008bdc7-ed30-4e87-8975-1eeef6a5d44f tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.189s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2047.909707] env[62824]: DEBUG oslo_concurrency.lockutils [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.065s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2047.912322] env[62824]: INFO nova.compute.claims [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: b0456b8a-348a-4503-a92c-58e5ab455d1c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2047.920032] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2047.920032] env[62824]: value = "task-2146113" [ 2047.920032] env[62824]: _type = "Task" [ 2047.920032] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2047.946764] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146113, 'name': CreateVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2047.947164] env[62824]: DEBUG oslo_vmware.api [None req-76aebf6c-e1c2-4914-b5ed-170d40366a5a tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Task: {'id': task-2146110, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2047.954247] env[62824]: INFO nova.scheduler.client.report [None req-6008bdc7-ed30-4e87-8975-1eeef6a5d44f tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Deleted allocations for instance e158b5d4-c120-4e6c-89c4-7668e097926d [ 2048.142108] env[62824]: DEBUG nova.network.neutron [req-b7b4d360-8efa-4dea-bd28-311a3af3b63b req-00e29436-a0dc-4d84-b80f-dfd371a5c650 service nova] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Updated VIF entry in instance network info cache for port b27bda53-5f70-40ff-bb04-9b22f2241093. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2048.142514] env[62824]: DEBUG nova.network.neutron [req-b7b4d360-8efa-4dea-bd28-311a3af3b63b req-00e29436-a0dc-4d84-b80f-dfd371a5c650 service nova] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Updating instance_info_cache with network_info: [{"id": "b27bda53-5f70-40ff-bb04-9b22f2241093", "address": "fa:16:3e:4d:09:75", "network": {"id": "3d2f0f85-4d51-4459-ad99-443b0821fd9d", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1138734647-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}, {"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "adf40520c1ef4d97b1ffe2c8cc9649b9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "10b81051-1eb1-406b-888c-4548c470c77e", "external-id": "nsx-vlan-transportzone-207", "segmentation_id": 207, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb27bda53-5f", "ovs_interfaceid": "b27bda53-5f70-40ff-bb04-9b22f2241093", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2048.215675] env[62824]: DEBUG oslo_vmware.api [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Task: {'id': task-2146109, 'name': PowerOnVM_Task, 'duration_secs': 0.547871} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2048.216024] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 21c05725-3dad-4965-98d0-0622ebcc8ece] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2048.216284] env[62824]: INFO nova.compute.manager [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 21c05725-3dad-4965-98d0-0622ebcc8ece] Took 6.91 seconds to spawn the instance on the hypervisor. [ 2048.216474] env[62824]: DEBUG nova.compute.manager [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 21c05725-3dad-4965-98d0-0622ebcc8ece] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2048.217478] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44491f26-920e-42ca-89ed-b0b6b9d16479 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2048.329510] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2048.329876] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a84abe05-03d5-475b-8030-eeb5d30ba24d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2048.338776] env[62824]: DEBUG oslo_vmware.api [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Waiting for the task: (returnval){ [ 2048.338776] env[62824]: value = "task-2146114" [ 2048.338776] env[62824]: _type = "Task" [ 2048.338776] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2048.350086] env[62824]: DEBUG oslo_vmware.api [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2146114, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2048.366384] env[62824]: DEBUG oslo_vmware.api [None req-28126ed4-3627-4b81-ab9c-0a91bbeea6cf tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146107, 'name': Destroy_Task} progress is 100%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2048.442267] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146113, 'name': CreateVM_Task, 'duration_secs': 0.516984} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2048.442942] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ed7d3e95-f26e-40fd-a370-e17922bbff8e] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2048.443454] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2048.443627] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2048.444175] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2048.444905] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-10e69365-be3c-4833-be55-cdb15de638e6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2048.451055] env[62824]: DEBUG oslo_vmware.api [None req-76aebf6c-e1c2-4914-b5ed-170d40366a5a tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Task: {'id': task-2146110, 'name': RemoveSnapshot_Task} progress is 98%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2048.458155] env[62824]: DEBUG oslo_vmware.api [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Waiting for the task: (returnval){ [ 2048.458155] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]528b098e-686e-01a7-cd28-6f9dbc12994c" [ 2048.458155] env[62824]: _type = "Task" [ 2048.458155] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2048.468198] env[62824]: DEBUG oslo_concurrency.lockutils [None req-6008bdc7-ed30-4e87-8975-1eeef6a5d44f tempest-MigrationsAdminTest-736701413 tempest-MigrationsAdminTest-736701413-project-member] Lock "e158b5d4-c120-4e6c-89c4-7668e097926d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 38.981s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2048.472416] env[62824]: DEBUG oslo_vmware.api [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]528b098e-686e-01a7-cd28-6f9dbc12994c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2048.645389] env[62824]: DEBUG oslo_concurrency.lockutils [req-b7b4d360-8efa-4dea-bd28-311a3af3b63b req-00e29436-a0dc-4d84-b80f-dfd371a5c650 service nova] Releasing lock "refresh_cache-7febb8b4-f984-4d79-a888-b2829f2a9df6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2048.703516] env[62824]: DEBUG oslo_concurrency.lockutils [None req-895a9f6d-c24d-4e89-b988-6024da95cb4d tempest-VolumesAssistedSnapshotsTest-163918615 tempest-VolumesAssistedSnapshotsTest-163918615-project-admin] Acquiring lock "29b8dd5f-1855-490c-a01b-54840073a753" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2048.703786] env[62824]: DEBUG oslo_concurrency.lockutils [None req-895a9f6d-c24d-4e89-b988-6024da95cb4d tempest-VolumesAssistedSnapshotsTest-163918615 tempest-VolumesAssistedSnapshotsTest-163918615-project-admin] Lock "29b8dd5f-1855-490c-a01b-54840073a753" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2048.704035] env[62824]: INFO nova.compute.manager [None req-895a9f6d-c24d-4e89-b988-6024da95cb4d tempest-VolumesAssistedSnapshotsTest-163918615 tempest-VolumesAssistedSnapshotsTest-163918615-project-admin] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Attaching volume f3a0e2f5-ded9-47fe-8fbe-8eeb061be84a to /dev/sdb [ 2048.743952] env[62824]: INFO nova.compute.manager [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 21c05725-3dad-4965-98d0-0622ebcc8ece] Took 42.94 seconds to build instance. [ 2048.754884] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-972545d9-8a99-436e-983e-72fdbbf4a05b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2048.769022] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a37487c1-1588-436b-b2d7-ce81a135617d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2048.787728] env[62824]: DEBUG nova.virt.block_device [None req-895a9f6d-c24d-4e89-b988-6024da95cb4d tempest-VolumesAssistedSnapshotsTest-163918615 tempest-VolumesAssistedSnapshotsTest-163918615-project-admin] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Updating existing volume attachment record: 4bce65f1-1d13-405b-b14d-a0b95d092347 {{(pid=62824) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 2048.796270] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74df1243-146d-46d2-9117-f6faf82e0c48 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2048.805553] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c45d28a1-60ff-488a-a2c6-b18422b58044 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 88bad2dd-dce7-41df-b56c-93a5d054c11e] Suspending the VM {{(pid=62824) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 2048.805955] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-7b76ed40-08c1-4a02-8155-c75eaa4fc6a6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2048.818100] env[62824]: DEBUG oslo_vmware.api [None req-c45d28a1-60ff-488a-a2c6-b18422b58044 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the task: (returnval){ [ 2048.818100] env[62824]: value = "task-2146115" [ 2048.818100] env[62824]: _type = "Task" [ 2048.818100] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2048.829865] env[62824]: DEBUG oslo_vmware.api [None req-c45d28a1-60ff-488a-a2c6-b18422b58044 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146115, 'name': SuspendVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2048.852065] env[62824]: DEBUG oslo_vmware.api [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2146114, 'name': PowerOffVM_Task, 'duration_secs': 0.361243} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2048.852550] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2048.853718] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1215aae5-7aca-4378-9743-4bc757b5afb5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2048.867897] env[62824]: DEBUG oslo_vmware.api [None req-28126ed4-3627-4b81-ab9c-0a91bbeea6cf tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146107, 'name': Destroy_Task, 'duration_secs': 2.123275} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2048.883741] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-28126ed4-3627-4b81-ab9c-0a91bbeea6cf tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: c65ccd20-b682-420a-9c1a-47a45959197e] Destroyed the VM [ 2048.883741] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-28126ed4-3627-4b81-ab9c-0a91bbeea6cf tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: c65ccd20-b682-420a-9c1a-47a45959197e] Deleting Snapshot of the VM instance {{(pid=62824) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 2048.883741] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-02c718d9-5539-4605-9c2c-e9b1ec625d01 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2048.886328] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-399d9719-4ce4-4a63-a4e1-fab1ba1fc734 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2048.904085] env[62824]: DEBUG oslo_vmware.api [None req-28126ed4-3627-4b81-ab9c-0a91bbeea6cf tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Waiting for the task: (returnval){ [ 2048.904085] env[62824]: value = "task-2146116" [ 2048.904085] env[62824]: _type = "Task" [ 2048.904085] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2048.918046] env[62824]: DEBUG oslo_vmware.api [None req-28126ed4-3627-4b81-ab9c-0a91bbeea6cf tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146116, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2048.923715] env[62824]: DEBUG nova.objects.instance [None req-cd4b7b59-8657-4fb4-94f8-f660379978d7 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Lazy-loading 'flavor' on Instance uuid 7febb8b4-f984-4d79-a888-b2829f2a9df6 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2048.942508] env[62824]: DEBUG oslo_vmware.api [None req-76aebf6c-e1c2-4914-b5ed-170d40366a5a tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Task: {'id': task-2146110, 'name': RemoveSnapshot_Task, 'duration_secs': 1.132025} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2048.943877] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-76aebf6c-e1c2-4914-b5ed-170d40366a5a tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Deleted Snapshot of the VM instance {{(pid=62824) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 2048.944218] env[62824]: INFO nova.compute.manager [None req-76aebf6c-e1c2-4914-b5ed-170d40366a5a tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Took 18.84 seconds to snapshot the instance on the hypervisor. [ 2048.978848] env[62824]: DEBUG oslo_vmware.api [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]528b098e-686e-01a7-cd28-6f9dbc12994c, 'name': SearchDatastore_Task, 'duration_secs': 0.016432} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2048.979524] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2048.979853] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] [instance: ed7d3e95-f26e-40fd-a370-e17922bbff8e] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2048.981011] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2048.981011] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2048.981011] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2048.981283] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5d02c636-bf7b-430b-8162-29d4b1bf23e1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2049.002781] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2049.003061] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2049.005627] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e45a38ab-c7e8-43f8-93b1-850d16c34ec4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2049.016734] env[62824]: DEBUG oslo_vmware.api [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Waiting for the task: (returnval){ [ 2049.016734] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52d9ecb5-84f2-08c0-c24c-256d88565b0e" [ 2049.016734] env[62824]: _type = "Task" [ 2049.016734] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2049.032464] env[62824]: DEBUG oslo_vmware.api [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52d9ecb5-84f2-08c0-c24c-256d88565b0e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2049.246073] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e465a3bb-fba2-4d71-9395-807884305696 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Lock "21c05725-3dad-4965-98d0-0622ebcc8ece" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 91.542s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2049.337823] env[62824]: DEBUG oslo_vmware.api [None req-c45d28a1-60ff-488a-a2c6-b18422b58044 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146115, 'name': SuspendVM_Task} progress is 100%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2049.407959] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Creating Snapshot of the VM instance {{(pid=62824) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 2049.408614] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-58a9cba3-2aa6-4202-b0fe-b8ddf5d89f83 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2049.434097] env[62824]: DEBUG oslo_vmware.api [None req-28126ed4-3627-4b81-ab9c-0a91bbeea6cf tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146116, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2049.434422] env[62824]: DEBUG oslo_vmware.api [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Waiting for the task: (returnval){ [ 2049.434422] env[62824]: value = "task-2146120" [ 2049.434422] env[62824]: _type = "Task" [ 2049.434422] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2049.434921] env[62824]: DEBUG oslo_concurrency.lockutils [None req-cd4b7b59-8657-4fb4-94f8-f660379978d7 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Acquiring lock "refresh_cache-7febb8b4-f984-4d79-a888-b2829f2a9df6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2049.438538] env[62824]: DEBUG oslo_concurrency.lockutils [None req-cd4b7b59-8657-4fb4-94f8-f660379978d7 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Acquired lock "refresh_cache-7febb8b4-f984-4d79-a888-b2829f2a9df6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2049.454884] env[62824]: DEBUG oslo_vmware.api [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2146120, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2049.533450] env[62824]: DEBUG oslo_vmware.api [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52d9ecb5-84f2-08c0-c24c-256d88565b0e, 'name': SearchDatastore_Task, 'duration_secs': 0.04078} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2049.534914] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-df8ac2a7-01ea-4b58-b348-7e4275a5f44c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2049.542521] env[62824]: DEBUG oslo_vmware.api [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Waiting for the task: (returnval){ [ 2049.542521] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52f95624-eea4-8560-cd91-11a51e180ed6" [ 2049.542521] env[62824]: _type = "Task" [ 2049.542521] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2049.557845] env[62824]: DEBUG oslo_vmware.api [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52f95624-eea4-8560-cd91-11a51e180ed6, 'name': SearchDatastore_Task} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2049.558208] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2049.558507] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] ed7d3e95-f26e-40fd-a370-e17922bbff8e/ed7d3e95-f26e-40fd-a370-e17922bbff8e.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2049.558813] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d2c98030-e92a-4e74-9267-359a935f1733 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2049.570844] env[62824]: DEBUG oslo_vmware.api [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Waiting for the task: (returnval){ [ 2049.570844] env[62824]: value = "task-2146121" [ 2049.570844] env[62824]: _type = "Task" [ 2049.570844] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2049.581287] env[62824]: DEBUG oslo_vmware.api [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Task: {'id': task-2146121, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2049.615519] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9349d935-f1a8-4b24-b9b8-0ba38ea1da7a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2049.625415] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-972e939f-0976-485b-9493-0d2bcae376a5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2049.659714] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69a7daef-dce7-484a-87e2-220df0d65cd7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2049.668070] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-031dfa21-22be-45ac-aae0-879c802f66f1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2049.684578] env[62824]: DEBUG nova.compute.provider_tree [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2049.833049] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9a50fcc7-25bf-46e5-b936-5cbc1f3c424b tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Acquiring lock "c63b05b7-1dbf-4637-8728-33673be0d3ea" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2049.833484] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9a50fcc7-25bf-46e5-b936-5cbc1f3c424b tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Lock "c63b05b7-1dbf-4637-8728-33673be0d3ea" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2049.833729] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9a50fcc7-25bf-46e5-b936-5cbc1f3c424b tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Acquiring lock "c63b05b7-1dbf-4637-8728-33673be0d3ea-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2049.833960] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9a50fcc7-25bf-46e5-b936-5cbc1f3c424b tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Lock "c63b05b7-1dbf-4637-8728-33673be0d3ea-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2049.834170] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9a50fcc7-25bf-46e5-b936-5cbc1f3c424b tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Lock "c63b05b7-1dbf-4637-8728-33673be0d3ea-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2049.836506] env[62824]: DEBUG oslo_vmware.api [None req-c45d28a1-60ff-488a-a2c6-b18422b58044 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146115, 'name': SuspendVM_Task, 'duration_secs': 0.903904} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2049.836691] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c45d28a1-60ff-488a-a2c6-b18422b58044 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 88bad2dd-dce7-41df-b56c-93a5d054c11e] Suspended the VM {{(pid=62824) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 2049.836984] env[62824]: DEBUG nova.compute.manager [None req-c45d28a1-60ff-488a-a2c6-b18422b58044 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 88bad2dd-dce7-41df-b56c-93a5d054c11e] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2049.837677] env[62824]: INFO nova.compute.manager [None req-9a50fcc7-25bf-46e5-b936-5cbc1f3c424b tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] [instance: c63b05b7-1dbf-4637-8728-33673be0d3ea] Terminating instance [ 2049.840529] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30301b4c-39f0-4b35-8959-4eed0cac2eb6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2049.925429] env[62824]: DEBUG oslo_vmware.api [None req-28126ed4-3627-4b81-ab9c-0a91bbeea6cf tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146116, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2049.958116] env[62824]: DEBUG oslo_vmware.api [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2146120, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2049.968379] env[62824]: INFO nova.compute.manager [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 21c05725-3dad-4965-98d0-0622ebcc8ece] Rebuilding instance [ 2050.054306] env[62824]: DEBUG nova.compute.manager [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 21c05725-3dad-4965-98d0-0622ebcc8ece] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2050.055523] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa542af6-cf61-4640-b0d3-0f230524482c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2050.090653] env[62824]: DEBUG oslo_vmware.api [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Task: {'id': task-2146121, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2050.176905] env[62824]: DEBUG oslo_vmware.rw_handles [None req-457a3e46-ac18-4c02-91aa-31023cbb56f4 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e5ec0c-f672-1439-f6cb-723be3e13a42/disk-0.vmdk. {{(pid=62824) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 2050.179762] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e214ea5c-e375-460d-89e4-fea485fda770 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2050.185343] env[62824]: DEBUG oslo_vmware.rw_handles [None req-457a3e46-ac18-4c02-91aa-31023cbb56f4 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e5ec0c-f672-1439-f6cb-723be3e13a42/disk-0.vmdk is in state: ready. {{(pid=62824) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 2050.185541] env[62824]: ERROR oslo_vmware.rw_handles [None req-457a3e46-ac18-4c02-91aa-31023cbb56f4 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e5ec0c-f672-1439-f6cb-723be3e13a42/disk-0.vmdk due to incomplete transfer. [ 2050.185811] env[62824]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-8a2d7fea-70a5-4f72-ae02-5efce4fd0743 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2050.188094] env[62824]: DEBUG nova.scheduler.client.report [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 173, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2050.202104] env[62824]: DEBUG oslo_vmware.rw_handles [None req-457a3e46-ac18-4c02-91aa-31023cbb56f4 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52e5ec0c-f672-1439-f6cb-723be3e13a42/disk-0.vmdk. {{(pid=62824) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 2050.202104] env[62824]: DEBUG nova.virt.vmwareapi.images [None req-457a3e46-ac18-4c02-91aa-31023cbb56f4 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Uploaded image 3ad686d1-b5e5-42bb-af64-cda5329ea3f2 to the Glance image server {{(pid=62824) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 2050.204807] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-457a3e46-ac18-4c02-91aa-31023cbb56f4 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Destroying the VM {{(pid=62824) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 2050.205182] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-32ee914e-f570-4241-b305-21a89d0793ff {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2050.213701] env[62824]: DEBUG oslo_vmware.api [None req-457a3e46-ac18-4c02-91aa-31023cbb56f4 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Waiting for the task: (returnval){ [ 2050.213701] env[62824]: value = "task-2146122" [ 2050.213701] env[62824]: _type = "Task" [ 2050.213701] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2050.224907] env[62824]: DEBUG oslo_vmware.api [None req-457a3e46-ac18-4c02-91aa-31023cbb56f4 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2146122, 'name': Destroy_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2050.347948] env[62824]: DEBUG nova.compute.manager [None req-9a50fcc7-25bf-46e5-b936-5cbc1f3c424b tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] [instance: c63b05b7-1dbf-4637-8728-33673be0d3ea] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2050.348234] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-9a50fcc7-25bf-46e5-b936-5cbc1f3c424b tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] [instance: c63b05b7-1dbf-4637-8728-33673be0d3ea] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2050.349549] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe31581d-6756-4040-ba7f-7219509ce08d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2050.360129] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a50fcc7-25bf-46e5-b936-5cbc1f3c424b tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] [instance: c63b05b7-1dbf-4637-8728-33673be0d3ea] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2050.361460] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5dfffeb5-e7e0-48dc-8eb7-69baeb0ab74d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2050.372728] env[62824]: DEBUG oslo_vmware.api [None req-9a50fcc7-25bf-46e5-b936-5cbc1f3c424b tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Waiting for the task: (returnval){ [ 2050.372728] env[62824]: value = "task-2146123" [ 2050.372728] env[62824]: _type = "Task" [ 2050.372728] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2050.384460] env[62824]: DEBUG oslo_vmware.api [None req-9a50fcc7-25bf-46e5-b936-5cbc1f3c424b tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Task: {'id': task-2146123, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2050.397958] env[62824]: DEBUG nova.network.neutron [None req-cd4b7b59-8657-4fb4-94f8-f660379978d7 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2050.421613] env[62824]: DEBUG oslo_vmware.api [None req-28126ed4-3627-4b81-ab9c-0a91bbeea6cf tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146116, 'name': RemoveSnapshot_Task, 'duration_secs': 1.191877} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2050.421892] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-28126ed4-3627-4b81-ab9c-0a91bbeea6cf tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: c65ccd20-b682-420a-9c1a-47a45959197e] Deleted Snapshot of the VM instance {{(pid=62824) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 2050.457500] env[62824]: DEBUG oslo_vmware.api [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2146120, 'name': CreateSnapshot_Task, 'duration_secs': 1.01753} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2050.457500] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Created Snapshot of the VM instance {{(pid=62824) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 2050.459649] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19e2b1d7-62f1-4d79-a84c-a2368969ea97 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2050.582308] env[62824]: DEBUG oslo_vmware.api [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Task: {'id': task-2146121, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.556565} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2050.582648] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] ed7d3e95-f26e-40fd-a370-e17922bbff8e/ed7d3e95-f26e-40fd-a370-e17922bbff8e.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2050.583026] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] [instance: ed7d3e95-f26e-40fd-a370-e17922bbff8e] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2050.583384] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-934b73df-87a0-4be6-bc90-e6649bbd4389 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2050.591671] env[62824]: DEBUG oslo_vmware.api [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Waiting for the task: (returnval){ [ 2050.591671] env[62824]: value = "task-2146124" [ 2050.591671] env[62824]: _type = "Task" [ 2050.591671] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2050.601766] env[62824]: DEBUG oslo_vmware.api [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Task: {'id': task-2146124, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2050.665986] env[62824]: DEBUG nova.compute.manager [req-6cb17be6-e1d1-4cd3-a2ff-709e1080b227 req-b4479a12-5259-4b92-bf02-7089c1e4a685 service nova] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Received event network-changed-b27bda53-5f70-40ff-bb04-9b22f2241093 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2050.666438] env[62824]: DEBUG nova.compute.manager [req-6cb17be6-e1d1-4cd3-a2ff-709e1080b227 req-b4479a12-5259-4b92-bf02-7089c1e4a685 service nova] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Refreshing instance network info cache due to event network-changed-b27bda53-5f70-40ff-bb04-9b22f2241093. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2050.666849] env[62824]: DEBUG oslo_concurrency.lockutils [req-6cb17be6-e1d1-4cd3-a2ff-709e1080b227 req-b4479a12-5259-4b92-bf02-7089c1e4a685 service nova] Acquiring lock "refresh_cache-7febb8b4-f984-4d79-a888-b2829f2a9df6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2050.696329] env[62824]: DEBUG oslo_concurrency.lockutils [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.784s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2050.696329] env[62824]: DEBUG nova.compute.manager [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: b0456b8a-348a-4503-a92c-58e5ab455d1c] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2050.697781] env[62824]: DEBUG oslo_concurrency.lockutils [None req-6c2d5169-8d71-4209-a7d7-4b4ea2abbc8d tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.424s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2050.697995] env[62824]: DEBUG nova.objects.instance [None req-6c2d5169-8d71-4209-a7d7-4b4ea2abbc8d tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Lazy-loading 'resources' on Instance uuid db018af0-7bdf-40c5-b129-fde15dfb8ab4 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2050.726862] env[62824]: DEBUG oslo_vmware.api [None req-457a3e46-ac18-4c02-91aa-31023cbb56f4 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2146122, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2050.883884] env[62824]: DEBUG oslo_vmware.api [None req-9a50fcc7-25bf-46e5-b936-5cbc1f3c424b tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Task: {'id': task-2146123, 'name': PowerOffVM_Task, 'duration_secs': 0.318815} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2050.884239] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a50fcc7-25bf-46e5-b936-5cbc1f3c424b tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] [instance: c63b05b7-1dbf-4637-8728-33673be0d3ea] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2050.884418] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-9a50fcc7-25bf-46e5-b936-5cbc1f3c424b tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] [instance: c63b05b7-1dbf-4637-8728-33673be0d3ea] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2050.884684] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-274d6f41-b8e6-4a42-b56b-ac84795489e2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2050.927892] env[62824]: WARNING nova.compute.manager [None req-28126ed4-3627-4b81-ab9c-0a91bbeea6cf tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: c65ccd20-b682-420a-9c1a-47a45959197e] Image not found during snapshot: nova.exception.ImageNotFound: Image 6c8b574b-aa50-4be7-896f-fe8af1b20d01 could not be found. [ 2050.982162] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Creating linked-clone VM from snapshot {{(pid=62824) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 2050.983031] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-ed9c943a-69b9-4d2b-b586-bef61dd3d301 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2050.994470] env[62824]: DEBUG oslo_vmware.api [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Waiting for the task: (returnval){ [ 2050.994470] env[62824]: value = "task-2146126" [ 2050.994470] env[62824]: _type = "Task" [ 2050.994470] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2050.994775] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-9a50fcc7-25bf-46e5-b936-5cbc1f3c424b tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] [instance: c63b05b7-1dbf-4637-8728-33673be0d3ea] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2050.995333] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-9a50fcc7-25bf-46e5-b936-5cbc1f3c424b tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] [instance: c63b05b7-1dbf-4637-8728-33673be0d3ea] Deleting contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2050.995333] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-9a50fcc7-25bf-46e5-b936-5cbc1f3c424b tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Deleting the datastore file [datastore2] c63b05b7-1dbf-4637-8728-33673be0d3ea {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2050.995553] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ad47dcec-0d55-4dde-8587-789f6c98c2f1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2051.024804] env[62824]: DEBUG oslo_vmware.api [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2146126, 'name': CloneVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2051.027063] env[62824]: DEBUG oslo_vmware.api [None req-9a50fcc7-25bf-46e5-b936-5cbc1f3c424b tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Waiting for the task: (returnval){ [ 2051.027063] env[62824]: value = "task-2146127" [ 2051.027063] env[62824]: _type = "Task" [ 2051.027063] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2051.037423] env[62824]: DEBUG oslo_vmware.api [None req-9a50fcc7-25bf-46e5-b936-5cbc1f3c424b tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Task: {'id': task-2146127, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2051.082671] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 21c05725-3dad-4965-98d0-0622ebcc8ece] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2051.083836] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d197f84c-442a-4061-836e-b172b40e0d9a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2051.090885] env[62824]: DEBUG oslo_vmware.api [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Waiting for the task: (returnval){ [ 2051.090885] env[62824]: value = "task-2146128" [ 2051.090885] env[62824]: _type = "Task" [ 2051.090885] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2051.106948] env[62824]: DEBUG oslo_vmware.api [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Task: {'id': task-2146124, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.29872} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2051.110744] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] [instance: ed7d3e95-f26e-40fd-a370-e17922bbff8e] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2051.111202] env[62824]: DEBUG oslo_vmware.api [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Task: {'id': task-2146128, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2051.112456] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b839e041-3936-49b2-8dc4-a3d7fba8d56f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2051.138160] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] [instance: ed7d3e95-f26e-40fd-a370-e17922bbff8e] Reconfiguring VM instance instance-00000037 to attach disk [datastore2] ed7d3e95-f26e-40fd-a370-e17922bbff8e/ed7d3e95-f26e-40fd-a370-e17922bbff8e.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2051.138524] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6e840655-dcf1-48b9-8abb-8bf12a742df2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2051.162318] env[62824]: DEBUG oslo_vmware.api [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Waiting for the task: (returnval){ [ 2051.162318] env[62824]: value = "task-2146130" [ 2051.162318] env[62824]: _type = "Task" [ 2051.162318] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2051.172190] env[62824]: DEBUG oslo_vmware.api [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Task: {'id': task-2146130, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2051.202258] env[62824]: DEBUG nova.compute.utils [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2051.208024] env[62824]: DEBUG nova.compute.manager [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: b0456b8a-348a-4503-a92c-58e5ab455d1c] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2051.208024] env[62824]: DEBUG nova.network.neutron [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: b0456b8a-348a-4503-a92c-58e5ab455d1c] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2051.228716] env[62824]: DEBUG oslo_vmware.api [None req-457a3e46-ac18-4c02-91aa-31023cbb56f4 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2146122, 'name': Destroy_Task, 'duration_secs': 0.76627} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2051.229131] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-457a3e46-ac18-4c02-91aa-31023cbb56f4 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Destroyed the VM [ 2051.229500] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-457a3e46-ac18-4c02-91aa-31023cbb56f4 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Deleting Snapshot of the VM instance {{(pid=62824) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 2051.229836] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-4e723648-f772-4e04-968a-dd9d46affe2f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2051.238319] env[62824]: DEBUG oslo_vmware.api [None req-457a3e46-ac18-4c02-91aa-31023cbb56f4 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Waiting for the task: (returnval){ [ 2051.238319] env[62824]: value = "task-2146131" [ 2051.238319] env[62824]: _type = "Task" [ 2051.238319] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2051.253433] env[62824]: DEBUG oslo_vmware.api [None req-457a3e46-ac18-4c02-91aa-31023cbb56f4 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2146131, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2051.279570] env[62824]: DEBUG nova.network.neutron [None req-cd4b7b59-8657-4fb4-94f8-f660379978d7 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Updating instance_info_cache with network_info: [{"id": "b27bda53-5f70-40ff-bb04-9b22f2241093", "address": "fa:16:3e:4d:09:75", "network": {"id": "3d2f0f85-4d51-4459-ad99-443b0821fd9d", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1138734647-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "adf40520c1ef4d97b1ffe2c8cc9649b9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "10b81051-1eb1-406b-888c-4548c470c77e", "external-id": "nsx-vlan-transportzone-207", "segmentation_id": 207, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb27bda53-5f", "ovs_interfaceid": "b27bda53-5f70-40ff-bb04-9b22f2241093", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2051.321022] env[62824]: DEBUG nova.policy [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1ba0f76cb10d41939831abe409efb30b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0503b434636c4ef5bae8db1b0c74d2ce', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2051.510430] env[62824]: DEBUG oslo_vmware.api [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2146126, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2051.537756] env[62824]: DEBUG oslo_vmware.api [None req-9a50fcc7-25bf-46e5-b936-5cbc1f3c424b tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Task: {'id': task-2146127, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.159672} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2051.538400] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-9a50fcc7-25bf-46e5-b936-5cbc1f3c424b tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2051.538591] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-9a50fcc7-25bf-46e5-b936-5cbc1f3c424b tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] [instance: c63b05b7-1dbf-4637-8728-33673be0d3ea] Deleted contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2051.538768] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-9a50fcc7-25bf-46e5-b936-5cbc1f3c424b tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] [instance: c63b05b7-1dbf-4637-8728-33673be0d3ea] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2051.538938] env[62824]: INFO nova.compute.manager [None req-9a50fcc7-25bf-46e5-b936-5cbc1f3c424b tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] [instance: c63b05b7-1dbf-4637-8728-33673be0d3ea] Took 1.19 seconds to destroy the instance on the hypervisor. [ 2051.539219] env[62824]: DEBUG oslo.service.loopingcall [None req-9a50fcc7-25bf-46e5-b936-5cbc1f3c424b tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2051.539404] env[62824]: DEBUG nova.compute.manager [-] [instance: c63b05b7-1dbf-4637-8728-33673be0d3ea] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2051.539495] env[62824]: DEBUG nova.network.neutron [-] [instance: c63b05b7-1dbf-4637-8728-33673be0d3ea] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2051.605493] env[62824]: DEBUG oslo_vmware.api [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Task: {'id': task-2146128, 'name': PowerOffVM_Task, 'duration_secs': 0.154596} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2051.608042] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 21c05725-3dad-4965-98d0-0622ebcc8ece] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2051.608042] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 21c05725-3dad-4965-98d0-0622ebcc8ece] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2051.608042] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1e9f208-33a4-4bc5-89ca-275a940022ee {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2051.617942] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 21c05725-3dad-4965-98d0-0622ebcc8ece] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2051.618232] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d8aad038-adae-475b-9703-cecab9ad9052 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2051.654421] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 21c05725-3dad-4965-98d0-0622ebcc8ece] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2051.654421] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 21c05725-3dad-4965-98d0-0622ebcc8ece] Deleting contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2051.654540] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Deleting the datastore file [datastore2] 21c05725-3dad-4965-98d0-0622ebcc8ece {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2051.654776] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3fc3346d-c80c-4b79-bdfd-9362ffd0bf07 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2051.662872] env[62824]: DEBUG oslo_vmware.api [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Waiting for the task: (returnval){ [ 2051.662872] env[62824]: value = "task-2146133" [ 2051.662872] env[62824]: _type = "Task" [ 2051.662872] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2051.678982] env[62824]: DEBUG oslo_vmware.api [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Task: {'id': task-2146133, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2051.684147] env[62824]: DEBUG oslo_vmware.api [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Task: {'id': task-2146130, 'name': ReconfigVM_Task, 'duration_secs': 0.353984} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2051.686102] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] [instance: ed7d3e95-f26e-40fd-a370-e17922bbff8e] Reconfigured VM instance instance-00000037 to attach disk [datastore2] ed7d3e95-f26e-40fd-a370-e17922bbff8e/ed7d3e95-f26e-40fd-a370-e17922bbff8e.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2051.686102] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bc7f46fb-138d-4563-b172-cfc3376414a0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2051.693045] env[62824]: DEBUG oslo_vmware.api [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Waiting for the task: (returnval){ [ 2051.693045] env[62824]: value = "task-2146134" [ 2051.693045] env[62824]: _type = "Task" [ 2051.693045] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2051.709317] env[62824]: DEBUG oslo_vmware.api [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Task: {'id': task-2146134, 'name': Rename_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2051.710233] env[62824]: DEBUG nova.compute.manager [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: b0456b8a-348a-4503-a92c-58e5ab455d1c] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2051.751731] env[62824]: DEBUG oslo_vmware.api [None req-457a3e46-ac18-4c02-91aa-31023cbb56f4 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2146131, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2051.780029] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bf6ca20b-2439-4b4d-ba93-061898ce061b tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Acquiring lock "35f0435c-fe46-46ab-bc71-c21bb0ad3d55" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2051.780029] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bf6ca20b-2439-4b4d-ba93-061898ce061b tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Lock "35f0435c-fe46-46ab-bc71-c21bb0ad3d55" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2051.780029] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bf6ca20b-2439-4b4d-ba93-061898ce061b tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Acquiring lock "35f0435c-fe46-46ab-bc71-c21bb0ad3d55-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2051.780029] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bf6ca20b-2439-4b4d-ba93-061898ce061b tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Lock "35f0435c-fe46-46ab-bc71-c21bb0ad3d55-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2051.780029] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bf6ca20b-2439-4b4d-ba93-061898ce061b tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Lock "35f0435c-fe46-46ab-bc71-c21bb0ad3d55-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2051.782792] env[62824]: DEBUG oslo_concurrency.lockutils [None req-cd4b7b59-8657-4fb4-94f8-f660379978d7 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Releasing lock "refresh_cache-7febb8b4-f984-4d79-a888-b2829f2a9df6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2051.783069] env[62824]: DEBUG nova.compute.manager [None req-cd4b7b59-8657-4fb4-94f8-f660379978d7 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Inject network info {{(pid=62824) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7737}} [ 2051.783386] env[62824]: DEBUG nova.compute.manager [None req-cd4b7b59-8657-4fb4-94f8-f660379978d7 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] network_info to inject: |[{"id": "b27bda53-5f70-40ff-bb04-9b22f2241093", "address": "fa:16:3e:4d:09:75", "network": {"id": "3d2f0f85-4d51-4459-ad99-443b0821fd9d", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1138734647-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "adf40520c1ef4d97b1ffe2c8cc9649b9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "10b81051-1eb1-406b-888c-4548c470c77e", "external-id": "nsx-vlan-transportzone-207", "segmentation_id": 207, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb27bda53-5f", "ovs_interfaceid": "b27bda53-5f70-40ff-bb04-9b22f2241093", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7738}} [ 2051.788456] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-cd4b7b59-8657-4fb4-94f8-f660379978d7 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Reconfiguring VM instance to set the machine id {{(pid=62824) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1796}} [ 2051.789097] env[62824]: INFO nova.compute.manager [None req-bf6ca20b-2439-4b4d-ba93-061898ce061b tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Terminating instance [ 2051.791281] env[62824]: DEBUG oslo_concurrency.lockutils [req-6cb17be6-e1d1-4cd3-a2ff-709e1080b227 req-b4479a12-5259-4b92-bf02-7089c1e4a685 service nova] Acquired lock "refresh_cache-7febb8b4-f984-4d79-a888-b2829f2a9df6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2051.791479] env[62824]: DEBUG nova.network.neutron [req-6cb17be6-e1d1-4cd3-a2ff-709e1080b227 req-b4479a12-5259-4b92-bf02-7089c1e4a685 service nova] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Refreshing network info cache for port b27bda53-5f70-40ff-bb04-9b22f2241093 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2051.792925] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-15c7d7f8-674e-4c81-b9e6-3542c53f3b23 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2051.810761] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5b6c437-6c3a-4a5e-bc60-28b5cf8cbfe1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2051.820556] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aef24138-8608-474f-91ae-6ee8a2042db0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2051.827712] env[62824]: DEBUG oslo_vmware.api [None req-cd4b7b59-8657-4fb4-94f8-f660379978d7 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Waiting for the task: (returnval){ [ 2051.827712] env[62824]: value = "task-2146135" [ 2051.827712] env[62824]: _type = "Task" [ 2051.827712] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2051.862177] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4763b3a-b994-45b4-9900-aa1e9ea73043 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2051.870445] env[62824]: DEBUG oslo_vmware.api [None req-cd4b7b59-8657-4fb4-94f8-f660379978d7 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Task: {'id': task-2146135, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2051.876384] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d9e7eae-cf17-492c-8ebd-90dcffd6fb69 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2051.891683] env[62824]: DEBUG nova.compute.provider_tree [None req-6c2d5169-8d71-4209-a7d7-4b4ea2abbc8d tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2052.002740] env[62824]: DEBUG nova.network.neutron [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: b0456b8a-348a-4503-a92c-58e5ab455d1c] Successfully created port: 0534c02e-d202-458d-9289-7d0ad2c5596f {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2052.010794] env[62824]: DEBUG oslo_vmware.api [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2146126, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2052.104877] env[62824]: DEBUG nova.network.neutron [req-6cb17be6-e1d1-4cd3-a2ff-709e1080b227 req-b4479a12-5259-4b92-bf02-7089c1e4a685 service nova] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Updated VIF entry in instance network info cache for port b27bda53-5f70-40ff-bb04-9b22f2241093. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2052.105320] env[62824]: DEBUG nova.network.neutron [req-6cb17be6-e1d1-4cd3-a2ff-709e1080b227 req-b4479a12-5259-4b92-bf02-7089c1e4a685 service nova] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Updating instance_info_cache with network_info: [{"id": "b27bda53-5f70-40ff-bb04-9b22f2241093", "address": "fa:16:3e:4d:09:75", "network": {"id": "3d2f0f85-4d51-4459-ad99-443b0821fd9d", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1138734647-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "adf40520c1ef4d97b1ffe2c8cc9649b9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "10b81051-1eb1-406b-888c-4548c470c77e", "external-id": "nsx-vlan-transportzone-207", "segmentation_id": 207, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb27bda53-5f", "ovs_interfaceid": "b27bda53-5f70-40ff-bb04-9b22f2241093", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2052.179179] env[62824]: DEBUG oslo_vmware.api [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Task: {'id': task-2146133, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.116161} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2052.179958] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2052.180174] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 21c05725-3dad-4965-98d0-0622ebcc8ece] Deleted contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2052.180430] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 21c05725-3dad-4965-98d0-0622ebcc8ece] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2052.185925] env[62824]: DEBUG nova.compute.manager [None req-752f903a-bdbd-4c93-8465-c657fedd4220 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 88bad2dd-dce7-41df-b56c-93a5d054c11e] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2052.186952] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-228ffcf4-b470-4181-bb3a-445dc15310b9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2052.204829] env[62824]: DEBUG oslo_vmware.api [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Task: {'id': task-2146134, 'name': Rename_Task, 'duration_secs': 0.187959} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2052.205329] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] [instance: ed7d3e95-f26e-40fd-a370-e17922bbff8e] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2052.205592] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-379ab686-8ecd-4a97-8746-b5cb25c1f8dd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2052.212840] env[62824]: DEBUG oslo_vmware.api [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Waiting for the task: (returnval){ [ 2052.212840] env[62824]: value = "task-2146136" [ 2052.212840] env[62824]: _type = "Task" [ 2052.212840] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2052.224760] env[62824]: DEBUG oslo_vmware.api [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Task: {'id': task-2146136, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2052.250365] env[62824]: DEBUG oslo_vmware.api [None req-457a3e46-ac18-4c02-91aa-31023cbb56f4 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2146131, 'name': RemoveSnapshot_Task, 'duration_secs': 0.840504} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2052.250622] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-457a3e46-ac18-4c02-91aa-31023cbb56f4 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Deleted Snapshot of the VM instance {{(pid=62824) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 2052.250864] env[62824]: INFO nova.compute.manager [None req-457a3e46-ac18-4c02-91aa-31023cbb56f4 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Took 16.24 seconds to snapshot the instance on the hypervisor. [ 2052.314775] env[62824]: DEBUG nova.compute.manager [None req-bf6ca20b-2439-4b4d-ba93-061898ce061b tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2052.315042] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-bf6ca20b-2439-4b4d-ba93-061898ce061b tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2052.316181] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fba3a4fb-a91f-4c02-85ef-a787dc862c51 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2052.324839] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf6ca20b-2439-4b4d-ba93-061898ce061b tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2052.325099] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a66a7861-a24f-4af5-ab87-06e3bb8cbd6e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2052.334136] env[62824]: DEBUG oslo_vmware.api [None req-bf6ca20b-2439-4b4d-ba93-061898ce061b tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Waiting for the task: (returnval){ [ 2052.334136] env[62824]: value = "task-2146137" [ 2052.334136] env[62824]: _type = "Task" [ 2052.334136] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2052.341427] env[62824]: DEBUG oslo_vmware.api [None req-cd4b7b59-8657-4fb4-94f8-f660379978d7 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Task: {'id': task-2146135, 'name': ReconfigVM_Task, 'duration_secs': 0.15708} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2052.342156] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-cd4b7b59-8657-4fb4-94f8-f660379978d7 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Reconfigured VM instance to set the machine id {{(pid=62824) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1799}} [ 2052.349124] env[62824]: DEBUG oslo_vmware.api [None req-bf6ca20b-2439-4b4d-ba93-061898ce061b tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Task: {'id': task-2146137, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2052.395503] env[62824]: DEBUG nova.scheduler.client.report [None req-6c2d5169-8d71-4209-a7d7-4b4ea2abbc8d tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 173, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2052.507644] env[62824]: DEBUG oslo_vmware.api [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2146126, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2052.605938] env[62824]: DEBUG nova.network.neutron [-] [instance: c63b05b7-1dbf-4637-8728-33673be0d3ea] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2052.609132] env[62824]: DEBUG oslo_concurrency.lockutils [req-6cb17be6-e1d1-4cd3-a2ff-709e1080b227 req-b4479a12-5259-4b92-bf02-7089c1e4a685 service nova] Releasing lock "refresh_cache-7febb8b4-f984-4d79-a888-b2829f2a9df6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2052.673256] env[62824]: DEBUG oslo_concurrency.lockutils [None req-02170dc2-e93c-4f62-a015-5cb11263272d tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Acquiring lock "c65ccd20-b682-420a-9c1a-47a45959197e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2052.673256] env[62824]: DEBUG oslo_concurrency.lockutils [None req-02170dc2-e93c-4f62-a015-5cb11263272d tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Lock "c65ccd20-b682-420a-9c1a-47a45959197e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2052.673256] env[62824]: DEBUG oslo_concurrency.lockutils [None req-02170dc2-e93c-4f62-a015-5cb11263272d tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Acquiring lock "c65ccd20-b682-420a-9c1a-47a45959197e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2052.673256] env[62824]: DEBUG oslo_concurrency.lockutils [None req-02170dc2-e93c-4f62-a015-5cb11263272d tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Lock "c65ccd20-b682-420a-9c1a-47a45959197e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2052.673256] env[62824]: DEBUG oslo_concurrency.lockutils [None req-02170dc2-e93c-4f62-a015-5cb11263272d tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Lock "c65ccd20-b682-420a-9c1a-47a45959197e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2052.674844] env[62824]: INFO nova.compute.manager [None req-02170dc2-e93c-4f62-a015-5cb11263272d tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: c65ccd20-b682-420a-9c1a-47a45959197e] Terminating instance [ 2052.702033] env[62824]: INFO nova.compute.manager [None req-752f903a-bdbd-4c93-8465-c657fedd4220 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 88bad2dd-dce7-41df-b56c-93a5d054c11e] instance snapshotting [ 2052.702033] env[62824]: WARNING nova.compute.manager [None req-752f903a-bdbd-4c93-8465-c657fedd4220 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 88bad2dd-dce7-41df-b56c-93a5d054c11e] trying to snapshot a non-running instance: (state: 7 expected: 1) [ 2052.704546] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8672038-d955-4a4b-9930-ef6584265ada {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2052.730168] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdeb7407-8d2a-4fab-9aa4-8c13c90acfde {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2052.734480] env[62824]: DEBUG nova.compute.manager [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: b0456b8a-348a-4503-a92c-58e5ab455d1c] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2052.748767] env[62824]: DEBUG oslo_vmware.api [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Task: {'id': task-2146136, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2052.763408] env[62824]: DEBUG nova.virt.hardware [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2052.763650] env[62824]: DEBUG nova.virt.hardware [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2052.763809] env[62824]: DEBUG nova.virt.hardware [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2052.763992] env[62824]: DEBUG nova.virt.hardware [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2052.764208] env[62824]: DEBUG nova.virt.hardware [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2052.764370] env[62824]: DEBUG nova.virt.hardware [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2052.764628] env[62824]: DEBUG nova.virt.hardware [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2052.764810] env[62824]: DEBUG nova.virt.hardware [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2052.765112] env[62824]: DEBUG nova.virt.hardware [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2052.766290] env[62824]: DEBUG nova.virt.hardware [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2052.766290] env[62824]: DEBUG nova.virt.hardware [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2052.766392] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c7c5f15-975b-43fc-a89a-6f21e664506b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2052.776151] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f729231a-e7d6-4542-a79c-9958cb5d1d93 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2052.845398] env[62824]: DEBUG oslo_vmware.api [None req-bf6ca20b-2439-4b4d-ba93-061898ce061b tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Task: {'id': task-2146137, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2052.888940] env[62824]: DEBUG nova.compute.manager [req-1e68b5f6-d00f-4cb3-9c8a-1a95e60548d1 req-7e7c65d5-dbfc-40ca-964b-18a438d4dfe2 service nova] [instance: c63b05b7-1dbf-4637-8728-33673be0d3ea] Received event network-vif-deleted-a0edea0b-d2cc-4b03-9d46-9cc5e6cd1794 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2052.900947] env[62824]: DEBUG oslo_concurrency.lockutils [None req-6c2d5169-8d71-4209-a7d7-4b4ea2abbc8d tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.203s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2052.904430] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a1ef20b1-87c8-4499-bf60-63e3997b1e64 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.212s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2052.904741] env[62824]: DEBUG nova.objects.instance [None req-a1ef20b1-87c8-4499-bf60-63e3997b1e64 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Lazy-loading 'resources' on Instance uuid 41f068d3-2c8d-46f0-8d84-78531d28b0dd {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2052.928998] env[62824]: INFO nova.scheduler.client.report [None req-6c2d5169-8d71-4209-a7d7-4b4ea2abbc8d tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Deleted allocations for instance db018af0-7bdf-40c5-b129-fde15dfb8ab4 [ 2053.018175] env[62824]: DEBUG oslo_vmware.api [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2146126, 'name': CloneVM_Task} progress is 95%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2053.110394] env[62824]: INFO nova.compute.manager [-] [instance: c63b05b7-1dbf-4637-8728-33673be0d3ea] Took 1.57 seconds to deallocate network for instance. [ 2053.179958] env[62824]: DEBUG nova.compute.manager [None req-02170dc2-e93c-4f62-a015-5cb11263272d tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: c65ccd20-b682-420a-9c1a-47a45959197e] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2053.180329] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-02170dc2-e93c-4f62-a015-5cb11263272d tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: c65ccd20-b682-420a-9c1a-47a45959197e] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2053.184889] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0021aab6-72f4-4cb2-b6f6-374a08cd20a0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2053.200093] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-02170dc2-e93c-4f62-a015-5cb11263272d tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: c65ccd20-b682-420a-9c1a-47a45959197e] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2053.200383] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f1f54824-c2d5-4f46-8e2b-33b5af120b01 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2053.208776] env[62824]: DEBUG oslo_vmware.api [None req-02170dc2-e93c-4f62-a015-5cb11263272d tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Waiting for the task: (returnval){ [ 2053.208776] env[62824]: value = "task-2146138" [ 2053.208776] env[62824]: _type = "Task" [ 2053.208776] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2053.220990] env[62824]: DEBUG oslo_vmware.api [None req-02170dc2-e93c-4f62-a015-5cb11263272d tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146138, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2053.224924] env[62824]: DEBUG nova.virt.hardware [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2053.225295] env[62824]: DEBUG nova.virt.hardware [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2053.225448] env[62824]: DEBUG nova.virt.hardware [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2053.225702] env[62824]: DEBUG nova.virt.hardware [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2053.225911] env[62824]: DEBUG nova.virt.hardware [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2053.226129] env[62824]: DEBUG nova.virt.hardware [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2053.226410] env[62824]: DEBUG nova.virt.hardware [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2053.226698] env[62824]: DEBUG nova.virt.hardware [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2053.227029] env[62824]: DEBUG nova.virt.hardware [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2053.227337] env[62824]: DEBUG nova.virt.hardware [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2053.227623] env[62824]: DEBUG nova.virt.hardware [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2053.228955] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c97ff73-f067-4ade-98ef-85eb0eed4724 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2053.243539] env[62824]: DEBUG oslo_vmware.api [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Task: {'id': task-2146136, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2053.244769] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b50ffc9-97e7-4ca5-b87a-2251acae5615 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2053.250395] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-752f903a-bdbd-4c93-8465-c657fedd4220 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 88bad2dd-dce7-41df-b56c-93a5d054c11e] Creating Snapshot of the VM instance {{(pid=62824) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 2053.250644] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-8d640e03-75fe-49e3-b95a-f4ad7ef07adb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2053.261486] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 21c05725-3dad-4965-98d0-0622ebcc8ece] Instance VIF info [] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2053.267943] env[62824]: DEBUG oslo.service.loopingcall [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2053.268708] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 21c05725-3dad-4965-98d0-0622ebcc8ece] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2053.269035] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cc142778-4922-40c5-9b61-ab606552047e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2053.283283] env[62824]: DEBUG oslo_vmware.api [None req-752f903a-bdbd-4c93-8465-c657fedd4220 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the task: (returnval){ [ 2053.283283] env[62824]: value = "task-2146139" [ 2053.283283] env[62824]: _type = "Task" [ 2053.283283] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2053.289856] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2053.289856] env[62824]: value = "task-2146140" [ 2053.289856] env[62824]: _type = "Task" [ 2053.289856] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2053.295962] env[62824]: DEBUG oslo_vmware.api [None req-752f903a-bdbd-4c93-8465-c657fedd4220 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146139, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2053.301792] env[62824]: DEBUG oslo_concurrency.lockutils [None req-24b1fd6d-11b0-414f-a359-57219816e958 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Acquiring lock "7febb8b4-f984-4d79-a888-b2829f2a9df6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2053.302087] env[62824]: DEBUG oslo_concurrency.lockutils [None req-24b1fd6d-11b0-414f-a359-57219816e958 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Lock "7febb8b4-f984-4d79-a888-b2829f2a9df6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2053.302259] env[62824]: DEBUG oslo_concurrency.lockutils [None req-24b1fd6d-11b0-414f-a359-57219816e958 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Acquiring lock "7febb8b4-f984-4d79-a888-b2829f2a9df6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2053.302441] env[62824]: DEBUG oslo_concurrency.lockutils [None req-24b1fd6d-11b0-414f-a359-57219816e958 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Lock "7febb8b4-f984-4d79-a888-b2829f2a9df6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2053.302609] env[62824]: DEBUG oslo_concurrency.lockutils [None req-24b1fd6d-11b0-414f-a359-57219816e958 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Lock "7febb8b4-f984-4d79-a888-b2829f2a9df6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2053.304293] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146140, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2053.304918] env[62824]: INFO nova.compute.manager [None req-24b1fd6d-11b0-414f-a359-57219816e958 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Terminating instance [ 2053.346740] env[62824]: DEBUG oslo_vmware.api [None req-bf6ca20b-2439-4b4d-ba93-061898ce061b tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Task: {'id': task-2146137, 'name': PowerOffVM_Task, 'duration_secs': 0.82761} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2053.347543] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf6ca20b-2439-4b4d-ba93-061898ce061b tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2053.347797] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-bf6ca20b-2439-4b4d-ba93-061898ce061b tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2053.348151] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c8b0499b-79dd-4cb8-aa5a-5f966eb9dbbc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2053.354893] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-895a9f6d-c24d-4e89-b988-6024da95cb4d tempest-VolumesAssistedSnapshotsTest-163918615 tempest-VolumesAssistedSnapshotsTest-163918615-project-admin] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Volume attach. Driver type: vmdk {{(pid=62824) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 2053.355146] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-895a9f6d-c24d-4e89-b988-6024da95cb4d tempest-VolumesAssistedSnapshotsTest-163918615 tempest-VolumesAssistedSnapshotsTest-163918615-project-admin] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-438645', 'volume_id': 'f3a0e2f5-ded9-47fe-8fbe-8eeb061be84a', 'name': 'volume-f3a0e2f5-ded9-47fe-8fbe-8eeb061be84a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '29b8dd5f-1855-490c-a01b-54840073a753', 'attached_at': '', 'detached_at': '', 'volume_id': 'f3a0e2f5-ded9-47fe-8fbe-8eeb061be84a', 'serial': 'f3a0e2f5-ded9-47fe-8fbe-8eeb061be84a'} {{(pid=62824) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 2053.356110] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f68e81e-1c18-41c1-91db-6af0a074bb1d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2053.377632] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1757f463-feb1-432e-88e9-c76b80f08d75 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2053.408309] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-895a9f6d-c24d-4e89-b988-6024da95cb4d tempest-VolumesAssistedSnapshotsTest-163918615 tempest-VolumesAssistedSnapshotsTest-163918615-project-admin] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Reconfiguring VM instance instance-0000000d to attach disk [datastore2] volume-f3a0e2f5-ded9-47fe-8fbe-8eeb061be84a/volume-f3a0e2f5-ded9-47fe-8fbe-8eeb061be84a.vmdk or device None with type thin {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2053.410691] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-20df4e61-f76b-4bf4-8fff-38f9e21a469e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2053.430905] env[62824]: DEBUG oslo_vmware.api [None req-895a9f6d-c24d-4e89-b988-6024da95cb4d tempest-VolumesAssistedSnapshotsTest-163918615 tempest-VolumesAssistedSnapshotsTest-163918615-project-admin] Waiting for the task: (returnval){ [ 2053.430905] env[62824]: value = "task-2146142" [ 2053.430905] env[62824]: _type = "Task" [ 2053.430905] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2053.442906] env[62824]: DEBUG oslo_vmware.api [None req-895a9f6d-c24d-4e89-b988-6024da95cb4d tempest-VolumesAssistedSnapshotsTest-163918615 tempest-VolumesAssistedSnapshotsTest-163918615-project-admin] Task: {'id': task-2146142, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2053.446135] env[62824]: DEBUG oslo_concurrency.lockutils [None req-6c2d5169-8d71-4209-a7d7-4b4ea2abbc8d tempest-ImagesNegativeTestJSON-243972461 tempest-ImagesNegativeTestJSON-243972461-project-member] Lock "db018af0-7bdf-40c5-b129-fde15dfb8ab4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 35.627s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2053.494115] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-bf6ca20b-2439-4b4d-ba93-061898ce061b tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2053.495008] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-bf6ca20b-2439-4b4d-ba93-061898ce061b tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Deleting contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2053.495008] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf6ca20b-2439-4b4d-ba93-061898ce061b tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Deleting the datastore file [datastore2] 35f0435c-fe46-46ab-bc71-c21bb0ad3d55 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2053.495008] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-98ab76ed-01e4-4ccc-b46a-dfdff5baa320 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2053.504721] env[62824]: DEBUG oslo_vmware.api [None req-bf6ca20b-2439-4b4d-ba93-061898ce061b tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Waiting for the task: (returnval){ [ 2053.504721] env[62824]: value = "task-2146143" [ 2053.504721] env[62824]: _type = "Task" [ 2053.504721] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2053.513986] env[62824]: DEBUG oslo_vmware.api [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2146126, 'name': CloneVM_Task} progress is 100%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2053.520038] env[62824]: DEBUG oslo_vmware.api [None req-bf6ca20b-2439-4b4d-ba93-061898ce061b tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Task: {'id': task-2146143, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2053.619494] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9a50fcc7-25bf-46e5-b936-5cbc1f3c424b tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2053.691605] env[62824]: DEBUG oslo_concurrency.lockutils [None req-74d3df01-c7bb-4adb-8f79-9713d2a8d4b2 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquiring lock "interface-f4d63a93-23af-470c-b36a-662af81dc386-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2053.691987] env[62824]: DEBUG oslo_concurrency.lockutils [None req-74d3df01-c7bb-4adb-8f79-9713d2a8d4b2 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lock "interface-f4d63a93-23af-470c-b36a-662af81dc386-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2053.692438] env[62824]: DEBUG nova.objects.instance [None req-74d3df01-c7bb-4adb-8f79-9713d2a8d4b2 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lazy-loading 'flavor' on Instance uuid f4d63a93-23af-470c-b36a-662af81dc386 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2053.723730] env[62824]: DEBUG oslo_vmware.api [None req-02170dc2-e93c-4f62-a015-5cb11263272d tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146138, 'name': PowerOffVM_Task, 'duration_secs': 0.201571} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2053.724023] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-02170dc2-e93c-4f62-a015-5cb11263272d tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: c65ccd20-b682-420a-9c1a-47a45959197e] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2053.724193] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-02170dc2-e93c-4f62-a015-5cb11263272d tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: c65ccd20-b682-420a-9c1a-47a45959197e] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2053.724440] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5a6a9127-5ccb-47f9-be95-25b4b4e1b353 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2053.740620] env[62824]: DEBUG oslo_vmware.api [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Task: {'id': task-2146136, 'name': PowerOnVM_Task, 'duration_secs': 1.17631} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2053.741067] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] [instance: ed7d3e95-f26e-40fd-a370-e17922bbff8e] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2053.741067] env[62824]: INFO nova.compute.manager [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] [instance: ed7d3e95-f26e-40fd-a370-e17922bbff8e] Took 6.01 seconds to spawn the instance on the hypervisor. [ 2053.741246] env[62824]: DEBUG nova.compute.manager [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] [instance: ed7d3e95-f26e-40fd-a370-e17922bbff8e] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2053.741994] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-559dedf2-61dc-45f2-882a-c346fd333021 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2053.799192] env[62824]: DEBUG oslo_vmware.api [None req-752f903a-bdbd-4c93-8465-c657fedd4220 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146139, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2053.803919] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146140, 'name': CreateVM_Task, 'duration_secs': 0.345727} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2053.806947] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 21c05725-3dad-4965-98d0-0622ebcc8ece] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2053.808211] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2053.808211] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2053.808211] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2053.811828] env[62824]: DEBUG nova.compute.manager [None req-24b1fd6d-11b0-414f-a359-57219816e958 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2053.811828] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-24b1fd6d-11b0-414f-a359-57219816e958 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2053.811828] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8a4c3447-70ba-49d9-a7f6-c0ed2f4dc452 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2053.811828] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2e59cd0-f611-4a06-b67b-306d4739f655 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2053.820180] env[62824]: DEBUG oslo_vmware.api [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Waiting for the task: (returnval){ [ 2053.820180] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]526492ef-07fe-6c6b-674d-9c938a5602ad" [ 2053.820180] env[62824]: _type = "Task" [ 2053.820180] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2053.822771] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-24b1fd6d-11b0-414f-a359-57219816e958 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2053.828402] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d3e29794-6019-4d20-a9f6-59299f71f412 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2053.837828] env[62824]: DEBUG oslo_vmware.api [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]526492ef-07fe-6c6b-674d-9c938a5602ad, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2053.839299] env[62824]: DEBUG oslo_vmware.api [None req-24b1fd6d-11b0-414f-a359-57219816e958 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Waiting for the task: (returnval){ [ 2053.839299] env[62824]: value = "task-2146145" [ 2053.839299] env[62824]: _type = "Task" [ 2053.839299] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2053.854536] env[62824]: DEBUG oslo_vmware.api [None req-24b1fd6d-11b0-414f-a359-57219816e958 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Task: {'id': task-2146145, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2053.866026] env[62824]: DEBUG nova.network.neutron [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: b0456b8a-348a-4503-a92c-58e5ab455d1c] Successfully updated port: 0534c02e-d202-458d-9289-7d0ad2c5596f {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2053.950295] env[62824]: DEBUG oslo_vmware.api [None req-895a9f6d-c24d-4e89-b988-6024da95cb4d tempest-VolumesAssistedSnapshotsTest-163918615 tempest-VolumesAssistedSnapshotsTest-163918615-project-admin] Task: {'id': task-2146142, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2053.968391] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55263a9f-39a2-461d-864d-daac385c9cca {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2053.978943] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60d9b85d-dcb5-42b9-b304-aa6a6fe9863b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2054.031668] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da6e77c8-0290-4cf7-a74c-2c83a8575bd9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2054.042384] env[62824]: DEBUG oslo_vmware.api [None req-bf6ca20b-2439-4b4d-ba93-061898ce061b tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Task: {'id': task-2146143, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.196764} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2054.042623] env[62824]: DEBUG oslo_vmware.api [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2146126, 'name': CloneVM_Task} progress is 100%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2054.044764] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf6ca20b-2439-4b4d-ba93-061898ce061b tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2054.044955] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-bf6ca20b-2439-4b4d-ba93-061898ce061b tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Deleted contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2054.045151] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-bf6ca20b-2439-4b4d-ba93-061898ce061b tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2054.045404] env[62824]: INFO nova.compute.manager [None req-bf6ca20b-2439-4b4d-ba93-061898ce061b tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Took 1.73 seconds to destroy the instance on the hypervisor. [ 2054.045648] env[62824]: DEBUG oslo.service.loopingcall [None req-bf6ca20b-2439-4b4d-ba93-061898ce061b tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2054.045897] env[62824]: DEBUG nova.compute.manager [-] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2054.045994] env[62824]: DEBUG nova.network.neutron [-] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2054.048706] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e820f95e-eae2-4d40-94ca-60039b8e3b13 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2054.063877] env[62824]: DEBUG nova.compute.provider_tree [None req-a1ef20b1-87c8-4499-bf60-63e3997b1e64 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2054.261235] env[62824]: INFO nova.compute.manager [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] [instance: ed7d3e95-f26e-40fd-a370-e17922bbff8e] Took 44.77 seconds to build instance. [ 2054.276614] env[62824]: DEBUG nova.objects.instance [None req-74d3df01-c7bb-4adb-8f79-9713d2a8d4b2 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lazy-loading 'pci_requests' on Instance uuid f4d63a93-23af-470c-b36a-662af81dc386 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2054.295207] env[62824]: DEBUG oslo_vmware.api [None req-752f903a-bdbd-4c93-8465-c657fedd4220 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146139, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2054.333709] env[62824]: DEBUG oslo_vmware.api [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]526492ef-07fe-6c6b-674d-9c938a5602ad, 'name': SearchDatastore_Task, 'duration_secs': 0.018154} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2054.333982] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2054.334228] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 21c05725-3dad-4965-98d0-0622ebcc8ece] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2054.334472] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2054.334596] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2054.334794] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2054.335095] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5b0bb88b-cc03-4c2c-b518-ee2659f75fee {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2054.345798] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2054.345979] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2054.353215] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aaef131c-04a1-4fe5-a315-c0ea860494c0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2054.353668] env[62824]: DEBUG oslo_vmware.api [None req-24b1fd6d-11b0-414f-a359-57219816e958 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Task: {'id': task-2146145, 'name': PowerOffVM_Task, 'duration_secs': 0.285799} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2054.354605] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-24b1fd6d-11b0-414f-a359-57219816e958 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2054.354605] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-24b1fd6d-11b0-414f-a359-57219816e958 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2054.354786] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fb7cdc60-4ffe-4947-ba69-0e50d73b0b4e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2054.360068] env[62824]: DEBUG oslo_vmware.api [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Waiting for the task: (returnval){ [ 2054.360068] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5279a39d-2b63-4f81-3d3c-94413ab86f51" [ 2054.360068] env[62824]: _type = "Task" [ 2054.360068] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2054.367472] env[62824]: DEBUG oslo_vmware.api [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5279a39d-2b63-4f81-3d3c-94413ab86f51, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2054.371485] env[62824]: DEBUG oslo_concurrency.lockutils [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Acquiring lock "refresh_cache-b0456b8a-348a-4503-a92c-58e5ab455d1c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2054.371485] env[62824]: DEBUG oslo_concurrency.lockutils [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Acquired lock "refresh_cache-b0456b8a-348a-4503-a92c-58e5ab455d1c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2054.371485] env[62824]: DEBUG nova.network.neutron [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: b0456b8a-348a-4503-a92c-58e5ab455d1c] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2054.443351] env[62824]: DEBUG oslo_vmware.api [None req-895a9f6d-c24d-4e89-b988-6024da95cb4d tempest-VolumesAssistedSnapshotsTest-163918615 tempest-VolumesAssistedSnapshotsTest-163918615-project-admin] Task: {'id': task-2146142, 'name': ReconfigVM_Task, 'duration_secs': 0.541937} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2054.443647] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-895a9f6d-c24d-4e89-b988-6024da95cb4d tempest-VolumesAssistedSnapshotsTest-163918615 tempest-VolumesAssistedSnapshotsTest-163918615-project-admin] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Reconfigured VM instance instance-0000000d to attach disk [datastore2] volume-f3a0e2f5-ded9-47fe-8fbe-8eeb061be84a/volume-f3a0e2f5-ded9-47fe-8fbe-8eeb061be84a.vmdk or device None with type thin {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2054.448620] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6c0f2f9b-99ee-457b-aca7-ca792491e4e0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2054.466373] env[62824]: DEBUG oslo_vmware.api [None req-895a9f6d-c24d-4e89-b988-6024da95cb4d tempest-VolumesAssistedSnapshotsTest-163918615 tempest-VolumesAssistedSnapshotsTest-163918615-project-admin] Waiting for the task: (returnval){ [ 2054.466373] env[62824]: value = "task-2146147" [ 2054.466373] env[62824]: _type = "Task" [ 2054.466373] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2054.475062] env[62824]: DEBUG oslo_vmware.api [None req-895a9f6d-c24d-4e89-b988-6024da95cb4d tempest-VolumesAssistedSnapshotsTest-163918615 tempest-VolumesAssistedSnapshotsTest-163918615-project-admin] Task: {'id': task-2146147, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2054.520122] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-02170dc2-e93c-4f62-a015-5cb11263272d tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: c65ccd20-b682-420a-9c1a-47a45959197e] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2054.522028] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-02170dc2-e93c-4f62-a015-5cb11263272d tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: c65ccd20-b682-420a-9c1a-47a45959197e] Deleting contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2054.522028] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-02170dc2-e93c-4f62-a015-5cb11263272d tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Deleting the datastore file [datastore1] c65ccd20-b682-420a-9c1a-47a45959197e {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2054.522420] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-17b1907a-f85a-4242-9bb2-44115fd79845 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2054.525497] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-24b1fd6d-11b0-414f-a359-57219816e958 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2054.525965] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-24b1fd6d-11b0-414f-a359-57219816e958 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Deleting contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2054.526436] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-24b1fd6d-11b0-414f-a359-57219816e958 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Deleting the datastore file [datastore1] 7febb8b4-f984-4d79-a888-b2829f2a9df6 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2054.526968] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-445cfc2a-c1b2-4363-ad36-87a42a876fab {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2054.541724] env[62824]: DEBUG oslo_vmware.api [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2146126, 'name': CloneVM_Task, 'duration_secs': 3.540406} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2054.544784] env[62824]: INFO nova.virt.vmwareapi.vmops [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Created linked-clone VM from snapshot [ 2054.548020] env[62824]: DEBUG oslo_vmware.api [None req-02170dc2-e93c-4f62-a015-5cb11263272d tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Waiting for the task: (returnval){ [ 2054.548020] env[62824]: value = "task-2146149" [ 2054.548020] env[62824]: _type = "Task" [ 2054.548020] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2054.548020] env[62824]: DEBUG oslo_vmware.api [None req-24b1fd6d-11b0-414f-a359-57219816e958 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Waiting for the task: (returnval){ [ 2054.548020] env[62824]: value = "task-2146148" [ 2054.548020] env[62824]: _type = "Task" [ 2054.548020] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2054.548020] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b62f324a-a650-4b28-9a0f-4bb1e0b771a7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2054.571226] env[62824]: DEBUG oslo_vmware.api [None req-02170dc2-e93c-4f62-a015-5cb11263272d tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146149, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2054.572573] env[62824]: DEBUG nova.scheduler.client.report [None req-a1ef20b1-87c8-4499-bf60-63e3997b1e64 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 173, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2054.577501] env[62824]: DEBUG nova.virt.vmwareapi.images [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Uploading image cc65c6c8-036d-4ab6-a8a0-770a7853932b {{(pid=62824) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 2054.585018] env[62824]: DEBUG oslo_vmware.api [None req-24b1fd6d-11b0-414f-a359-57219816e958 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Task: {'id': task-2146148, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2054.607694] env[62824]: DEBUG oslo_vmware.rw_handles [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 2054.607694] env[62824]: value = "vm-438647" [ 2054.607694] env[62824]: _type = "VirtualMachine" [ 2054.607694] env[62824]: }. {{(pid=62824) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 2054.608385] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-5bc9ce30-1301-4152-8c4a-fa302b3e4990 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2054.616476] env[62824]: DEBUG oslo_vmware.rw_handles [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Lease: (returnval){ [ 2054.616476] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5224ea4d-d247-c3f7-6713-1886b6997f97" [ 2054.616476] env[62824]: _type = "HttpNfcLease" [ 2054.616476] env[62824]: } obtained for exporting VM: (result){ [ 2054.616476] env[62824]: value = "vm-438647" [ 2054.616476] env[62824]: _type = "VirtualMachine" [ 2054.616476] env[62824]: }. {{(pid=62824) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 2054.616736] env[62824]: DEBUG oslo_vmware.api [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Waiting for the lease: (returnval){ [ 2054.616736] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5224ea4d-d247-c3f7-6713-1886b6997f97" [ 2054.616736] env[62824]: _type = "HttpNfcLease" [ 2054.616736] env[62824]: } to be ready. {{(pid=62824) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 2054.626761] env[62824]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 2054.626761] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5224ea4d-d247-c3f7-6713-1886b6997f97" [ 2054.626761] env[62824]: _type = "HttpNfcLease" [ 2054.626761] env[62824]: } is initializing. {{(pid=62824) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 2054.762775] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8c9c6f92-5bea-4350-b952-9090a4187012 tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Lock "ed7d3e95-f26e-40fd-a370-e17922bbff8e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 73.038s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2054.779582] env[62824]: DEBUG nova.objects.base [None req-74d3df01-c7bb-4adb-8f79-9713d2a8d4b2 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=62824) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 2054.779829] env[62824]: DEBUG nova.network.neutron [None req-74d3df01-c7bb-4adb-8f79-9713d2a8d4b2 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: f4d63a93-23af-470c-b36a-662af81dc386] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2054.798016] env[62824]: DEBUG oslo_vmware.api [None req-752f903a-bdbd-4c93-8465-c657fedd4220 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146139, 'name': CreateSnapshot_Task, 'duration_secs': 1.046421} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2054.798366] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-752f903a-bdbd-4c93-8465-c657fedd4220 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 88bad2dd-dce7-41df-b56c-93a5d054c11e] Created Snapshot of the VM instance {{(pid=62824) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 2054.799326] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b985c747-8141-45f4-acb5-3bea5bef8e99 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2054.833445] env[62824]: DEBUG nova.policy [None req-74d3df01-c7bb-4adb-8f79-9713d2a8d4b2 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b989eb89498b45dab6604116870d2fa1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c987c5f844be42c99cc9d57a9ca84c31', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2054.868855] env[62824]: DEBUG oslo_vmware.api [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5279a39d-2b63-4f81-3d3c-94413ab86f51, 'name': SearchDatastore_Task, 'duration_secs': 0.02762} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2054.869623] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d58c12ad-bea4-46d9-8ec2-21523994648e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2054.876403] env[62824]: DEBUG oslo_vmware.api [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Waiting for the task: (returnval){ [ 2054.876403] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5266f74c-bc2a-684b-3a84-9d4882357001" [ 2054.876403] env[62824]: _type = "Task" [ 2054.876403] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2054.884402] env[62824]: DEBUG oslo_vmware.api [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5266f74c-bc2a-684b-3a84-9d4882357001, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2054.900219] env[62824]: DEBUG nova.network.neutron [-] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2054.902890] env[62824]: DEBUG nova.network.neutron [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: b0456b8a-348a-4503-a92c-58e5ab455d1c] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2054.982053] env[62824]: DEBUG oslo_vmware.api [None req-895a9f6d-c24d-4e89-b988-6024da95cb4d tempest-VolumesAssistedSnapshotsTest-163918615 tempest-VolumesAssistedSnapshotsTest-163918615-project-admin] Task: {'id': task-2146147, 'name': ReconfigVM_Task, 'duration_secs': 0.185701} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2054.982349] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-895a9f6d-c24d-4e89-b988-6024da95cb4d tempest-VolumesAssistedSnapshotsTest-163918615 tempest-VolumesAssistedSnapshotsTest-163918615-project-admin] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-438645', 'volume_id': 'f3a0e2f5-ded9-47fe-8fbe-8eeb061be84a', 'name': 'volume-f3a0e2f5-ded9-47fe-8fbe-8eeb061be84a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '29b8dd5f-1855-490c-a01b-54840073a753', 'attached_at': '', 'detached_at': '', 'volume_id': 'f3a0e2f5-ded9-47fe-8fbe-8eeb061be84a', 'serial': 'f3a0e2f5-ded9-47fe-8fbe-8eeb061be84a'} {{(pid=62824) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 2055.058773] env[62824]: DEBUG nova.network.neutron [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: b0456b8a-348a-4503-a92c-58e5ab455d1c] Updating instance_info_cache with network_info: [{"id": "0534c02e-d202-458d-9289-7d0ad2c5596f", "address": "fa:16:3e:b0:2c:b9", "network": {"id": "bcf30d77-cac1-4deb-971d-efb7c1fb8354", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1030043097-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "0503b434636c4ef5bae8db1b0c74d2ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3cc0a33d-17c0-4b87-b48f-413a87a4cc6a", "external-id": "nsx-vlan-transportzone-865", "segmentation_id": 865, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0534c02e-d2", "ovs_interfaceid": "0534c02e-d202-458d-9289-7d0ad2c5596f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2055.066217] env[62824]: DEBUG oslo_vmware.api [None req-02170dc2-e93c-4f62-a015-5cb11263272d tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146149, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.258108} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2055.069217] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-02170dc2-e93c-4f62-a015-5cb11263272d tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2055.069417] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-02170dc2-e93c-4f62-a015-5cb11263272d tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: c65ccd20-b682-420a-9c1a-47a45959197e] Deleted contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2055.069598] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-02170dc2-e93c-4f62-a015-5cb11263272d tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: c65ccd20-b682-420a-9c1a-47a45959197e] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2055.069809] env[62824]: INFO nova.compute.manager [None req-02170dc2-e93c-4f62-a015-5cb11263272d tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: c65ccd20-b682-420a-9c1a-47a45959197e] Took 1.89 seconds to destroy the instance on the hypervisor. [ 2055.070031] env[62824]: DEBUG oslo.service.loopingcall [None req-02170dc2-e93c-4f62-a015-5cb11263272d tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2055.070236] env[62824]: DEBUG oslo_vmware.api [None req-24b1fd6d-11b0-414f-a359-57219816e958 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Task: {'id': task-2146148, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.248823} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2055.070428] env[62824]: DEBUG nova.compute.manager [-] [instance: c65ccd20-b682-420a-9c1a-47a45959197e] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2055.070524] env[62824]: DEBUG nova.network.neutron [-] [instance: c65ccd20-b682-420a-9c1a-47a45959197e] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2055.072107] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-24b1fd6d-11b0-414f-a359-57219816e958 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2055.072290] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-24b1fd6d-11b0-414f-a359-57219816e958 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Deleted contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2055.072467] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-24b1fd6d-11b0-414f-a359-57219816e958 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2055.072631] env[62824]: INFO nova.compute.manager [None req-24b1fd6d-11b0-414f-a359-57219816e958 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Took 1.26 seconds to destroy the instance on the hypervisor. [ 2055.072840] env[62824]: DEBUG oslo.service.loopingcall [None req-24b1fd6d-11b0-414f-a359-57219816e958 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2055.073596] env[62824]: DEBUG nova.compute.manager [-] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2055.073693] env[62824]: DEBUG nova.network.neutron [-] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2055.086318] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a1ef20b1-87c8-4499-bf60-63e3997b1e64 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.183s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2055.088432] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 33.369s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2055.088611] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2055.088798] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62824) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 2055.089099] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.757s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2055.092710] env[62824]: INFO nova.compute.claims [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2055.094817] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a51a3b38-23a2-4b17-8911-9d234e53cb41 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2055.106352] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-328346ad-26ba-436d-b0d6-d566ad481035 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2055.114960] env[62824]: INFO nova.scheduler.client.report [None req-a1ef20b1-87c8-4499-bf60-63e3997b1e64 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Deleted allocations for instance 41f068d3-2c8d-46f0-8d84-78531d28b0dd [ 2055.135187] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df1c3f44-0e74-4814-851e-35c6e201e987 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2055.140925] env[62824]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 2055.140925] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5224ea4d-d247-c3f7-6713-1886b6997f97" [ 2055.140925] env[62824]: _type = "HttpNfcLease" [ 2055.140925] env[62824]: } is ready. {{(pid=62824) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 2055.140925] env[62824]: DEBUG oslo_vmware.rw_handles [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 2055.140925] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5224ea4d-d247-c3f7-6713-1886b6997f97" [ 2055.140925] env[62824]: _type = "HttpNfcLease" [ 2055.140925] env[62824]: }. {{(pid=62824) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 2055.141552] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1de5a6a-a826-4146-8b51-e01da3e9d97a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2055.150017] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d343a1ce-00c9-4e12-ae46-ccf44aa0a49d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2055.155523] env[62824]: DEBUG oslo_vmware.rw_handles [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52031a0f-3082-1103-10b0-ff5f977f397f/disk-0.vmdk from lease info. {{(pid=62824) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 2055.155900] env[62824]: DEBUG oslo_vmware.rw_handles [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52031a0f-3082-1103-10b0-ff5f977f397f/disk-0.vmdk for reading. {{(pid=62824) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 2055.194317] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=178779MB free_disk=173GB free_vcpus=48 pci_devices=None {{(pid=62824) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 2055.194317] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2055.248967] env[62824]: DEBUG oslo_concurrency.lockutils [None req-abb349ec-c188-4834-ae78-a5dd5c6e4af7 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Acquiring lock "4423d172-acdd-4c69-b3b8-ff166e1b8548" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2055.248967] env[62824]: DEBUG oslo_concurrency.lockutils [None req-abb349ec-c188-4834-ae78-a5dd5c6e4af7 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Lock "4423d172-acdd-4c69-b3b8-ff166e1b8548" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2055.248967] env[62824]: DEBUG oslo_concurrency.lockutils [None req-abb349ec-c188-4834-ae78-a5dd5c6e4af7 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Acquiring lock "4423d172-acdd-4c69-b3b8-ff166e1b8548-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2055.248967] env[62824]: DEBUG oslo_concurrency.lockutils [None req-abb349ec-c188-4834-ae78-a5dd5c6e4af7 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Lock "4423d172-acdd-4c69-b3b8-ff166e1b8548-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2055.248967] env[62824]: DEBUG oslo_concurrency.lockutils [None req-abb349ec-c188-4834-ae78-a5dd5c6e4af7 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Lock "4423d172-acdd-4c69-b3b8-ff166e1b8548-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2055.250695] env[62824]: INFO nova.compute.manager [None req-abb349ec-c188-4834-ae78-a5dd5c6e4af7 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] [instance: 4423d172-acdd-4c69-b3b8-ff166e1b8548] Terminating instance [ 2055.302240] env[62824]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-3cf70244-040f-4661-b59b-5eb821376573 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2055.318376] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-752f903a-bdbd-4c93-8465-c657fedd4220 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 88bad2dd-dce7-41df-b56c-93a5d054c11e] Creating linked-clone VM from snapshot {{(pid=62824) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 2055.320864] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-ea2b5f1c-f2cf-4f19-9ea6-32ee8667c4e4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2055.329168] env[62824]: DEBUG nova.network.neutron [None req-74d3df01-c7bb-4adb-8f79-9713d2a8d4b2 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Successfully created port: c34f030d-46da-4a7a-9703-5acb2e526d50 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2055.335109] env[62824]: DEBUG oslo_vmware.api [None req-752f903a-bdbd-4c93-8465-c657fedd4220 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the task: (returnval){ [ 2055.335109] env[62824]: value = "task-2146151" [ 2055.335109] env[62824]: _type = "Task" [ 2055.335109] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2055.349428] env[62824]: DEBUG oslo_vmware.api [None req-752f903a-bdbd-4c93-8465-c657fedd4220 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146151, 'name': CloneVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2055.390104] env[62824]: DEBUG oslo_vmware.api [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5266f74c-bc2a-684b-3a84-9d4882357001, 'name': SearchDatastore_Task, 'duration_secs': 0.017481} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2055.390104] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2055.390104] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 21c05725-3dad-4965-98d0-0622ebcc8ece/21c05725-3dad-4965-98d0-0622ebcc8ece.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2055.390243] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5d7ad1b5-5c96-45c5-b35b-214018009fad {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2055.397630] env[62824]: DEBUG oslo_vmware.api [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Waiting for the task: (returnval){ [ 2055.397630] env[62824]: value = "task-2146152" [ 2055.397630] env[62824]: _type = "Task" [ 2055.397630] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2055.403536] env[62824]: INFO nova.compute.manager [-] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Took 1.36 seconds to deallocate network for instance. [ 2055.414879] env[62824]: DEBUG oslo_vmware.api [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Task: {'id': task-2146152, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2055.561866] env[62824]: DEBUG oslo_concurrency.lockutils [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Releasing lock "refresh_cache-b0456b8a-348a-4503-a92c-58e5ab455d1c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2055.564042] env[62824]: DEBUG nova.compute.manager [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: b0456b8a-348a-4503-a92c-58e5ab455d1c] Instance network_info: |[{"id": "0534c02e-d202-458d-9289-7d0ad2c5596f", "address": "fa:16:3e:b0:2c:b9", "network": {"id": "bcf30d77-cac1-4deb-971d-efb7c1fb8354", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1030043097-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "0503b434636c4ef5bae8db1b0c74d2ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3cc0a33d-17c0-4b87-b48f-413a87a4cc6a", "external-id": "nsx-vlan-transportzone-865", "segmentation_id": 865, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0534c02e-d2", "ovs_interfaceid": "0534c02e-d202-458d-9289-7d0ad2c5596f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2055.564042] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: b0456b8a-348a-4503-a92c-58e5ab455d1c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b0:2c:b9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3cc0a33d-17c0-4b87-b48f-413a87a4cc6a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0534c02e-d202-458d-9289-7d0ad2c5596f', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2055.571291] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Creating folder: Project (0503b434636c4ef5bae8db1b0c74d2ce). Parent ref: group-v438503. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 2055.573321] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e50ae6b7-c096-49fc-ad32-3edd3569e7df {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2055.590127] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Created folder: Project (0503b434636c4ef5bae8db1b0c74d2ce) in parent group-v438503. [ 2055.590127] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Creating folder: Instances. Parent ref: group-v438651. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 2055.590127] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-23cedeb9-e97a-45af-bff6-9561cebad381 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2055.602669] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Created folder: Instances in parent group-v438651. [ 2055.602918] env[62824]: DEBUG oslo.service.loopingcall [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2055.603531] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b0456b8a-348a-4503-a92c-58e5ab455d1c] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2055.603690] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-314ad470-6259-447d-9c5b-168988c41f39 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2055.626759] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a1ef20b1-87c8-4499-bf60-63e3997b1e64 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Lock "41f068d3-2c8d-46f0-8d84-78531d28b0dd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 37.363s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2055.629856] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2055.629856] env[62824]: value = "task-2146155" [ 2055.629856] env[62824]: _type = "Task" [ 2055.629856] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2055.639678] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146155, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2055.742253] env[62824]: DEBUG nova.compute.manager [req-a685c793-3c43-4d49-9d9a-7a5a8f044071 req-98e4c097-3fb3-4d54-bb33-7d1401a280ba service nova] [instance: b0456b8a-348a-4503-a92c-58e5ab455d1c] Received event network-vif-plugged-0534c02e-d202-458d-9289-7d0ad2c5596f {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2055.742565] env[62824]: DEBUG oslo_concurrency.lockutils [req-a685c793-3c43-4d49-9d9a-7a5a8f044071 req-98e4c097-3fb3-4d54-bb33-7d1401a280ba service nova] Acquiring lock "b0456b8a-348a-4503-a92c-58e5ab455d1c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2055.742877] env[62824]: DEBUG oslo_concurrency.lockutils [req-a685c793-3c43-4d49-9d9a-7a5a8f044071 req-98e4c097-3fb3-4d54-bb33-7d1401a280ba service nova] Lock "b0456b8a-348a-4503-a92c-58e5ab455d1c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2055.743151] env[62824]: DEBUG oslo_concurrency.lockutils [req-a685c793-3c43-4d49-9d9a-7a5a8f044071 req-98e4c097-3fb3-4d54-bb33-7d1401a280ba service nova] Lock "b0456b8a-348a-4503-a92c-58e5ab455d1c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2055.743413] env[62824]: DEBUG nova.compute.manager [req-a685c793-3c43-4d49-9d9a-7a5a8f044071 req-98e4c097-3fb3-4d54-bb33-7d1401a280ba service nova] [instance: b0456b8a-348a-4503-a92c-58e5ab455d1c] No waiting events found dispatching network-vif-plugged-0534c02e-d202-458d-9289-7d0ad2c5596f {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2055.743620] env[62824]: WARNING nova.compute.manager [req-a685c793-3c43-4d49-9d9a-7a5a8f044071 req-98e4c097-3fb3-4d54-bb33-7d1401a280ba service nova] [instance: b0456b8a-348a-4503-a92c-58e5ab455d1c] Received unexpected event network-vif-plugged-0534c02e-d202-458d-9289-7d0ad2c5596f for instance with vm_state building and task_state spawning. [ 2055.743891] env[62824]: DEBUG nova.compute.manager [req-a685c793-3c43-4d49-9d9a-7a5a8f044071 req-98e4c097-3fb3-4d54-bb33-7d1401a280ba service nova] [instance: b0456b8a-348a-4503-a92c-58e5ab455d1c] Received event network-changed-0534c02e-d202-458d-9289-7d0ad2c5596f {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2055.744139] env[62824]: DEBUG nova.compute.manager [req-a685c793-3c43-4d49-9d9a-7a5a8f044071 req-98e4c097-3fb3-4d54-bb33-7d1401a280ba service nova] [instance: b0456b8a-348a-4503-a92c-58e5ab455d1c] Refreshing instance network info cache due to event network-changed-0534c02e-d202-458d-9289-7d0ad2c5596f. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2055.744407] env[62824]: DEBUG oslo_concurrency.lockutils [req-a685c793-3c43-4d49-9d9a-7a5a8f044071 req-98e4c097-3fb3-4d54-bb33-7d1401a280ba service nova] Acquiring lock "refresh_cache-b0456b8a-348a-4503-a92c-58e5ab455d1c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2055.744628] env[62824]: DEBUG oslo_concurrency.lockutils [req-a685c793-3c43-4d49-9d9a-7a5a8f044071 req-98e4c097-3fb3-4d54-bb33-7d1401a280ba service nova] Acquired lock "refresh_cache-b0456b8a-348a-4503-a92c-58e5ab455d1c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2055.744955] env[62824]: DEBUG nova.network.neutron [req-a685c793-3c43-4d49-9d9a-7a5a8f044071 req-98e4c097-3fb3-4d54-bb33-7d1401a280ba service nova] [instance: b0456b8a-348a-4503-a92c-58e5ab455d1c] Refreshing network info cache for port 0534c02e-d202-458d-9289-7d0ad2c5596f {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2055.758077] env[62824]: DEBUG nova.compute.manager [None req-abb349ec-c188-4834-ae78-a5dd5c6e4af7 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] [instance: 4423d172-acdd-4c69-b3b8-ff166e1b8548] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2055.758316] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-abb349ec-c188-4834-ae78-a5dd5c6e4af7 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] [instance: 4423d172-acdd-4c69-b3b8-ff166e1b8548] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2055.759487] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-884e3d79-21d3-4fb5-b783-2cd560ee5f8a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2055.771505] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-abb349ec-c188-4834-ae78-a5dd5c6e4af7 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] [instance: 4423d172-acdd-4c69-b3b8-ff166e1b8548] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2055.771845] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-85de33f0-cb93-4380-b708-ca5f2ac1d77f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2055.852853] env[62824]: DEBUG oslo_vmware.api [None req-752f903a-bdbd-4c93-8465-c657fedd4220 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146151, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2055.911201] env[62824]: DEBUG oslo_vmware.api [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Task: {'id': task-2146152, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2055.916772] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bf6ca20b-2439-4b4d-ba93-061898ce061b tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2055.953540] env[62824]: DEBUG nova.network.neutron [-] [instance: c65ccd20-b682-420a-9c1a-47a45959197e] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2055.985135] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-abb349ec-c188-4834-ae78-a5dd5c6e4af7 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] [instance: 4423d172-acdd-4c69-b3b8-ff166e1b8548] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2055.985598] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-abb349ec-c188-4834-ae78-a5dd5c6e4af7 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] [instance: 4423d172-acdd-4c69-b3b8-ff166e1b8548] Deleting contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2055.985832] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-abb349ec-c188-4834-ae78-a5dd5c6e4af7 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Deleting the datastore file [datastore1] 4423d172-acdd-4c69-b3b8-ff166e1b8548 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2055.987450] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-34e1ab36-60be-4088-ae85-f35d441a1799 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2055.998283] env[62824]: DEBUG oslo_vmware.api [None req-abb349ec-c188-4834-ae78-a5dd5c6e4af7 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Waiting for the task: (returnval){ [ 2055.998283] env[62824]: value = "task-2146157" [ 2055.998283] env[62824]: _type = "Task" [ 2055.998283] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2056.015485] env[62824]: DEBUG oslo_vmware.api [None req-abb349ec-c188-4834-ae78-a5dd5c6e4af7 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Task: {'id': task-2146157, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2056.041736] env[62824]: DEBUG nova.objects.instance [None req-895a9f6d-c24d-4e89-b988-6024da95cb4d tempest-VolumesAssistedSnapshotsTest-163918615 tempest-VolumesAssistedSnapshotsTest-163918615-project-admin] Lazy-loading 'flavor' on Instance uuid 29b8dd5f-1855-490c-a01b-54840073a753 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2056.147965] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146155, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2056.352920] env[62824]: DEBUG oslo_vmware.api [None req-752f903a-bdbd-4c93-8465-c657fedd4220 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146151, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2056.412998] env[62824]: DEBUG oslo_vmware.api [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Task: {'id': task-2146152, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.598948} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2056.416857] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 21c05725-3dad-4965-98d0-0622ebcc8ece/21c05725-3dad-4965-98d0-0622ebcc8ece.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2056.417767] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 21c05725-3dad-4965-98d0-0622ebcc8ece] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2056.424022] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8112f10a-62a0-4de6-a9c9-12e1f409795a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2056.432797] env[62824]: DEBUG oslo_vmware.api [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Waiting for the task: (returnval){ [ 2056.432797] env[62824]: value = "task-2146158" [ 2056.432797] env[62824]: _type = "Task" [ 2056.432797] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2056.448647] env[62824]: DEBUG oslo_vmware.api [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Task: {'id': task-2146158, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2056.459950] env[62824]: INFO nova.compute.manager [-] [instance: c65ccd20-b682-420a-9c1a-47a45959197e] Took 1.39 seconds to deallocate network for instance. [ 2056.479067] env[62824]: DEBUG nova.compute.manager [req-03bfb8dd-0f2f-4545-a08b-154f90005a8a req-974d77c2-0ae6-47a8-9377-f69c57b2a7a6 service nova] [instance: c65ccd20-b682-420a-9c1a-47a45959197e] Received event network-vif-deleted-344e044b-aab1-485b-bd89-5477bc394098 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2056.509462] env[62824]: DEBUG nova.network.neutron [-] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2056.520838] env[62824]: DEBUG oslo_vmware.api [None req-abb349ec-c188-4834-ae78-a5dd5c6e4af7 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Task: {'id': task-2146157, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.388161} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2056.521250] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-abb349ec-c188-4834-ae78-a5dd5c6e4af7 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2056.521475] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-abb349ec-c188-4834-ae78-a5dd5c6e4af7 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] [instance: 4423d172-acdd-4c69-b3b8-ff166e1b8548] Deleted contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2056.521657] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-abb349ec-c188-4834-ae78-a5dd5c6e4af7 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] [instance: 4423d172-acdd-4c69-b3b8-ff166e1b8548] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2056.521826] env[62824]: INFO nova.compute.manager [None req-abb349ec-c188-4834-ae78-a5dd5c6e4af7 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] [instance: 4423d172-acdd-4c69-b3b8-ff166e1b8548] Took 0.76 seconds to destroy the instance on the hypervisor. [ 2056.522078] env[62824]: DEBUG oslo.service.loopingcall [None req-abb349ec-c188-4834-ae78-a5dd5c6e4af7 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2056.522279] env[62824]: DEBUG nova.compute.manager [-] [instance: 4423d172-acdd-4c69-b3b8-ff166e1b8548] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2056.522436] env[62824]: DEBUG nova.network.neutron [-] [instance: 4423d172-acdd-4c69-b3b8-ff166e1b8548] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2056.547262] env[62824]: DEBUG oslo_concurrency.lockutils [None req-895a9f6d-c24d-4e89-b988-6024da95cb4d tempest-VolumesAssistedSnapshotsTest-163918615 tempest-VolumesAssistedSnapshotsTest-163918615-project-admin] Lock "29b8dd5f-1855-490c-a01b-54840073a753" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.843s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2056.579657] env[62824]: DEBUG nova.network.neutron [req-a685c793-3c43-4d49-9d9a-7a5a8f044071 req-98e4c097-3fb3-4d54-bb33-7d1401a280ba service nova] [instance: b0456b8a-348a-4503-a92c-58e5ab455d1c] Updated VIF entry in instance network info cache for port 0534c02e-d202-458d-9289-7d0ad2c5596f. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2056.580309] env[62824]: DEBUG nova.network.neutron [req-a685c793-3c43-4d49-9d9a-7a5a8f044071 req-98e4c097-3fb3-4d54-bb33-7d1401a280ba service nova] [instance: b0456b8a-348a-4503-a92c-58e5ab455d1c] Updating instance_info_cache with network_info: [{"id": "0534c02e-d202-458d-9289-7d0ad2c5596f", "address": "fa:16:3e:b0:2c:b9", "network": {"id": "bcf30d77-cac1-4deb-971d-efb7c1fb8354", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1030043097-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "0503b434636c4ef5bae8db1b0c74d2ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3cc0a33d-17c0-4b87-b48f-413a87a4cc6a", "external-id": "nsx-vlan-transportzone-865", "segmentation_id": 865, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0534c02e-d2", "ovs_interfaceid": "0534c02e-d202-458d-9289-7d0ad2c5596f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2056.648753] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146155, 'name': CreateVM_Task, 'duration_secs': 0.664421} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2056.649178] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b0456b8a-348a-4503-a92c-58e5ab455d1c] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2056.650210] env[62824]: DEBUG oslo_concurrency.lockutils [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2056.650511] env[62824]: DEBUG oslo_concurrency.lockutils [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2056.651156] env[62824]: DEBUG oslo_concurrency.lockutils [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2056.651517] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b703fe48-5a2e-465d-aef0-a31ad0b1e51d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2056.659713] env[62824]: DEBUG oslo_vmware.api [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Waiting for the task: (returnval){ [ 2056.659713] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52fd2d5f-c0d3-d32f-1079-522e389e5045" [ 2056.659713] env[62824]: _type = "Task" [ 2056.659713] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2056.674740] env[62824]: DEBUG oslo_vmware.api [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52fd2d5f-c0d3-d32f-1079-522e389e5045, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2056.769381] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-176b5895-1fd0-48ba-9cd2-150ef888ea39 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2056.778334] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-909180d9-3808-49da-ab53-8f4fb25982d8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2056.815984] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8e3495f-4a1d-48e0-8442-bd42a239cb83 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2056.826667] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8db518d9-3738-4a33-848c-43c12fa4975d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2056.843824] env[62824]: DEBUG nova.compute.provider_tree [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2056.857452] env[62824]: DEBUG oslo_vmware.api [None req-752f903a-bdbd-4c93-8465-c657fedd4220 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146151, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2056.947057] env[62824]: DEBUG oslo_vmware.api [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Task: {'id': task-2146158, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.096675} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2056.947432] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 21c05725-3dad-4965-98d0-0622ebcc8ece] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2056.948420] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6da8c246-b3bc-4e04-93e0-6129d3d954e2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2056.973759] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 21c05725-3dad-4965-98d0-0622ebcc8ece] Reconfiguring VM instance instance-00000036 to attach disk [datastore1] 21c05725-3dad-4965-98d0-0622ebcc8ece/21c05725-3dad-4965-98d0-0622ebcc8ece.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2056.974938] env[62824]: DEBUG oslo_concurrency.lockutils [None req-02170dc2-e93c-4f62-a015-5cb11263272d tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2056.975199] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-35c73be1-3cd3-4b37-b1ee-58429f6ed73b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2056.998041] env[62824]: DEBUG oslo_vmware.api [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Waiting for the task: (returnval){ [ 2056.998041] env[62824]: value = "task-2146159" [ 2056.998041] env[62824]: _type = "Task" [ 2056.998041] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2057.014824] env[62824]: INFO nova.compute.manager [-] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Took 1.94 seconds to deallocate network for instance. [ 2057.015299] env[62824]: DEBUG oslo_vmware.api [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Task: {'id': task-2146159, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2057.087857] env[62824]: DEBUG oslo_concurrency.lockutils [req-a685c793-3c43-4d49-9d9a-7a5a8f044071 req-98e4c097-3fb3-4d54-bb33-7d1401a280ba service nova] Releasing lock "refresh_cache-b0456b8a-348a-4503-a92c-58e5ab455d1c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2057.088337] env[62824]: DEBUG nova.compute.manager [req-a685c793-3c43-4d49-9d9a-7a5a8f044071 req-98e4c097-3fb3-4d54-bb33-7d1401a280ba service nova] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Received event network-vif-deleted-a1d55778-1594-4d8c-943a-350198200d3a {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2057.176504] env[62824]: DEBUG oslo_vmware.api [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52fd2d5f-c0d3-d32f-1079-522e389e5045, 'name': SearchDatastore_Task, 'duration_secs': 0.016289} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2057.178112] env[62824]: DEBUG oslo_concurrency.lockutils [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2057.178112] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: b0456b8a-348a-4503-a92c-58e5ab455d1c] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2057.178342] env[62824]: DEBUG oslo_concurrency.lockutils [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2057.178510] env[62824]: DEBUG oslo_concurrency.lockutils [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2057.178722] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2057.180564] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-50fd639e-5d33-4843-a798-05194855db35 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2057.190718] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2057.190927] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2057.191836] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a90a7a3a-f6f1-487b-8547-ee40ccfd2d05 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2057.200084] env[62824]: DEBUG oslo_vmware.api [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Waiting for the task: (returnval){ [ 2057.200084] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]525537ce-0888-3bfe-5a4b-85888b527934" [ 2057.200084] env[62824]: _type = "Task" [ 2057.200084] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2057.211850] env[62824]: DEBUG oslo_vmware.api [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]525537ce-0888-3bfe-5a4b-85888b527934, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2057.319522] env[62824]: DEBUG nova.network.neutron [-] [instance: 4423d172-acdd-4c69-b3b8-ff166e1b8548] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2057.350621] env[62824]: DEBUG nova.scheduler.client.report [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 173, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2057.364998] env[62824]: DEBUG oslo_vmware.api [None req-752f903a-bdbd-4c93-8465-c657fedd4220 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146151, 'name': CloneVM_Task, 'duration_secs': 2.019624} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2057.365912] env[62824]: INFO nova.virt.vmwareapi.vmops [None req-752f903a-bdbd-4c93-8465-c657fedd4220 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 88bad2dd-dce7-41df-b56c-93a5d054c11e] Created linked-clone VM from snapshot [ 2057.368238] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c39f2973-762c-4402-941f-44c92bac8222 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2057.385248] env[62824]: DEBUG nova.virt.vmwareapi.images [None req-752f903a-bdbd-4c93-8465-c657fedd4220 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 88bad2dd-dce7-41df-b56c-93a5d054c11e] Uploading image d11d1462-ceff-4a2d-bb14-0cf931545d6f {{(pid=62824) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 2057.417249] env[62824]: DEBUG oslo_vmware.rw_handles [None req-752f903a-bdbd-4c93-8465-c657fedd4220 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 2057.417249] env[62824]: value = "vm-438650" [ 2057.417249] env[62824]: _type = "VirtualMachine" [ 2057.417249] env[62824]: }. {{(pid=62824) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 2057.417816] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-27f36230-628a-4b61-b115-c8119837ad3c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2057.428440] env[62824]: DEBUG oslo_vmware.rw_handles [None req-752f903a-bdbd-4c93-8465-c657fedd4220 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Lease: (returnval){ [ 2057.428440] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52a3e345-0c8e-7a20-ab7f-058ca1a06a25" [ 2057.428440] env[62824]: _type = "HttpNfcLease" [ 2057.428440] env[62824]: } obtained for exporting VM: (result){ [ 2057.428440] env[62824]: value = "vm-438650" [ 2057.428440] env[62824]: _type = "VirtualMachine" [ 2057.428440] env[62824]: }. {{(pid=62824) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 2057.428440] env[62824]: DEBUG oslo_vmware.api [None req-752f903a-bdbd-4c93-8465-c657fedd4220 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the lease: (returnval){ [ 2057.428440] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52a3e345-0c8e-7a20-ab7f-058ca1a06a25" [ 2057.428440] env[62824]: _type = "HttpNfcLease" [ 2057.428440] env[62824]: } to be ready. {{(pid=62824) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 2057.437022] env[62824]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 2057.437022] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52a3e345-0c8e-7a20-ab7f-058ca1a06a25" [ 2057.437022] env[62824]: _type = "HttpNfcLease" [ 2057.437022] env[62824]: } is initializing. {{(pid=62824) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 2057.509513] env[62824]: DEBUG oslo_vmware.api [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Task: {'id': task-2146159, 'name': ReconfigVM_Task, 'duration_secs': 0.373559} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2057.509914] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 21c05725-3dad-4965-98d0-0622ebcc8ece] Reconfigured VM instance instance-00000036 to attach disk [datastore1] 21c05725-3dad-4965-98d0-0622ebcc8ece/21c05725-3dad-4965-98d0-0622ebcc8ece.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2057.510655] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c622d85e-acbb-42a0-8734-6d641e807b71 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2057.519458] env[62824]: DEBUG oslo_vmware.api [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Waiting for the task: (returnval){ [ 2057.519458] env[62824]: value = "task-2146161" [ 2057.519458] env[62824]: _type = "Task" [ 2057.519458] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2057.528555] env[62824]: DEBUG oslo_concurrency.lockutils [None req-24b1fd6d-11b0-414f-a359-57219816e958 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2057.539295] env[62824]: DEBUG oslo_vmware.api [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Task: {'id': task-2146161, 'name': Rename_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2057.540227] env[62824]: DEBUG nova.network.neutron [None req-74d3df01-c7bb-4adb-8f79-9713d2a8d4b2 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Successfully updated port: c34f030d-46da-4a7a-9703-5acb2e526d50 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2057.711967] env[62824]: DEBUG oslo_vmware.api [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]525537ce-0888-3bfe-5a4b-85888b527934, 'name': SearchDatastore_Task, 'duration_secs': 0.016448} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2057.712990] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b13482fd-6f1c-4d6b-8365-f4e7e67e1734 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2057.719533] env[62824]: DEBUG oslo_vmware.api [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Waiting for the task: (returnval){ [ 2057.719533] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]526063a8-6329-3aa2-7a4f-bf8e46826634" [ 2057.719533] env[62824]: _type = "Task" [ 2057.719533] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2057.728143] env[62824]: DEBUG oslo_vmware.api [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]526063a8-6329-3aa2-7a4f-bf8e46826634, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2057.790078] env[62824]: INFO nova.compute.manager [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] [instance: ed7d3e95-f26e-40fd-a370-e17922bbff8e] Rebuilding instance [ 2057.823112] env[62824]: INFO nova.compute.manager [-] [instance: 4423d172-acdd-4c69-b3b8-ff166e1b8548] Took 1.30 seconds to deallocate network for instance. [ 2057.834862] env[62824]: DEBUG nova.compute.manager [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] [instance: ed7d3e95-f26e-40fd-a370-e17922bbff8e] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2057.835843] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca4fb6cc-43f9-4c80-af41-a318c3441126 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2057.866547] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.777s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2057.867097] env[62824]: DEBUG nova.compute.manager [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2057.869734] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9d78c884-a693-423a-b835-e548c8af80b6 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.350s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2057.869945] env[62824]: DEBUG nova.objects.instance [None req-9d78c884-a693-423a-b835-e548c8af80b6 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Lazy-loading 'resources' on Instance uuid b96d1351-f3a7-4bac-998c-a34ab2606041 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2057.936976] env[62824]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 2057.936976] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52a3e345-0c8e-7a20-ab7f-058ca1a06a25" [ 2057.936976] env[62824]: _type = "HttpNfcLease" [ 2057.936976] env[62824]: } is ready. {{(pid=62824) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 2057.937288] env[62824]: DEBUG oslo_vmware.rw_handles [None req-752f903a-bdbd-4c93-8465-c657fedd4220 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 2057.937288] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52a3e345-0c8e-7a20-ab7f-058ca1a06a25" [ 2057.937288] env[62824]: _type = "HttpNfcLease" [ 2057.937288] env[62824]: }. {{(pid=62824) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 2057.938915] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8433589-94e0-4fe4-a6bd-ae7c28b92ff9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2057.948521] env[62824]: DEBUG oslo_vmware.rw_handles [None req-752f903a-bdbd-4c93-8465-c657fedd4220 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52cfba3a-f3ed-caee-c5f8-4849298710d3/disk-0.vmdk from lease info. {{(pid=62824) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 2057.948757] env[62824]: DEBUG oslo_vmware.rw_handles [None req-752f903a-bdbd-4c93-8465-c657fedd4220 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52cfba3a-f3ed-caee-c5f8-4849298710d3/disk-0.vmdk for reading. {{(pid=62824) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 2058.031066] env[62824]: DEBUG oslo_vmware.api [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Task: {'id': task-2146161, 'name': Rename_Task, 'duration_secs': 0.142611} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2058.031382] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 21c05725-3dad-4965-98d0-0622ebcc8ece] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2058.031671] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f2ab5692-a342-490e-94fe-3fd3d29c38c1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2058.039198] env[62824]: DEBUG oslo_vmware.api [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Waiting for the task: (returnval){ [ 2058.039198] env[62824]: value = "task-2146162" [ 2058.039198] env[62824]: _type = "Task" [ 2058.039198] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2058.043080] env[62824]: DEBUG oslo_concurrency.lockutils [None req-74d3df01-c7bb-4adb-8f79-9713d2a8d4b2 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquiring lock "refresh_cache-f4d63a93-23af-470c-b36a-662af81dc386" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2058.043286] env[62824]: DEBUG oslo_concurrency.lockutils [None req-74d3df01-c7bb-4adb-8f79-9713d2a8d4b2 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquired lock "refresh_cache-f4d63a93-23af-470c-b36a-662af81dc386" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2058.043539] env[62824]: DEBUG nova.network.neutron [None req-74d3df01-c7bb-4adb-8f79-9713d2a8d4b2 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2058.050627] env[62824]: DEBUG oslo_vmware.api [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Task: {'id': task-2146162, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2058.063451] env[62824]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-23874bc7-d863-4dfb-9350-63ddb5ce3982 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2058.231443] env[62824]: DEBUG oslo_vmware.api [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]526063a8-6329-3aa2-7a4f-bf8e46826634, 'name': SearchDatastore_Task, 'duration_secs': 0.022358} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2058.231744] env[62824]: DEBUG oslo_concurrency.lockutils [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2058.232083] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] b0456b8a-348a-4503-a92c-58e5ab455d1c/b0456b8a-348a-4503-a92c-58e5ab455d1c.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2058.232314] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ab2e8623-fd56-4407-85de-ddd72abd06e6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2058.240756] env[62824]: DEBUG oslo_vmware.api [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Waiting for the task: (returnval){ [ 2058.240756] env[62824]: value = "task-2146163" [ 2058.240756] env[62824]: _type = "Task" [ 2058.240756] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2058.250256] env[62824]: DEBUG oslo_vmware.api [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146163, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2058.329487] env[62824]: DEBUG oslo_concurrency.lockutils [None req-abb349ec-c188-4834-ae78-a5dd5c6e4af7 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2058.373597] env[62824]: DEBUG nova.compute.utils [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2058.378293] env[62824]: DEBUG nova.compute.manager [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2058.378521] env[62824]: DEBUG nova.network.neutron [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2058.380774] env[62824]: DEBUG oslo_concurrency.lockutils [None req-23d5dc9c-2a7f-430b-a467-cfa30a8b4ca0 tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Acquiring lock "82382932-7302-4441-a6f8-9aa2300ec0f6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2058.381147] env[62824]: DEBUG oslo_concurrency.lockutils [None req-23d5dc9c-2a7f-430b-a467-cfa30a8b4ca0 tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Lock "82382932-7302-4441-a6f8-9aa2300ec0f6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2058.381412] env[62824]: DEBUG oslo_concurrency.lockutils [None req-23d5dc9c-2a7f-430b-a467-cfa30a8b4ca0 tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Acquiring lock "82382932-7302-4441-a6f8-9aa2300ec0f6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2058.381686] env[62824]: DEBUG oslo_concurrency.lockutils [None req-23d5dc9c-2a7f-430b-a467-cfa30a8b4ca0 tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Lock "82382932-7302-4441-a6f8-9aa2300ec0f6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2058.382377] env[62824]: DEBUG oslo_concurrency.lockutils [None req-23d5dc9c-2a7f-430b-a467-cfa30a8b4ca0 tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Lock "82382932-7302-4441-a6f8-9aa2300ec0f6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2058.384690] env[62824]: INFO nova.compute.manager [None req-23d5dc9c-2a7f-430b-a467-cfa30a8b4ca0 tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] [instance: 82382932-7302-4441-a6f8-9aa2300ec0f6] Terminating instance [ 2058.446715] env[62824]: DEBUG nova.policy [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6ac24e1886434785b90dfba135fb7c8f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e10208040df34eb5adfe8dcbc76043d5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2058.552670] env[62824]: DEBUG oslo_vmware.api [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Task: {'id': task-2146162, 'name': PowerOnVM_Task, 'duration_secs': 0.494725} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2058.556676] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 21c05725-3dad-4965-98d0-0622ebcc8ece] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2058.558517] env[62824]: DEBUG nova.compute.manager [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 21c05725-3dad-4965-98d0-0622ebcc8ece] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2058.561688] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd558bfd-2d0d-4db7-8c61-8935c6150a90 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2058.596553] env[62824]: WARNING nova.network.neutron [None req-74d3df01-c7bb-4adb-8f79-9713d2a8d4b2 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: f4d63a93-23af-470c-b36a-662af81dc386] aa1c09e5-8d72-43ad-a903-a0f5e711da80 already exists in list: networks containing: ['aa1c09e5-8d72-43ad-a903-a0f5e711da80']. ignoring it [ 2058.762144] env[62824]: DEBUG oslo_vmware.api [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146163, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2058.770890] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8b5da446-9f53-4569-8936-1a55fec9a1e7 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Acquiring lock "2afecec5-763d-4616-b690-41d3101cfc47" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2058.771166] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8b5da446-9f53-4569-8936-1a55fec9a1e7 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Lock "2afecec5-763d-4616-b690-41d3101cfc47" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2058.771400] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8b5da446-9f53-4569-8936-1a55fec9a1e7 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Acquiring lock "2afecec5-763d-4616-b690-41d3101cfc47-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2058.771730] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8b5da446-9f53-4569-8936-1a55fec9a1e7 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Lock "2afecec5-763d-4616-b690-41d3101cfc47-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2058.771985] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8b5da446-9f53-4569-8936-1a55fec9a1e7 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Lock "2afecec5-763d-4616-b690-41d3101cfc47-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2058.774340] env[62824]: INFO nova.compute.manager [None req-8b5da446-9f53-4569-8936-1a55fec9a1e7 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 2afecec5-763d-4616-b690-41d3101cfc47] Terminating instance [ 2058.853546] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] [instance: ed7d3e95-f26e-40fd-a370-e17922bbff8e] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2058.853877] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-09cc9b68-1fa8-4a44-a821-183c15f9bea2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2058.868564] env[62824]: DEBUG oslo_vmware.api [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Waiting for the task: (returnval){ [ 2058.868564] env[62824]: value = "task-2146164" [ 2058.868564] env[62824]: _type = "Task" [ 2058.868564] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2058.871011] env[62824]: DEBUG nova.network.neutron [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Successfully created port: d85a0acc-fed6-4797-9f81-2aafa5bb6967 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2058.884746] env[62824]: DEBUG nova.compute.manager [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2058.889656] env[62824]: DEBUG nova.compute.manager [None req-23d5dc9c-2a7f-430b-a467-cfa30a8b4ca0 tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] [instance: 82382932-7302-4441-a6f8-9aa2300ec0f6] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2058.890108] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-23d5dc9c-2a7f-430b-a467-cfa30a8b4ca0 tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] [instance: 82382932-7302-4441-a6f8-9aa2300ec0f6] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2058.893342] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f1a93cd-d768-4569-b94d-4f09e134968f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2058.904370] env[62824]: DEBUG oslo_vmware.api [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Task: {'id': task-2146164, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2058.911599] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-23d5dc9c-2a7f-430b-a467-cfa30a8b4ca0 tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] [instance: 82382932-7302-4441-a6f8-9aa2300ec0f6] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2058.913282] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1c4cc354-2f0d-446a-affd-2b227b07a4d6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2058.924855] env[62824]: DEBUG oslo_vmware.api [None req-23d5dc9c-2a7f-430b-a467-cfa30a8b4ca0 tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Waiting for the task: (returnval){ [ 2058.924855] env[62824]: value = "task-2146165" [ 2058.924855] env[62824]: _type = "Task" [ 2058.924855] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2058.938624] env[62824]: DEBUG oslo_vmware.api [None req-23d5dc9c-2a7f-430b-a467-cfa30a8b4ca0 tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Task: {'id': task-2146165, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2059.018204] env[62824]: DEBUG nova.network.neutron [None req-74d3df01-c7bb-4adb-8f79-9713d2a8d4b2 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Updating instance_info_cache with network_info: [{"id": "5a010977-7d05-47e0-8688-bbf9a0e70900", "address": "fa:16:3e:1e:14:06", "network": {"id": "aa1c09e5-8d72-43ad-a903-a0f5e711da80", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1643616572-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.251", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c987c5f844be42c99cc9d57a9ca84c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9bc2632-36f9-4912-8782-8bbb789f909d", "external-id": "nsx-vlan-transportzone-897", "segmentation_id": 897, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5a010977-7d", "ovs_interfaceid": "5a010977-7d05-47e0-8688-bbf9a0e70900", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "c34f030d-46da-4a7a-9703-5acb2e526d50", "address": "fa:16:3e:78:b8:36", "network": {"id": "aa1c09e5-8d72-43ad-a903-a0f5e711da80", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1643616572-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c987c5f844be42c99cc9d57a9ca84c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9bc2632-36f9-4912-8782-8bbb789f909d", "external-id": "nsx-vlan-transportzone-897", "segmentation_id": 897, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc34f030d-46", "ovs_interfaceid": "c34f030d-46da-4a7a-9703-5acb2e526d50", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2059.066690] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f48b74c-bad6-4014-9d3c-6c8b918ecbce {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2059.075848] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-281bb04d-4483-4e15-a66e-7f710bf57d2f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2059.087379] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2059.115263] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14c1a3df-b7bc-4f0f-9862-296c23c0747e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2059.124182] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3470c74c-462a-4c3e-8fd7-0e290d6b0c52 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2059.140189] env[62824]: DEBUG nova.compute.provider_tree [None req-9d78c884-a693-423a-b835-e548c8af80b6 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2059.254906] env[62824]: DEBUG oslo_vmware.api [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146163, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.691978} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2059.255696] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] b0456b8a-348a-4503-a92c-58e5ab455d1c/b0456b8a-348a-4503-a92c-58e5ab455d1c.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2059.256151] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: b0456b8a-348a-4503-a92c-58e5ab455d1c] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2059.256561] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-70a0eea7-1bb7-472a-9c62-d7035b084b05 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2059.265096] env[62824]: DEBUG oslo_vmware.api [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Waiting for the task: (returnval){ [ 2059.265096] env[62824]: value = "task-2146166" [ 2059.265096] env[62824]: _type = "Task" [ 2059.265096] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2059.275845] env[62824]: DEBUG oslo_vmware.api [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146166, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2059.278732] env[62824]: DEBUG nova.compute.manager [None req-8b5da446-9f53-4569-8936-1a55fec9a1e7 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 2afecec5-763d-4616-b690-41d3101cfc47] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2059.278997] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-8b5da446-9f53-4569-8936-1a55fec9a1e7 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 2afecec5-763d-4616-b690-41d3101cfc47] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2059.280274] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d37da6f-cc9a-4299-a512-2d616c8a5e1b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2059.289692] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b5da446-9f53-4569-8936-1a55fec9a1e7 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 2afecec5-763d-4616-b690-41d3101cfc47] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2059.290524] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a79aacea-1c93-4f1a-b3a3-6ebfd351da26 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2059.298051] env[62824]: DEBUG oslo_vmware.api [None req-8b5da446-9f53-4569-8936-1a55fec9a1e7 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Waiting for the task: (returnval){ [ 2059.298051] env[62824]: value = "task-2146167" [ 2059.298051] env[62824]: _type = "Task" [ 2059.298051] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2059.308396] env[62824]: DEBUG oslo_vmware.api [None req-8b5da446-9f53-4569-8936-1a55fec9a1e7 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': task-2146167, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2059.385531] env[62824]: DEBUG oslo_vmware.api [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Task: {'id': task-2146164, 'name': PowerOffVM_Task, 'duration_secs': 0.186811} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2059.385904] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] [instance: ed7d3e95-f26e-40fd-a370-e17922bbff8e] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2059.386215] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] [instance: ed7d3e95-f26e-40fd-a370-e17922bbff8e] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2059.387076] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acfa3264-7886-417d-980b-0eceef644402 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2059.398644] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] [instance: ed7d3e95-f26e-40fd-a370-e17922bbff8e] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2059.399204] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fe63a5b7-61ea-4a44-840f-a97a512a1355 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2059.435757] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] [instance: ed7d3e95-f26e-40fd-a370-e17922bbff8e] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2059.436114] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] [instance: ed7d3e95-f26e-40fd-a370-e17922bbff8e] Deleting contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2059.436452] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Deleting the datastore file [datastore2] ed7d3e95-f26e-40fd-a370-e17922bbff8e {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2059.436930] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2dfb0429-79f0-4b0e-acae-6d9e7b4ef735 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2059.443237] env[62824]: DEBUG oslo_vmware.api [None req-23d5dc9c-2a7f-430b-a467-cfa30a8b4ca0 tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Task: {'id': task-2146165, 'name': PowerOffVM_Task, 'duration_secs': 0.234242} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2059.444340] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-23d5dc9c-2a7f-430b-a467-cfa30a8b4ca0 tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] [instance: 82382932-7302-4441-a6f8-9aa2300ec0f6] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2059.444685] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-23d5dc9c-2a7f-430b-a467-cfa30a8b4ca0 tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] [instance: 82382932-7302-4441-a6f8-9aa2300ec0f6] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2059.445085] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d25c27c7-5984-481c-9115-5b9b89b15032 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2059.450630] env[62824]: DEBUG oslo_vmware.api [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Waiting for the task: (returnval){ [ 2059.450630] env[62824]: value = "task-2146169" [ 2059.450630] env[62824]: _type = "Task" [ 2059.450630] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2059.466979] env[62824]: DEBUG oslo_vmware.api [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Task: {'id': task-2146169, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2059.521315] env[62824]: DEBUG oslo_concurrency.lockutils [None req-74d3df01-c7bb-4adb-8f79-9713d2a8d4b2 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Releasing lock "refresh_cache-f4d63a93-23af-470c-b36a-662af81dc386" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2059.522036] env[62824]: DEBUG oslo_concurrency.lockutils [None req-74d3df01-c7bb-4adb-8f79-9713d2a8d4b2 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquiring lock "f4d63a93-23af-470c-b36a-662af81dc386" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2059.522202] env[62824]: DEBUG oslo_concurrency.lockutils [None req-74d3df01-c7bb-4adb-8f79-9713d2a8d4b2 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquired lock "f4d63a93-23af-470c-b36a-662af81dc386" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2059.523178] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69ffb60c-2c96-4a69-8343-d3fee9175faa {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2059.543191] env[62824]: DEBUG nova.virt.hardware [None req-74d3df01-c7bb-4adb-8f79-9713d2a8d4b2 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2059.543541] env[62824]: DEBUG nova.virt.hardware [None req-74d3df01-c7bb-4adb-8f79-9713d2a8d4b2 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2059.543758] env[62824]: DEBUG nova.virt.hardware [None req-74d3df01-c7bb-4adb-8f79-9713d2a8d4b2 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2059.544030] env[62824]: DEBUG nova.virt.hardware [None req-74d3df01-c7bb-4adb-8f79-9713d2a8d4b2 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2059.544242] env[62824]: DEBUG nova.virt.hardware [None req-74d3df01-c7bb-4adb-8f79-9713d2a8d4b2 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2059.544440] env[62824]: DEBUG nova.virt.hardware [None req-74d3df01-c7bb-4adb-8f79-9713d2a8d4b2 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2059.544719] env[62824]: DEBUG nova.virt.hardware [None req-74d3df01-c7bb-4adb-8f79-9713d2a8d4b2 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2059.544943] env[62824]: DEBUG nova.virt.hardware [None req-74d3df01-c7bb-4adb-8f79-9713d2a8d4b2 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2059.545184] env[62824]: DEBUG nova.virt.hardware [None req-74d3df01-c7bb-4adb-8f79-9713d2a8d4b2 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2059.545420] env[62824]: DEBUG nova.virt.hardware [None req-74d3df01-c7bb-4adb-8f79-9713d2a8d4b2 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2059.545673] env[62824]: DEBUG nova.virt.hardware [None req-74d3df01-c7bb-4adb-8f79-9713d2a8d4b2 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2059.552363] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-74d3df01-c7bb-4adb-8f79-9713d2a8d4b2 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Reconfiguring VM to attach interface {{(pid=62824) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 2059.553173] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1bea0ef8-b304-4621-888a-db996d888d4f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2059.573297] env[62824]: DEBUG oslo_vmware.api [None req-74d3df01-c7bb-4adb-8f79-9713d2a8d4b2 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Waiting for the task: (returnval){ [ 2059.573297] env[62824]: value = "task-2146171" [ 2059.573297] env[62824]: _type = "Task" [ 2059.573297] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2059.582807] env[62824]: DEBUG oslo_vmware.api [None req-74d3df01-c7bb-4adb-8f79-9713d2a8d4b2 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146171, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2059.608360] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-23d5dc9c-2a7f-430b-a467-cfa30a8b4ca0 tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] [instance: 82382932-7302-4441-a6f8-9aa2300ec0f6] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2059.608659] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-23d5dc9c-2a7f-430b-a467-cfa30a8b4ca0 tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] [instance: 82382932-7302-4441-a6f8-9aa2300ec0f6] Deleting contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2059.608869] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-23d5dc9c-2a7f-430b-a467-cfa30a8b4ca0 tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Deleting the datastore file [datastore2] 82382932-7302-4441-a6f8-9aa2300ec0f6 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2059.609168] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a7c3976f-4f68-449e-bdbd-eeb5f1ff02f4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2059.619280] env[62824]: DEBUG oslo_vmware.api [None req-23d5dc9c-2a7f-430b-a467-cfa30a8b4ca0 tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Waiting for the task: (returnval){ [ 2059.619280] env[62824]: value = "task-2146172" [ 2059.619280] env[62824]: _type = "Task" [ 2059.619280] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2059.628168] env[62824]: DEBUG oslo_vmware.api [None req-23d5dc9c-2a7f-430b-a467-cfa30a8b4ca0 tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Task: {'id': task-2146172, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2059.643453] env[62824]: DEBUG nova.scheduler.client.report [None req-9d78c884-a693-423a-b835-e548c8af80b6 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 173, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2059.779195] env[62824]: DEBUG oslo_vmware.api [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146166, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.123433} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2059.779576] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: b0456b8a-348a-4503-a92c-58e5ab455d1c] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2059.780794] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f216862-ba6c-4aee-9ffe-16d3b13f46d5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2059.808288] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: b0456b8a-348a-4503-a92c-58e5ab455d1c] Reconfiguring VM instance instance-00000038 to attach disk [datastore1] b0456b8a-348a-4503-a92c-58e5ab455d1c/b0456b8a-348a-4503-a92c-58e5ab455d1c.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2059.811938] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9061e866-3c9b-4d8e-9aab-f12689ed8f10 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2059.833997] env[62824]: DEBUG oslo_vmware.api [None req-8b5da446-9f53-4569-8936-1a55fec9a1e7 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': task-2146167, 'name': PowerOffVM_Task, 'duration_secs': 0.253242} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2059.836560] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b5da446-9f53-4569-8936-1a55fec9a1e7 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 2afecec5-763d-4616-b690-41d3101cfc47] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2059.836560] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-8b5da446-9f53-4569-8936-1a55fec9a1e7 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 2afecec5-763d-4616-b690-41d3101cfc47] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2059.836838] env[62824]: DEBUG oslo_vmware.api [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Waiting for the task: (returnval){ [ 2059.836838] env[62824]: value = "task-2146173" [ 2059.836838] env[62824]: _type = "Task" [ 2059.836838] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2059.837137] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3969dae8-9e43-44b8-8363-afa367ea7c8a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2059.849470] env[62824]: DEBUG oslo_vmware.api [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146173, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2059.900695] env[62824]: DEBUG nova.compute.manager [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2059.961947] env[62824]: DEBUG oslo_vmware.api [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Task: {'id': task-2146169, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.154083} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2059.962704] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2059.962704] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] [instance: ed7d3e95-f26e-40fd-a370-e17922bbff8e] Deleted contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2059.962704] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] [instance: ed7d3e95-f26e-40fd-a370-e17922bbff8e] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2060.021192] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-8b5da446-9f53-4569-8936-1a55fec9a1e7 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 2afecec5-763d-4616-b690-41d3101cfc47] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2060.021418] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-8b5da446-9f53-4569-8936-1a55fec9a1e7 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 2afecec5-763d-4616-b690-41d3101cfc47] Deleting contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2060.021618] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-8b5da446-9f53-4569-8936-1a55fec9a1e7 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Deleting the datastore file [datastore2] 2afecec5-763d-4616-b690-41d3101cfc47 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2060.022630] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-adad6d7d-7ccc-4bef-855c-8f792089820c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2060.031183] env[62824]: DEBUG oslo_vmware.api [None req-8b5da446-9f53-4569-8936-1a55fec9a1e7 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Waiting for the task: (returnval){ [ 2060.031183] env[62824]: value = "task-2146175" [ 2060.031183] env[62824]: _type = "Task" [ 2060.031183] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2060.041228] env[62824]: DEBUG oslo_vmware.api [None req-8b5da446-9f53-4569-8936-1a55fec9a1e7 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': task-2146175, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2060.084559] env[62824]: DEBUG oslo_vmware.api [None req-74d3df01-c7bb-4adb-8f79-9713d2a8d4b2 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146171, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2060.129345] env[62824]: DEBUG oslo_vmware.api [None req-23d5dc9c-2a7f-430b-a467-cfa30a8b4ca0 tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Task: {'id': task-2146172, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.199261} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2060.129641] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-23d5dc9c-2a7f-430b-a467-cfa30a8b4ca0 tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2060.129839] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-23d5dc9c-2a7f-430b-a467-cfa30a8b4ca0 tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] [instance: 82382932-7302-4441-a6f8-9aa2300ec0f6] Deleted contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2060.130048] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-23d5dc9c-2a7f-430b-a467-cfa30a8b4ca0 tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] [instance: 82382932-7302-4441-a6f8-9aa2300ec0f6] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2060.130233] env[62824]: INFO nova.compute.manager [None req-23d5dc9c-2a7f-430b-a467-cfa30a8b4ca0 tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] [instance: 82382932-7302-4441-a6f8-9aa2300ec0f6] Took 1.24 seconds to destroy the instance on the hypervisor. [ 2060.130493] env[62824]: DEBUG oslo.service.loopingcall [None req-23d5dc9c-2a7f-430b-a467-cfa30a8b4ca0 tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2060.130707] env[62824]: DEBUG nova.compute.manager [-] [instance: 82382932-7302-4441-a6f8-9aa2300ec0f6] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2060.130805] env[62824]: DEBUG nova.network.neutron [-] [instance: 82382932-7302-4441-a6f8-9aa2300ec0f6] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2060.149332] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9d78c884-a693-423a-b835-e548c8af80b6 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.279s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2060.151960] env[62824]: DEBUG oslo_concurrency.lockutils [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.710s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2060.155617] env[62824]: INFO nova.compute.claims [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2060.178779] env[62824]: INFO nova.scheduler.client.report [None req-9d78c884-a693-423a-b835-e548c8af80b6 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Deleted allocations for instance b96d1351-f3a7-4bac-998c-a34ab2606041 [ 2060.351061] env[62824]: DEBUG oslo_vmware.api [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146173, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2060.493212] env[62824]: DEBUG nova.network.neutron [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Successfully updated port: d85a0acc-fed6-4797-9f81-2aafa5bb6967 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2060.542658] env[62824]: DEBUG oslo_vmware.api [None req-8b5da446-9f53-4569-8936-1a55fec9a1e7 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': task-2146175, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.195118} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2060.542859] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-8b5da446-9f53-4569-8936-1a55fec9a1e7 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2060.543056] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-8b5da446-9f53-4569-8936-1a55fec9a1e7 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 2afecec5-763d-4616-b690-41d3101cfc47] Deleted contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2060.543242] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-8b5da446-9f53-4569-8936-1a55fec9a1e7 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 2afecec5-763d-4616-b690-41d3101cfc47] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2060.543440] env[62824]: INFO nova.compute.manager [None req-8b5da446-9f53-4569-8936-1a55fec9a1e7 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 2afecec5-763d-4616-b690-41d3101cfc47] Took 1.26 seconds to destroy the instance on the hypervisor. [ 2060.543679] env[62824]: DEBUG oslo.service.loopingcall [None req-8b5da446-9f53-4569-8936-1a55fec9a1e7 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2060.543865] env[62824]: DEBUG nova.compute.manager [-] [instance: 2afecec5-763d-4616-b690-41d3101cfc47] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2060.543966] env[62824]: DEBUG nova.network.neutron [-] [instance: 2afecec5-763d-4616-b690-41d3101cfc47] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2060.586131] env[62824]: DEBUG oslo_vmware.api [None req-74d3df01-c7bb-4adb-8f79-9713d2a8d4b2 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146171, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2060.688900] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9d78c884-a693-423a-b835-e548c8af80b6 tempest-ServersTestBootFromVolume-1799904681 tempest-ServersTestBootFromVolume-1799904681-project-member] Lock "b96d1351-f3a7-4bac-998c-a34ab2606041" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.855s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2060.850852] env[62824]: DEBUG oslo_vmware.api [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146173, 'name': ReconfigVM_Task, 'duration_secs': 0.625106} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2060.852819] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: b0456b8a-348a-4503-a92c-58e5ab455d1c] Reconfigured VM instance instance-00000038 to attach disk [datastore1] b0456b8a-348a-4503-a92c-58e5ab455d1c/b0456b8a-348a-4503-a92c-58e5ab455d1c.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2060.853652] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-831924fb-8b75-42ea-8442-689c352f0f45 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2060.861994] env[62824]: DEBUG oslo_vmware.api [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Waiting for the task: (returnval){ [ 2060.861994] env[62824]: value = "task-2146176" [ 2060.861994] env[62824]: _type = "Task" [ 2060.861994] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2060.873847] env[62824]: DEBUG oslo_vmware.api [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146176, 'name': Rename_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2060.998867] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Acquiring lock "refresh_cache-bbfcb2e3-9326-4548-b15b-e054cbfd192e" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2060.999160] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Acquired lock "refresh_cache-bbfcb2e3-9326-4548-b15b-e054cbfd192e" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2060.999324] env[62824]: DEBUG nova.network.neutron [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2061.085721] env[62824]: DEBUG oslo_vmware.api [None req-74d3df01-c7bb-4adb-8f79-9713d2a8d4b2 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146171, 'name': ReconfigVM_Task, 'duration_secs': 1.071303} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2061.086283] env[62824]: DEBUG oslo_concurrency.lockutils [None req-74d3df01-c7bb-4adb-8f79-9713d2a8d4b2 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Releasing lock "f4d63a93-23af-470c-b36a-662af81dc386" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2061.086546] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-74d3df01-c7bb-4adb-8f79-9713d2a8d4b2 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Reconfigured VM to attach interface {{(pid=62824) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 2061.300450] env[62824]: DEBUG nova.network.neutron [-] [instance: 2afecec5-763d-4616-b690-41d3101cfc47] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2061.354305] env[62824]: DEBUG nova.network.neutron [-] [instance: 82382932-7302-4441-a6f8-9aa2300ec0f6] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2061.376028] env[62824]: DEBUG oslo_vmware.api [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146176, 'name': Rename_Task, 'duration_secs': 0.206167} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2061.376267] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: b0456b8a-348a-4503-a92c-58e5ab455d1c] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2061.376559] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-35d7b00a-dcb5-461d-9c87-137217dba550 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2061.385268] env[62824]: DEBUG oslo_vmware.api [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Waiting for the task: (returnval){ [ 2061.385268] env[62824]: value = "task-2146177" [ 2061.385268] env[62824]: _type = "Task" [ 2061.385268] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2061.394269] env[62824]: DEBUG oslo_vmware.api [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146177, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2061.536440] env[62824]: DEBUG nova.network.neutron [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2061.592556] env[62824]: DEBUG oslo_concurrency.lockutils [None req-74d3df01-c7bb-4adb-8f79-9713d2a8d4b2 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lock "interface-f4d63a93-23af-470c-b36a-662af81dc386-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.900s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2061.634934] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba7942c3-1ccd-4d8c-a78c-1b4482019235 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2061.644631] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16cb3e2f-f1f0-4096-9730-6e1efb0e0fa2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2061.679436] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6304b29c-ed8b-4863-bf69-d3d848d8a55f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2061.688239] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcb5676e-9151-4c17-aa20-2d883e402c70 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2061.693362] env[62824]: DEBUG nova.network.neutron [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Updating instance_info_cache with network_info: [{"id": "d85a0acc-fed6-4797-9f81-2aafa5bb6967", "address": "fa:16:3e:d8:c8:ab", "network": {"id": "a410f4f3-d072-4034-b797-8210e1a79093", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1482574391-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e10208040df34eb5adfe8dcbc76043d5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f4399275-8e92-4448-be9e-d4984e93e89c", "external-id": "nsx-vlan-transportzone-192", "segmentation_id": 192, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd85a0acc-fe", "ovs_interfaceid": "d85a0acc-fed6-4797-9f81-2aafa5bb6967", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2061.711176] env[62824]: DEBUG nova.compute.provider_tree [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Updating inventory in ProviderTree for provider bf95157b-a274-42de-9ccf-9851128a44a1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 2061.803297] env[62824]: INFO nova.compute.manager [-] [instance: 2afecec5-763d-4616-b690-41d3101cfc47] Took 1.26 seconds to deallocate network for instance. [ 2061.859395] env[62824]: INFO nova.compute.manager [-] [instance: 82382932-7302-4441-a6f8-9aa2300ec0f6] Took 1.73 seconds to deallocate network for instance. [ 2061.898033] env[62824]: DEBUG oslo_vmware.api [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146177, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2062.195852] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Releasing lock "refresh_cache-bbfcb2e3-9326-4548-b15b-e054cbfd192e" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2062.196270] env[62824]: DEBUG nova.compute.manager [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Instance network_info: |[{"id": "d85a0acc-fed6-4797-9f81-2aafa5bb6967", "address": "fa:16:3e:d8:c8:ab", "network": {"id": "a410f4f3-d072-4034-b797-8210e1a79093", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1482574391-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e10208040df34eb5adfe8dcbc76043d5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f4399275-8e92-4448-be9e-d4984e93e89c", "external-id": "nsx-vlan-transportzone-192", "segmentation_id": 192, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd85a0acc-fe", "ovs_interfaceid": "d85a0acc-fed6-4797-9f81-2aafa5bb6967", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2062.245827] env[62824]: DEBUG nova.scheduler.client.report [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Updated inventory for provider bf95157b-a274-42de-9ccf-9851128a44a1 with generation 77 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 2062.246109] env[62824]: DEBUG nova.compute.provider_tree [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Updating resource provider bf95157b-a274-42de-9ccf-9851128a44a1 generation from 77 to 78 during operation: update_inventory {{(pid=62824) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 2062.246296] env[62824]: DEBUG nova.compute.provider_tree [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Updating inventory in ProviderTree for provider bf95157b-a274-42de-9ccf-9851128a44a1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 2062.311623] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8b5da446-9f53-4569-8936-1a55fec9a1e7 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2062.338497] env[62824]: DEBUG nova.compute.manager [req-637ef91c-59b8-43a9-904a-a13389632a9e req-ea0614ac-cff1-47fc-ae3d-f23377639f69 service nova] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Received event network-vif-deleted-b27bda53-5f70-40ff-bb04-9b22f2241093 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2062.338706] env[62824]: DEBUG nova.compute.manager [req-637ef91c-59b8-43a9-904a-a13389632a9e req-ea0614ac-cff1-47fc-ae3d-f23377639f69 service nova] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Received event network-vif-plugged-c34f030d-46da-4a7a-9703-5acb2e526d50 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2062.339380] env[62824]: DEBUG oslo_concurrency.lockutils [req-637ef91c-59b8-43a9-904a-a13389632a9e req-ea0614ac-cff1-47fc-ae3d-f23377639f69 service nova] Acquiring lock "f4d63a93-23af-470c-b36a-662af81dc386-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2062.339380] env[62824]: DEBUG oslo_concurrency.lockutils [req-637ef91c-59b8-43a9-904a-a13389632a9e req-ea0614ac-cff1-47fc-ae3d-f23377639f69 service nova] Lock "f4d63a93-23af-470c-b36a-662af81dc386-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2062.339380] env[62824]: DEBUG oslo_concurrency.lockutils [req-637ef91c-59b8-43a9-904a-a13389632a9e req-ea0614ac-cff1-47fc-ae3d-f23377639f69 service nova] Lock "f4d63a93-23af-470c-b36a-662af81dc386-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2062.339658] env[62824]: DEBUG nova.compute.manager [req-637ef91c-59b8-43a9-904a-a13389632a9e req-ea0614ac-cff1-47fc-ae3d-f23377639f69 service nova] [instance: f4d63a93-23af-470c-b36a-662af81dc386] No waiting events found dispatching network-vif-plugged-c34f030d-46da-4a7a-9703-5acb2e526d50 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2062.339658] env[62824]: WARNING nova.compute.manager [req-637ef91c-59b8-43a9-904a-a13389632a9e req-ea0614ac-cff1-47fc-ae3d-f23377639f69 service nova] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Received unexpected event network-vif-plugged-c34f030d-46da-4a7a-9703-5acb2e526d50 for instance with vm_state active and task_state None. [ 2062.339717] env[62824]: DEBUG nova.compute.manager [req-637ef91c-59b8-43a9-904a-a13389632a9e req-ea0614ac-cff1-47fc-ae3d-f23377639f69 service nova] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Received event network-changed-c34f030d-46da-4a7a-9703-5acb2e526d50 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2062.339866] env[62824]: DEBUG nova.compute.manager [req-637ef91c-59b8-43a9-904a-a13389632a9e req-ea0614ac-cff1-47fc-ae3d-f23377639f69 service nova] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Refreshing instance network info cache due to event network-changed-c34f030d-46da-4a7a-9703-5acb2e526d50. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2062.340064] env[62824]: DEBUG oslo_concurrency.lockutils [req-637ef91c-59b8-43a9-904a-a13389632a9e req-ea0614ac-cff1-47fc-ae3d-f23377639f69 service nova] Acquiring lock "refresh_cache-f4d63a93-23af-470c-b36a-662af81dc386" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2062.340201] env[62824]: DEBUG oslo_concurrency.lockutils [req-637ef91c-59b8-43a9-904a-a13389632a9e req-ea0614ac-cff1-47fc-ae3d-f23377639f69 service nova] Acquired lock "refresh_cache-f4d63a93-23af-470c-b36a-662af81dc386" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2062.340354] env[62824]: DEBUG nova.network.neutron [req-637ef91c-59b8-43a9-904a-a13389632a9e req-ea0614ac-cff1-47fc-ae3d-f23377639f69 service nova] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Refreshing network info cache for port c34f030d-46da-4a7a-9703-5acb2e526d50 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2062.366425] env[62824]: DEBUG oslo_concurrency.lockutils [None req-23d5dc9c-2a7f-430b-a467-cfa30a8b4ca0 tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2062.397272] env[62824]: DEBUG oslo_vmware.api [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146177, 'name': PowerOnVM_Task, 'duration_secs': 0.714722} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2062.397597] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: b0456b8a-348a-4503-a92c-58e5ab455d1c] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2062.398460] env[62824]: INFO nova.compute.manager [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: b0456b8a-348a-4503-a92c-58e5ab455d1c] Took 9.66 seconds to spawn the instance on the hypervisor. [ 2062.398460] env[62824]: DEBUG nova.compute.manager [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: b0456b8a-348a-4503-a92c-58e5ab455d1c] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2062.398745] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0add8ad-f552-4c24-bc1b-21c6a4610860 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2062.752281] env[62824]: DEBUG oslo_concurrency.lockutils [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.600s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2062.752708] env[62824]: DEBUG nova.compute.manager [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2062.756237] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.857s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2062.758185] env[62824]: INFO nova.compute.claims [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 919e8a45-7810-4a8d-a1aa-5046b5ab059c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2062.919228] env[62824]: INFO nova.compute.manager [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: b0456b8a-348a-4503-a92c-58e5ab455d1c] Took 44.09 seconds to build instance. [ 2063.092719] env[62824]: DEBUG nova.network.neutron [req-637ef91c-59b8-43a9-904a-a13389632a9e req-ea0614ac-cff1-47fc-ae3d-f23377639f69 service nova] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Updated VIF entry in instance network info cache for port c34f030d-46da-4a7a-9703-5acb2e526d50. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2063.093272] env[62824]: DEBUG nova.network.neutron [req-637ef91c-59b8-43a9-904a-a13389632a9e req-ea0614ac-cff1-47fc-ae3d-f23377639f69 service nova] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Updating instance_info_cache with network_info: [{"id": "5a010977-7d05-47e0-8688-bbf9a0e70900", "address": "fa:16:3e:1e:14:06", "network": {"id": "aa1c09e5-8d72-43ad-a903-a0f5e711da80", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1643616572-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.251", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c987c5f844be42c99cc9d57a9ca84c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9bc2632-36f9-4912-8782-8bbb789f909d", "external-id": "nsx-vlan-transportzone-897", "segmentation_id": 897, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5a010977-7d", "ovs_interfaceid": "5a010977-7d05-47e0-8688-bbf9a0e70900", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "c34f030d-46da-4a7a-9703-5acb2e526d50", "address": "fa:16:3e:78:b8:36", "network": {"id": "aa1c09e5-8d72-43ad-a903-a0f5e711da80", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1643616572-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c987c5f844be42c99cc9d57a9ca84c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9bc2632-36f9-4912-8782-8bbb789f909d", "external-id": "nsx-vlan-transportzone-897", "segmentation_id": 897, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc34f030d-46", "ovs_interfaceid": "c34f030d-46da-4a7a-9703-5acb2e526d50", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2063.266540] env[62824]: DEBUG nova.compute.utils [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2063.268169] env[62824]: DEBUG nova.compute.manager [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2063.268359] env[62824]: DEBUG nova.network.neutron [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2063.314477] env[62824]: DEBUG nova.policy [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dcabddbc1f804889a257dae580593fe4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '87239f55d6d44606b23ca0c25d9fb4d9', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2063.421984] env[62824]: DEBUG oslo_concurrency.lockutils [None req-46546cba-ec48-47be-b049-ec42fa63c00e tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Lock "b0456b8a-348a-4503-a92c-58e5ab455d1c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 78.299s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2063.596486] env[62824]: DEBUG oslo_concurrency.lockutils [req-637ef91c-59b8-43a9-904a-a13389632a9e req-ea0614ac-cff1-47fc-ae3d-f23377639f69 service nova] Releasing lock "refresh_cache-f4d63a93-23af-470c-b36a-662af81dc386" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2063.606664] env[62824]: DEBUG nova.network.neutron [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Successfully created port: d358ab3c-f4fb-41f3-954b-19de8b22ef15 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2063.772621] env[62824]: DEBUG nova.compute.manager [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2064.279361] env[62824]: INFO nova.virt.block_device [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Booting with volume 8d462122-e7ae-4257-8ce4-0a79a4acae13 at /dev/sda [ 2064.306207] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3c714cc-f3d2-41ce-ad59-60e8ca39ab69 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2064.315331] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47ababae-2736-43d5-88aa-9b3e74fc50ff {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2064.322322] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-78726121-ec15-417e-a338-54d27d9cef29 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2064.354341] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60eb1767-5335-4a94-a921-244937526030 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2064.364485] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54c832c2-2385-4c81-a5c0-1086c83ed64b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2064.376534] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ec59b36-f697-4370-8c5d-e91099adfee3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2064.392649] env[62824]: DEBUG nova.compute.provider_tree [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2064.412159] env[62824]: DEBUG nova.scheduler.client.report [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2064.415539] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-36d225ff-49d6-45d6-a9b0-6c8c3aa35373 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2064.426634] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48987535-b8a7-458a-8653-4e905210da1c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2064.474149] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a7f7330-074c-4838-b227-571d83d48993 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2064.482878] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-578d5bfb-e80c-464d-aab2-cc50cde9d23c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2064.499716] env[62824]: DEBUG nova.virt.block_device [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Updating existing volume attachment record: adcebb43-d174-446e-98ea-edc5509fce77 {{(pid=62824) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 2064.660466] env[62824]: DEBUG nova.virt.hardware [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2064.661289] env[62824]: DEBUG nova.virt.hardware [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2064.661655] env[62824]: DEBUG nova.virt.hardware [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2064.662347] env[62824]: DEBUG nova.virt.hardware [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2064.662975] env[62824]: DEBUG nova.virt.hardware [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2064.663652] env[62824]: DEBUG nova.virt.hardware [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2064.663997] env[62824]: DEBUG nova.virt.hardware [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2064.665539] env[62824]: DEBUG nova.virt.hardware [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2064.665539] env[62824]: DEBUG nova.virt.hardware [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2064.665539] env[62824]: DEBUG nova.virt.hardware [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2064.665539] env[62824]: DEBUG nova.virt.hardware [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2064.668791] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eea0f0c1-e20a-4fab-83c7-bc19e7025033 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2064.681206] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-491be291-aa6e-4e8e-9c90-65728201e5ed {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2064.687572] env[62824]: DEBUG nova.virt.hardware [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2064.687889] env[62824]: DEBUG nova.virt.hardware [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2064.688111] env[62824]: DEBUG nova.virt.hardware [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2064.688942] env[62824]: DEBUG nova.virt.hardware [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2064.688942] env[62824]: DEBUG nova.virt.hardware [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2064.688942] env[62824]: DEBUG nova.virt.hardware [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2064.688942] env[62824]: DEBUG nova.virt.hardware [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2064.689146] env[62824]: DEBUG nova.virt.hardware [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2064.689612] env[62824]: DEBUG nova.virt.hardware [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2064.689612] env[62824]: DEBUG nova.virt.hardware [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2064.689718] env[62824]: DEBUG nova.virt.hardware [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2064.692125] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c610c5c-09da-472c-8d39-f1049208cc21 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2064.711753] env[62824]: DEBUG oslo_vmware.rw_handles [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52031a0f-3082-1103-10b0-ff5f977f397f/disk-0.vmdk. {{(pid=62824) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 2064.712459] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d8:c8:ab', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f4399275-8e92-4448-be9e-d4984e93e89c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd85a0acc-fed6-4797-9f81-2aafa5bb6967', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2064.720293] env[62824]: DEBUG oslo.service.loopingcall [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2064.721162] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6a5dfe9-74a9-4244-9372-9ce9bda2ebcc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2064.725039] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6bc7736-6bae-4c49-a798-3bd68813c7cc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2064.731030] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2064.731321] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c8d17010-d2f9-4587-8c54-a4a2dfe36ae6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2064.752560] env[62824]: DEBUG oslo_vmware.rw_handles [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52031a0f-3082-1103-10b0-ff5f977f397f/disk-0.vmdk is in state: ready. {{(pid=62824) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 2064.752859] env[62824]: ERROR oslo_vmware.rw_handles [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52031a0f-3082-1103-10b0-ff5f977f397f/disk-0.vmdk due to incomplete transfer. [ 2064.762279] env[62824]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-93a069b1-c00c-49b1-ae27-47394e309e7f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2064.764194] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2064.764194] env[62824]: value = "task-2146178" [ 2064.764194] env[62824]: _type = "Task" [ 2064.764194] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2064.764581] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] [instance: ed7d3e95-f26e-40fd-a370-e17922bbff8e] Instance VIF info [] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2064.770421] env[62824]: DEBUG oslo.service.loopingcall [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2064.770779] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ed7d3e95-f26e-40fd-a370-e17922bbff8e] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2064.774565] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e4980808-286b-43db-905f-48d67b99b94d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2064.788142] env[62824]: DEBUG oslo_vmware.rw_handles [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52031a0f-3082-1103-10b0-ff5f977f397f/disk-0.vmdk. {{(pid=62824) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 2064.788306] env[62824]: DEBUG nova.virt.vmwareapi.images [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Uploaded image cc65c6c8-036d-4ab6-a8a0-770a7853932b to the Glance image server {{(pid=62824) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 2064.790690] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Destroying the VM {{(pid=62824) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 2064.791686] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-aeaadeae-5921-4b27-a9d2-5e3099377d25 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2064.797784] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146178, 'name': CreateVM_Task} progress is 15%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2064.799289] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2064.799289] env[62824]: value = "task-2146179" [ 2064.799289] env[62824]: _type = "Task" [ 2064.799289] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2064.800724] env[62824]: DEBUG oslo_vmware.api [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Waiting for the task: (returnval){ [ 2064.800724] env[62824]: value = "task-2146180" [ 2064.800724] env[62824]: _type = "Task" [ 2064.800724] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2064.816534] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146179, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2064.819973] env[62824]: DEBUG oslo_vmware.api [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2146180, 'name': Destroy_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2064.919590] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.164s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2064.920377] env[62824]: DEBUG nova.compute.manager [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 919e8a45-7810-4a8d-a1aa-5046b5ab059c] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2064.924041] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.859s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2064.926175] env[62824]: INFO nova.compute.claims [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 15479062-af75-4925-99b3-77d6a49751ad] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2064.960222] env[62824]: DEBUG nova.compute.manager [req-fc06ff86-eb4c-4a38-b0f5-cac73a7f4e42 req-db028232-0292-469c-9ca4-f44ae9c61b02 service nova] [instance: 4423d172-acdd-4c69-b3b8-ff166e1b8548] Received event network-vif-deleted-0f1c4fc4-1d12-4800-9861-68137c05d00f {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2065.212061] env[62824]: DEBUG nova.network.neutron [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Successfully updated port: d358ab3c-f4fb-41f3-954b-19de8b22ef15 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2065.283421] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146178, 'name': CreateVM_Task, 'duration_secs': 0.513162} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2065.283760] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2065.284597] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2065.284597] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2065.284778] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2065.285066] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-544bafa2-458a-4fd7-bc02-a5b6716d6196 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2065.290570] env[62824]: DEBUG oslo_vmware.api [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for the task: (returnval){ [ 2065.290570] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52fd9d5e-2c87-c64f-6369-c488f5086a5d" [ 2065.290570] env[62824]: _type = "Task" [ 2065.290570] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2065.300686] env[62824]: DEBUG oslo_vmware.api [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52fd9d5e-2c87-c64f-6369-c488f5086a5d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2065.317168] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146179, 'name': CreateVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2065.317441] env[62824]: DEBUG oslo_vmware.api [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2146180, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2065.432487] env[62824]: DEBUG nova.compute.utils [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2065.434105] env[62824]: DEBUG nova.compute.manager [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 919e8a45-7810-4a8d-a1aa-5046b5ab059c] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2065.434335] env[62824]: DEBUG nova.network.neutron [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 919e8a45-7810-4a8d-a1aa-5046b5ab059c] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2065.438654] env[62824]: DEBUG oslo_concurrency.lockutils [None req-70aefc57-09da-40ff-bfce-730a89b771bc tempest-VolumesAssistedSnapshotsTest-163918615 tempest-VolumesAssistedSnapshotsTest-163918615-project-admin] Acquiring lock "29b8dd5f-1855-490c-a01b-54840073a753" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2065.439013] env[62824]: DEBUG oslo_concurrency.lockutils [None req-70aefc57-09da-40ff-bfce-730a89b771bc tempest-VolumesAssistedSnapshotsTest-163918615 tempest-VolumesAssistedSnapshotsTest-163918615-project-admin] Lock "29b8dd5f-1855-490c-a01b-54840073a753" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2065.489919] env[62824]: DEBUG nova.policy [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c379c824a178444e99bb121716fc7887', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0c11aef7fb6247cb9b5272a6a063cd12', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2065.643092] env[62824]: DEBUG nova.compute.manager [req-a00710a0-d2e1-4cef-96d0-a3f96f30a64d req-6616bd3d-9275-4afd-924a-d45a0616e781 service nova] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Received event network-vif-plugged-d85a0acc-fed6-4797-9f81-2aafa5bb6967 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2065.643464] env[62824]: DEBUG oslo_concurrency.lockutils [req-a00710a0-d2e1-4cef-96d0-a3f96f30a64d req-6616bd3d-9275-4afd-924a-d45a0616e781 service nova] Acquiring lock "bbfcb2e3-9326-4548-b15b-e054cbfd192e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2065.643979] env[62824]: DEBUG oslo_concurrency.lockutils [req-a00710a0-d2e1-4cef-96d0-a3f96f30a64d req-6616bd3d-9275-4afd-924a-d45a0616e781 service nova] Lock "bbfcb2e3-9326-4548-b15b-e054cbfd192e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2065.644235] env[62824]: DEBUG oslo_concurrency.lockutils [req-a00710a0-d2e1-4cef-96d0-a3f96f30a64d req-6616bd3d-9275-4afd-924a-d45a0616e781 service nova] Lock "bbfcb2e3-9326-4548-b15b-e054cbfd192e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2065.644466] env[62824]: DEBUG nova.compute.manager [req-a00710a0-d2e1-4cef-96d0-a3f96f30a64d req-6616bd3d-9275-4afd-924a-d45a0616e781 service nova] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] No waiting events found dispatching network-vif-plugged-d85a0acc-fed6-4797-9f81-2aafa5bb6967 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2065.644675] env[62824]: WARNING nova.compute.manager [req-a00710a0-d2e1-4cef-96d0-a3f96f30a64d req-6616bd3d-9275-4afd-924a-d45a0616e781 service nova] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Received unexpected event network-vif-plugged-d85a0acc-fed6-4797-9f81-2aafa5bb6967 for instance with vm_state building and task_state spawning. [ 2065.644865] env[62824]: DEBUG nova.compute.manager [req-a00710a0-d2e1-4cef-96d0-a3f96f30a64d req-6616bd3d-9275-4afd-924a-d45a0616e781 service nova] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Received event network-changed-d85a0acc-fed6-4797-9f81-2aafa5bb6967 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2065.645094] env[62824]: DEBUG nova.compute.manager [req-a00710a0-d2e1-4cef-96d0-a3f96f30a64d req-6616bd3d-9275-4afd-924a-d45a0616e781 service nova] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Refreshing instance network info cache due to event network-changed-d85a0acc-fed6-4797-9f81-2aafa5bb6967. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2065.645335] env[62824]: DEBUG oslo_concurrency.lockutils [req-a00710a0-d2e1-4cef-96d0-a3f96f30a64d req-6616bd3d-9275-4afd-924a-d45a0616e781 service nova] Acquiring lock "refresh_cache-bbfcb2e3-9326-4548-b15b-e054cbfd192e" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2065.645578] env[62824]: DEBUG oslo_concurrency.lockutils [req-a00710a0-d2e1-4cef-96d0-a3f96f30a64d req-6616bd3d-9275-4afd-924a-d45a0616e781 service nova] Acquired lock "refresh_cache-bbfcb2e3-9326-4548-b15b-e054cbfd192e" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2065.645829] env[62824]: DEBUG nova.network.neutron [req-a00710a0-d2e1-4cef-96d0-a3f96f30a64d req-6616bd3d-9275-4afd-924a-d45a0616e781 service nova] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Refreshing network info cache for port d85a0acc-fed6-4797-9f81-2aafa5bb6967 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2065.716243] env[62824]: DEBUG oslo_concurrency.lockutils [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Acquiring lock "refresh_cache-2f315f9e-94e4-47ef-9503-ee92b59e5452" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2065.716395] env[62824]: DEBUG oslo_concurrency.lockutils [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Acquired lock "refresh_cache-2f315f9e-94e4-47ef-9503-ee92b59e5452" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2065.716571] env[62824]: DEBUG nova.network.neutron [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2065.808267] env[62824]: DEBUG oslo_vmware.api [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52fd9d5e-2c87-c64f-6369-c488f5086a5d, 'name': SearchDatastore_Task, 'duration_secs': 0.014453} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2065.817313] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2065.817688] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2065.818052] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2065.818933] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2065.818933] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2065.818933] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-29afae35-a22c-41c6-b94a-35882dcede34 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2065.828124] env[62824]: DEBUG oslo_vmware.api [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2146180, 'name': Destroy_Task, 'duration_secs': 0.606805} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2065.831710] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Destroyed the VM [ 2065.832232] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Deleting Snapshot of the VM instance {{(pid=62824) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 2065.832387] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146179, 'name': CreateVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2065.834323] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-8e0e4172-163f-4b73-a5ec-35f0a800794e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2065.835493] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2065.835926] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2065.836829] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-36353448-ad39-4540-8e77-b4222567d4b6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2065.844120] env[62824]: DEBUG oslo_vmware.api [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for the task: (returnval){ [ 2065.844120] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52b6feb0-2513-b29e-a193-14d1abddfab7" [ 2065.844120] env[62824]: _type = "Task" [ 2065.844120] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2065.851044] env[62824]: DEBUG oslo_vmware.api [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Waiting for the task: (returnval){ [ 2065.851044] env[62824]: value = "task-2146181" [ 2065.851044] env[62824]: _type = "Task" [ 2065.851044] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2065.861464] env[62824]: DEBUG oslo_vmware.api [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52b6feb0-2513-b29e-a193-14d1abddfab7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2065.868794] env[62824]: DEBUG oslo_vmware.api [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2146181, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2065.941242] env[62824]: DEBUG nova.compute.manager [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 919e8a45-7810-4a8d-a1aa-5046b5ab059c] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2065.947703] env[62824]: INFO nova.compute.manager [None req-70aefc57-09da-40ff-bfce-730a89b771bc tempest-VolumesAssistedSnapshotsTest-163918615 tempest-VolumesAssistedSnapshotsTest-163918615-project-admin] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Detaching volume f3a0e2f5-ded9-47fe-8fbe-8eeb061be84a [ 2065.974293] env[62824]: DEBUG nova.network.neutron [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 919e8a45-7810-4a8d-a1aa-5046b5ab059c] Successfully created port: ec8bab9e-6a51-4cc3-9fe8-4764dfd97d45 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2066.011955] env[62824]: INFO nova.virt.block_device [None req-70aefc57-09da-40ff-bfce-730a89b771bc tempest-VolumesAssistedSnapshotsTest-163918615 tempest-VolumesAssistedSnapshotsTest-163918615-project-admin] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Attempting to driver detach volume f3a0e2f5-ded9-47fe-8fbe-8eeb061be84a from mountpoint /dev/sdb [ 2066.012270] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-70aefc57-09da-40ff-bfce-730a89b771bc tempest-VolumesAssistedSnapshotsTest-163918615 tempest-VolumesAssistedSnapshotsTest-163918615-project-admin] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Volume detach. Driver type: vmdk {{(pid=62824) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 2066.012508] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-70aefc57-09da-40ff-bfce-730a89b771bc tempest-VolumesAssistedSnapshotsTest-163918615 tempest-VolumesAssistedSnapshotsTest-163918615-project-admin] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-438645', 'volume_id': 'f3a0e2f5-ded9-47fe-8fbe-8eeb061be84a', 'name': 'volume-f3a0e2f5-ded9-47fe-8fbe-8eeb061be84a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '29b8dd5f-1855-490c-a01b-54840073a753', 'attached_at': '', 'detached_at': '', 'volume_id': 'f3a0e2f5-ded9-47fe-8fbe-8eeb061be84a', 'serial': 'f3a0e2f5-ded9-47fe-8fbe-8eeb061be84a'} {{(pid=62824) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 2066.013533] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2837c85f-3f0b-4035-947f-bd8eba140273 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2066.043372] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31c81055-3244-45d7-80d1-dac53c5cb2fe {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2066.055661] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af3c7f5d-25c7-48a3-a25e-5ff762ce9cf2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2066.085699] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09eb45b8-846c-43e2-9a10-97e598ec7100 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2066.103979] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-70aefc57-09da-40ff-bfce-730a89b771bc tempest-VolumesAssistedSnapshotsTest-163918615 tempest-VolumesAssistedSnapshotsTest-163918615-project-admin] The volume has not been displaced from its original location: [datastore2] volume-f3a0e2f5-ded9-47fe-8fbe-8eeb061be84a/volume-f3a0e2f5-ded9-47fe-8fbe-8eeb061be84a.vmdk. No consolidation needed. {{(pid=62824) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 2066.109427] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-70aefc57-09da-40ff-bfce-730a89b771bc tempest-VolumesAssistedSnapshotsTest-163918615 tempest-VolumesAssistedSnapshotsTest-163918615-project-admin] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Reconfiguring VM instance instance-0000000d to detach disk 2001 {{(pid=62824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 2066.116415] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c7c7f24b-55a0-41b7-b706-e58c0ba284a7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2066.138220] env[62824]: DEBUG oslo_vmware.rw_handles [None req-752f903a-bdbd-4c93-8465-c657fedd4220 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52cfba3a-f3ed-caee-c5f8-4849298710d3/disk-0.vmdk. {{(pid=62824) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 2066.139759] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a1520d5-4a92-46e6-9641-8fd7c8d87f29 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2066.148486] env[62824]: DEBUG oslo_vmware.api [None req-70aefc57-09da-40ff-bfce-730a89b771bc tempest-VolumesAssistedSnapshotsTest-163918615 tempest-VolumesAssistedSnapshotsTest-163918615-project-admin] Waiting for the task: (returnval){ [ 2066.148486] env[62824]: value = "task-2146182" [ 2066.148486] env[62824]: _type = "Task" [ 2066.148486] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2066.156568] env[62824]: DEBUG oslo_vmware.rw_handles [None req-752f903a-bdbd-4c93-8465-c657fedd4220 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52cfba3a-f3ed-caee-c5f8-4849298710d3/disk-0.vmdk is in state: ready. {{(pid=62824) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 2066.156725] env[62824]: ERROR oslo_vmware.rw_handles [None req-752f903a-bdbd-4c93-8465-c657fedd4220 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52cfba3a-f3ed-caee-c5f8-4849298710d3/disk-0.vmdk due to incomplete transfer. [ 2066.157899] env[62824]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-68588443-bb4e-4453-a96c-2ffc288b5938 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2066.167739] env[62824]: DEBUG oslo_vmware.api [None req-70aefc57-09da-40ff-bfce-730a89b771bc tempest-VolumesAssistedSnapshotsTest-163918615 tempest-VolumesAssistedSnapshotsTest-163918615-project-admin] Task: {'id': task-2146182, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2066.169475] env[62824]: DEBUG oslo_vmware.rw_handles [None req-752f903a-bdbd-4c93-8465-c657fedd4220 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52cfba3a-f3ed-caee-c5f8-4849298710d3/disk-0.vmdk. {{(pid=62824) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 2066.169775] env[62824]: DEBUG nova.virt.vmwareapi.images [None req-752f903a-bdbd-4c93-8465-c657fedd4220 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 88bad2dd-dce7-41df-b56c-93a5d054c11e] Uploaded image d11d1462-ceff-4a2d-bb14-0cf931545d6f to the Glance image server {{(pid=62824) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 2066.171533] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-752f903a-bdbd-4c93-8465-c657fedd4220 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 88bad2dd-dce7-41df-b56c-93a5d054c11e] Destroying the VM {{(pid=62824) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 2066.172103] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-a6e09dd0-4b68-4c3e-90e4-b4e52898207f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2066.181735] env[62824]: DEBUG oslo_vmware.api [None req-752f903a-bdbd-4c93-8465-c657fedd4220 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the task: (returnval){ [ 2066.181735] env[62824]: value = "task-2146183" [ 2066.181735] env[62824]: _type = "Task" [ 2066.181735] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2066.191091] env[62824]: DEBUG oslo_vmware.api [None req-752f903a-bdbd-4c93-8465-c657fedd4220 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146183, 'name': Destroy_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2066.319889] env[62824]: DEBUG nova.network.neutron [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2066.326212] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146179, 'name': CreateVM_Task, 'duration_secs': 1.072463} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2066.327376] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ed7d3e95-f26e-40fd-a370-e17922bbff8e] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2066.332032] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2066.332032] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2066.332032] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2066.333781] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ff015125-d022-4b2c-bb85-dddcfd1fd436 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2066.341821] env[62824]: DEBUG oslo_vmware.api [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Waiting for the task: (returnval){ [ 2066.341821] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]520fe4b9-cfbf-3f95-f604-5cdf14ac5974" [ 2066.341821] env[62824]: _type = "Task" [ 2066.341821] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2066.355601] env[62824]: DEBUG oslo_vmware.api [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]520fe4b9-cfbf-3f95-f604-5cdf14ac5974, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2066.370485] env[62824]: DEBUG oslo_vmware.api [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52b6feb0-2513-b29e-a193-14d1abddfab7, 'name': SearchDatastore_Task, 'duration_secs': 0.01584} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2066.370890] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9ebf8131-0487-43bc-b97e-fd82fdfa60ef {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2066.380141] env[62824]: INFO nova.compute.manager [None req-03bb66d7-d558-4473-adb2-183a408a723b tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: b0456b8a-348a-4503-a92c-58e5ab455d1c] Rescuing [ 2066.380426] env[62824]: DEBUG oslo_concurrency.lockutils [None req-03bb66d7-d558-4473-adb2-183a408a723b tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Acquiring lock "refresh_cache-b0456b8a-348a-4503-a92c-58e5ab455d1c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2066.380580] env[62824]: DEBUG oslo_concurrency.lockutils [None req-03bb66d7-d558-4473-adb2-183a408a723b tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Acquired lock "refresh_cache-b0456b8a-348a-4503-a92c-58e5ab455d1c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2066.380796] env[62824]: DEBUG nova.network.neutron [None req-03bb66d7-d558-4473-adb2-183a408a723b tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: b0456b8a-348a-4503-a92c-58e5ab455d1c] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2066.381994] env[62824]: DEBUG oslo_vmware.api [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2146181, 'name': RemoveSnapshot_Task, 'duration_secs': 0.459352} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2066.383544] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Deleted Snapshot of the VM instance {{(pid=62824) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 2066.383544] env[62824]: DEBUG nova.compute.manager [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2066.385257] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3156840e-628a-4920-9b32-4bc87cc92833 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2066.391952] env[62824]: DEBUG oslo_vmware.api [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for the task: (returnval){ [ 2066.391952] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52c76005-8a46-e0aa-d831-fe67672c8a07" [ 2066.391952] env[62824]: _type = "Task" [ 2066.391952] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2066.412264] env[62824]: DEBUG oslo_vmware.api [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52c76005-8a46-e0aa-d831-fe67672c8a07, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2066.555035] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e99f51a7-e71c-443f-812d-3cdaa01ba099 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Acquiring lock "21c05725-3dad-4965-98d0-0622ebcc8ece" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2066.555035] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e99f51a7-e71c-443f-812d-3cdaa01ba099 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Lock "21c05725-3dad-4965-98d0-0622ebcc8ece" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2066.555035] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e99f51a7-e71c-443f-812d-3cdaa01ba099 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Acquiring lock "21c05725-3dad-4965-98d0-0622ebcc8ece-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2066.555035] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e99f51a7-e71c-443f-812d-3cdaa01ba099 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Lock "21c05725-3dad-4965-98d0-0622ebcc8ece-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2066.555035] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e99f51a7-e71c-443f-812d-3cdaa01ba099 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Lock "21c05725-3dad-4965-98d0-0622ebcc8ece-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2066.556802] env[62824]: INFO nova.compute.manager [None req-e99f51a7-e71c-443f-812d-3cdaa01ba099 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 21c05725-3dad-4965-98d0-0622ebcc8ece] Terminating instance [ 2066.632911] env[62824]: DEBUG nova.compute.manager [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2066.633537] env[62824]: DEBUG nova.virt.hardware [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2066.633786] env[62824]: DEBUG nova.virt.hardware [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2066.633943] env[62824]: DEBUG nova.virt.hardware [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2066.634146] env[62824]: DEBUG nova.virt.hardware [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2066.634293] env[62824]: DEBUG nova.virt.hardware [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2066.634509] env[62824]: DEBUG nova.virt.hardware [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2066.634691] env[62824]: DEBUG nova.virt.hardware [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2066.634776] env[62824]: DEBUG nova.virt.hardware [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2066.634941] env[62824]: DEBUG nova.virt.hardware [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2066.635117] env[62824]: DEBUG nova.virt.hardware [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2066.635302] env[62824]: DEBUG nova.virt.hardware [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2066.636177] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0496091-03b5-46bd-b283-19cd873c8b85 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2066.639428] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26e63a56-a622-42b2-973a-e2f64bf859cb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2066.650053] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dda22c2e-808b-4f62-b35d-19982071c60e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2066.654712] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cce56013-6f5e-432a-ac02-8af59f2af5a4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2066.677155] env[62824]: DEBUG oslo_vmware.api [None req-70aefc57-09da-40ff-bfce-730a89b771bc tempest-VolumesAssistedSnapshotsTest-163918615 tempest-VolumesAssistedSnapshotsTest-163918615-project-admin] Task: {'id': task-2146182, 'name': ReconfigVM_Task, 'duration_secs': 0.352379} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2066.701200] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-70aefc57-09da-40ff-bfce-730a89b771bc tempest-VolumesAssistedSnapshotsTest-163918615 tempest-VolumesAssistedSnapshotsTest-163918615-project-admin] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Reconfigured VM instance instance-0000000d to detach disk 2001 {{(pid=62824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 2066.706025] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e50a52d8-5f30-41f2-8d8d-47e1c3358cf0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2066.718997] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-209d9436-6d60-41cd-bbe0-3d809c203e64 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2066.732444] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9212b7db-b39e-4f72-a9ab-2bbdfae9243e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2066.736406] env[62824]: DEBUG oslo_vmware.api [None req-752f903a-bdbd-4c93-8465-c657fedd4220 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146183, 'name': Destroy_Task, 'duration_secs': 0.512516} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2066.736690] env[62824]: DEBUG oslo_vmware.api [None req-70aefc57-09da-40ff-bfce-730a89b771bc tempest-VolumesAssistedSnapshotsTest-163918615 tempest-VolumesAssistedSnapshotsTest-163918615-project-admin] Waiting for the task: (returnval){ [ 2066.736690] env[62824]: value = "task-2146184" [ 2066.736690] env[62824]: _type = "Task" [ 2066.736690] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2066.737067] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-752f903a-bdbd-4c93-8465-c657fedd4220 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 88bad2dd-dce7-41df-b56c-93a5d054c11e] Destroyed the VM [ 2066.737157] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-752f903a-bdbd-4c93-8465-c657fedd4220 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 88bad2dd-dce7-41df-b56c-93a5d054c11e] Deleting Snapshot of the VM instance {{(pid=62824) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 2066.737797] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-8454cfe3-94b8-47e0-b85f-8d47bcb0735c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2066.751759] env[62824]: DEBUG nova.compute.provider_tree [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2066.759766] env[62824]: DEBUG oslo_vmware.api [None req-70aefc57-09da-40ff-bfce-730a89b771bc tempest-VolumesAssistedSnapshotsTest-163918615 tempest-VolumesAssistedSnapshotsTest-163918615-project-admin] Task: {'id': task-2146184, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2066.761154] env[62824]: DEBUG oslo_vmware.api [None req-752f903a-bdbd-4c93-8465-c657fedd4220 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the task: (returnval){ [ 2066.761154] env[62824]: value = "task-2146185" [ 2066.761154] env[62824]: _type = "Task" [ 2066.761154] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2066.771671] env[62824]: DEBUG oslo_vmware.api [None req-752f903a-bdbd-4c93-8465-c657fedd4220 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146185, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2066.851307] env[62824]: DEBUG oslo_vmware.api [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]520fe4b9-cfbf-3f95-f604-5cdf14ac5974, 'name': SearchDatastore_Task, 'duration_secs': 0.031662} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2066.851616] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2066.851849] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] [instance: ed7d3e95-f26e-40fd-a370-e17922bbff8e] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2066.852070] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2066.904366] env[62824]: DEBUG oslo_vmware.api [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52c76005-8a46-e0aa-d831-fe67672c8a07, 'name': SearchDatastore_Task, 'duration_secs': 0.024823} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2066.904575] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2066.904805] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] bbfcb2e3-9326-4548-b15b-e054cbfd192e/bbfcb2e3-9326-4548-b15b-e054cbfd192e.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2066.905060] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2066.905282] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2066.907741] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a9ab8af4-2328-4f65-9806-f672d8d768d2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2066.910170] env[62824]: INFO nova.compute.manager [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Shelve offloading [ 2066.911670] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-85472896-8337-4511-87ee-1e9d41794d4c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2066.920991] env[62824]: DEBUG oslo_vmware.api [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for the task: (returnval){ [ 2066.920991] env[62824]: value = "task-2146186" [ 2066.920991] env[62824]: _type = "Task" [ 2066.920991] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2066.927914] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2066.928225] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2066.929589] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e28064e8-302b-480d-b238-ccf2ffe31c59 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2066.936608] env[62824]: DEBUG oslo_vmware.api [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2146186, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2066.938370] env[62824]: DEBUG nova.network.neutron [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Updating instance_info_cache with network_info: [{"id": "d358ab3c-f4fb-41f3-954b-19de8b22ef15", "address": "fa:16:3e:12:72:e1", "network": {"id": "2ed8a86a-9c14-47b4-8cb4-e8496d289a6e", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1207171730-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "87239f55d6d44606b23ca0c25d9fb4d9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7514a465-f1a4-4a8b-b76b-726b1a9d7e2f", "external-id": "nsx-vlan-transportzone-36", "segmentation_id": 36, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd358ab3c-f4", "ovs_interfaceid": "d358ab3c-f4fb-41f3-954b-19de8b22ef15", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2066.940568] env[62824]: DEBUG nova.network.neutron [req-a00710a0-d2e1-4cef-96d0-a3f96f30a64d req-6616bd3d-9275-4afd-924a-d45a0616e781 service nova] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Updated VIF entry in instance network info cache for port d85a0acc-fed6-4797-9f81-2aafa5bb6967. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2066.940947] env[62824]: DEBUG nova.network.neutron [req-a00710a0-d2e1-4cef-96d0-a3f96f30a64d req-6616bd3d-9275-4afd-924a-d45a0616e781 service nova] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Updating instance_info_cache with network_info: [{"id": "d85a0acc-fed6-4797-9f81-2aafa5bb6967", "address": "fa:16:3e:d8:c8:ab", "network": {"id": "a410f4f3-d072-4034-b797-8210e1a79093", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1482574391-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e10208040df34eb5adfe8dcbc76043d5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f4399275-8e92-4448-be9e-d4984e93e89c", "external-id": "nsx-vlan-transportzone-192", "segmentation_id": 192, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd85a0acc-fe", "ovs_interfaceid": "d85a0acc-fed6-4797-9f81-2aafa5bb6967", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2066.943781] env[62824]: DEBUG oslo_vmware.api [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Waiting for the task: (returnval){ [ 2066.943781] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5208b219-ed4c-9121-e57e-0ecdf7efb743" [ 2066.943781] env[62824]: _type = "Task" [ 2066.943781] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2066.956968] env[62824]: DEBUG oslo_vmware.api [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5208b219-ed4c-9121-e57e-0ecdf7efb743, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2066.960237] env[62824]: DEBUG nova.compute.manager [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 919e8a45-7810-4a8d-a1aa-5046b5ab059c] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2066.992103] env[62824]: DEBUG nova.virt.hardware [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2066.992282] env[62824]: DEBUG nova.virt.hardware [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2066.992416] env[62824]: DEBUG nova.virt.hardware [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2066.992596] env[62824]: DEBUG nova.virt.hardware [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2066.992743] env[62824]: DEBUG nova.virt.hardware [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2066.992907] env[62824]: DEBUG nova.virt.hardware [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2066.993148] env[62824]: DEBUG nova.virt.hardware [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2066.993312] env[62824]: DEBUG nova.virt.hardware [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2066.993524] env[62824]: DEBUG nova.virt.hardware [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2066.993828] env[62824]: DEBUG nova.virt.hardware [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2066.993939] env[62824]: DEBUG nova.virt.hardware [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2066.994831] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d71e5a96-ae40-48f4-98c7-d994c9bce2c4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2067.018768] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-505cebac-dd9a-447f-be55-a563ebf25726 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2067.060881] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e99f51a7-e71c-443f-812d-3cdaa01ba099 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Acquiring lock "refresh_cache-21c05725-3dad-4965-98d0-0622ebcc8ece" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2067.061165] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e99f51a7-e71c-443f-812d-3cdaa01ba099 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Acquired lock "refresh_cache-21c05725-3dad-4965-98d0-0622ebcc8ece" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2067.061367] env[62824]: DEBUG nova.network.neutron [None req-e99f51a7-e71c-443f-812d-3cdaa01ba099 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 21c05725-3dad-4965-98d0-0622ebcc8ece] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2067.250078] env[62824]: DEBUG oslo_vmware.api [None req-70aefc57-09da-40ff-bfce-730a89b771bc tempest-VolumesAssistedSnapshotsTest-163918615 tempest-VolumesAssistedSnapshotsTest-163918615-project-admin] Task: {'id': task-2146184, 'name': ReconfigVM_Task, 'duration_secs': 0.142395} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2067.250444] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-70aefc57-09da-40ff-bfce-730a89b771bc tempest-VolumesAssistedSnapshotsTest-163918615 tempest-VolumesAssistedSnapshotsTest-163918615-project-admin] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-438645', 'volume_id': 'f3a0e2f5-ded9-47fe-8fbe-8eeb061be84a', 'name': 'volume-f3a0e2f5-ded9-47fe-8fbe-8eeb061be84a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '29b8dd5f-1855-490c-a01b-54840073a753', 'attached_at': '', 'detached_at': '', 'volume_id': 'f3a0e2f5-ded9-47fe-8fbe-8eeb061be84a', 'serial': 'f3a0e2f5-ded9-47fe-8fbe-8eeb061be84a'} {{(pid=62824) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 2067.255093] env[62824]: DEBUG nova.scheduler.client.report [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2067.284178] env[62824]: DEBUG oslo_vmware.api [None req-752f903a-bdbd-4c93-8465-c657fedd4220 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146185, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2067.287501] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c94e74ba-f9b2-469b-be3d-b77fc48e94ea tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquiring lock "interface-f4d63a93-23af-470c-b36a-662af81dc386-c34f030d-46da-4a7a-9703-5acb2e526d50" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2067.287648] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c94e74ba-f9b2-469b-be3d-b77fc48e94ea tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lock "interface-f4d63a93-23af-470c-b36a-662af81dc386-c34f030d-46da-4a7a-9703-5acb2e526d50" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2067.415856] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2067.416215] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ac3020fd-7016-4b08-a4ce-223d7d6e9050 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2067.422474] env[62824]: DEBUG nova.network.neutron [None req-03bb66d7-d558-4473-adb2-183a408a723b tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: b0456b8a-348a-4503-a92c-58e5ab455d1c] Updating instance_info_cache with network_info: [{"id": "0534c02e-d202-458d-9289-7d0ad2c5596f", "address": "fa:16:3e:b0:2c:b9", "network": {"id": "bcf30d77-cac1-4deb-971d-efb7c1fb8354", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1030043097-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "0503b434636c4ef5bae8db1b0c74d2ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3cc0a33d-17c0-4b87-b48f-413a87a4cc6a", "external-id": "nsx-vlan-transportzone-865", "segmentation_id": 865, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0534c02e-d2", "ovs_interfaceid": "0534c02e-d202-458d-9289-7d0ad2c5596f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2067.429518] env[62824]: DEBUG oslo_vmware.api [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Waiting for the task: (returnval){ [ 2067.429518] env[62824]: value = "task-2146187" [ 2067.429518] env[62824]: _type = "Task" [ 2067.429518] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2067.432645] env[62824]: DEBUG nova.compute.manager [req-d9d0bdfd-3f51-44f3-bc13-375198d3d49d req-23492b4b-2ec3-4e43-9d25-d4b9e9f6dfdc service nova] [instance: 82382932-7302-4441-a6f8-9aa2300ec0f6] Received event network-vif-deleted-760062c3-a8b6-4863-a557-0d44a27e3bb3 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2067.432814] env[62824]: DEBUG nova.compute.manager [req-d9d0bdfd-3f51-44f3-bc13-375198d3d49d req-23492b4b-2ec3-4e43-9d25-d4b9e9f6dfdc service nova] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Received event network-vif-plugged-d358ab3c-f4fb-41f3-954b-19de8b22ef15 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2067.432996] env[62824]: DEBUG oslo_concurrency.lockutils [req-d9d0bdfd-3f51-44f3-bc13-375198d3d49d req-23492b4b-2ec3-4e43-9d25-d4b9e9f6dfdc service nova] Acquiring lock "2f315f9e-94e4-47ef-9503-ee92b59e5452-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2067.433219] env[62824]: DEBUG oslo_concurrency.lockutils [req-d9d0bdfd-3f51-44f3-bc13-375198d3d49d req-23492b4b-2ec3-4e43-9d25-d4b9e9f6dfdc service nova] Lock "2f315f9e-94e4-47ef-9503-ee92b59e5452-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2067.433374] env[62824]: DEBUG oslo_concurrency.lockutils [req-d9d0bdfd-3f51-44f3-bc13-375198d3d49d req-23492b4b-2ec3-4e43-9d25-d4b9e9f6dfdc service nova] Lock "2f315f9e-94e4-47ef-9503-ee92b59e5452-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2067.433618] env[62824]: DEBUG nova.compute.manager [req-d9d0bdfd-3f51-44f3-bc13-375198d3d49d req-23492b4b-2ec3-4e43-9d25-d4b9e9f6dfdc service nova] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] No waiting events found dispatching network-vif-plugged-d358ab3c-f4fb-41f3-954b-19de8b22ef15 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2067.433811] env[62824]: WARNING nova.compute.manager [req-d9d0bdfd-3f51-44f3-bc13-375198d3d49d req-23492b4b-2ec3-4e43-9d25-d4b9e9f6dfdc service nova] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Received unexpected event network-vif-plugged-d358ab3c-f4fb-41f3-954b-19de8b22ef15 for instance with vm_state building and task_state spawning. [ 2067.434029] env[62824]: DEBUG nova.compute.manager [req-d9d0bdfd-3f51-44f3-bc13-375198d3d49d req-23492b4b-2ec3-4e43-9d25-d4b9e9f6dfdc service nova] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Received event network-changed-d358ab3c-f4fb-41f3-954b-19de8b22ef15 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2067.434130] env[62824]: DEBUG nova.compute.manager [req-d9d0bdfd-3f51-44f3-bc13-375198d3d49d req-23492b4b-2ec3-4e43-9d25-d4b9e9f6dfdc service nova] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Refreshing instance network info cache due to event network-changed-d358ab3c-f4fb-41f3-954b-19de8b22ef15. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2067.434297] env[62824]: DEBUG oslo_concurrency.lockutils [req-d9d0bdfd-3f51-44f3-bc13-375198d3d49d req-23492b4b-2ec3-4e43-9d25-d4b9e9f6dfdc service nova] Acquiring lock "refresh_cache-2f315f9e-94e4-47ef-9503-ee92b59e5452" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2067.443023] env[62824]: DEBUG oslo_concurrency.lockutils [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Releasing lock "refresh_cache-2f315f9e-94e4-47ef-9503-ee92b59e5452" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2067.443352] env[62824]: DEBUG nova.compute.manager [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Instance network_info: |[{"id": "d358ab3c-f4fb-41f3-954b-19de8b22ef15", "address": "fa:16:3e:12:72:e1", "network": {"id": "2ed8a86a-9c14-47b4-8cb4-e8496d289a6e", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1207171730-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "87239f55d6d44606b23ca0c25d9fb4d9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7514a465-f1a4-4a8b-b76b-726b1a9d7e2f", "external-id": "nsx-vlan-transportzone-36", "segmentation_id": 36, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd358ab3c-f4", "ovs_interfaceid": "d358ab3c-f4fb-41f3-954b-19de8b22ef15", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2067.444291] env[62824]: DEBUG oslo_vmware.api [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2146186, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2067.444675] env[62824]: DEBUG oslo_concurrency.lockutils [req-a00710a0-d2e1-4cef-96d0-a3f96f30a64d req-6616bd3d-9275-4afd-924a-d45a0616e781 service nova] Releasing lock "refresh_cache-bbfcb2e3-9326-4548-b15b-e054cbfd192e" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2067.444834] env[62824]: DEBUG nova.compute.manager [req-a00710a0-d2e1-4cef-96d0-a3f96f30a64d req-6616bd3d-9275-4afd-924a-d45a0616e781 service nova] [instance: 2afecec5-763d-4616-b690-41d3101cfc47] Received event network-vif-deleted-3d71b7f8-4ccd-4ae9-87c1-c4f03a8a22ed {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2067.445230] env[62824]: DEBUG oslo_concurrency.lockutils [req-d9d0bdfd-3f51-44f3-bc13-375198d3d49d req-23492b4b-2ec3-4e43-9d25-d4b9e9f6dfdc service nova] Acquired lock "refresh_cache-2f315f9e-94e4-47ef-9503-ee92b59e5452" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2067.445402] env[62824]: DEBUG nova.network.neutron [req-d9d0bdfd-3f51-44f3-bc13-375198d3d49d req-23492b4b-2ec3-4e43-9d25-d4b9e9f6dfdc service nova] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Refreshing network info cache for port d358ab3c-f4fb-41f3-954b-19de8b22ef15 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2067.446667] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:12:72:e1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7514a465-f1a4-4a8b-b76b-726b1a9d7e2f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd358ab3c-f4fb-41f3-954b-19de8b22ef15', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2067.454348] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Creating folder: Project (87239f55d6d44606b23ca0c25d9fb4d9). Parent ref: group-v438503. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 2067.455925] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e928a666-37aa-48ce-87ac-9a6efd890430 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2067.468113] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] VM already powered off {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 2067.468348] env[62824]: DEBUG nova.compute.manager [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2067.470019] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3cf5b4e-bf6f-40f6-8ef6-1c6f67254eca {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2067.477949] env[62824]: DEBUG oslo_vmware.api [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5208b219-ed4c-9121-e57e-0ecdf7efb743, 'name': SearchDatastore_Task, 'duration_secs': 0.021312} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2067.479366] env[62824]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 2067.479573] env[62824]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=62824) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 2067.480943] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Folder already exists: Project (87239f55d6d44606b23ca0c25d9fb4d9). Parent ref: group-v438503. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1609}} [ 2067.481193] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Creating folder: Instances. Parent ref: group-v438582. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 2067.481592] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9878fcaa-412d-4503-8878-f7366d8079bc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2067.484826] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7d78d687-2ff7-4781-a768-bf74dd0c9cbb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2067.493926] env[62824]: DEBUG oslo_concurrency.lockutils [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Acquiring lock "refresh_cache-b8cc8cd3-ea03-40bf-b867-7ad193365552" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2067.493926] env[62824]: DEBUG oslo_concurrency.lockutils [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Acquired lock "refresh_cache-b8cc8cd3-ea03-40bf-b867-7ad193365552" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2067.494086] env[62824]: DEBUG nova.network.neutron [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2067.499191] env[62824]: DEBUG oslo_vmware.api [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Waiting for the task: (returnval){ [ 2067.499191] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5284887c-44e8-bba2-ff0b-3e2f59861300" [ 2067.499191] env[62824]: _type = "Task" [ 2067.499191] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2067.506343] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Created folder: Instances in parent group-v438582. [ 2067.506965] env[62824]: DEBUG oslo.service.loopingcall [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2067.511045] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2067.511045] env[62824]: DEBUG oslo_vmware.api [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5284887c-44e8-bba2-ff0b-3e2f59861300, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2067.511045] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-acb041e5-82c0-45e3-b2e4-d47042269a57 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2067.533349] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2067.533349] env[62824]: value = "task-2146190" [ 2067.533349] env[62824]: _type = "Task" [ 2067.533349] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2067.543444] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146190, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2067.592135] env[62824]: DEBUG nova.network.neutron [None req-e99f51a7-e71c-443f-812d-3cdaa01ba099 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 21c05725-3dad-4965-98d0-0622ebcc8ece] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2067.740448] env[62824]: DEBUG nova.network.neutron [None req-e99f51a7-e71c-443f-812d-3cdaa01ba099 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 21c05725-3dad-4965-98d0-0622ebcc8ece] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2067.761023] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.837s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2067.762455] env[62824]: DEBUG nova.compute.manager [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 15479062-af75-4925-99b3-77d6a49751ad] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2067.765956] env[62824]: DEBUG oslo_concurrency.lockutils [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.289s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2067.768051] env[62824]: INFO nova.compute.claims [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: a6cd2032-de60-4f78-bf1e-79801d049df0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2067.785476] env[62824]: DEBUG oslo_vmware.api [None req-752f903a-bdbd-4c93-8465-c657fedd4220 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146185, 'name': RemoveSnapshot_Task, 'duration_secs': 0.627108} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2067.785707] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-752f903a-bdbd-4c93-8465-c657fedd4220 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 88bad2dd-dce7-41df-b56c-93a5d054c11e] Deleted Snapshot of the VM instance {{(pid=62824) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 2067.786198] env[62824]: INFO nova.compute.manager [None req-752f903a-bdbd-4c93-8465-c657fedd4220 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 88bad2dd-dce7-41df-b56c-93a5d054c11e] Took 15.08 seconds to snapshot the instance on the hypervisor. [ 2067.791049] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c94e74ba-f9b2-469b-be3d-b77fc48e94ea tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquiring lock "f4d63a93-23af-470c-b36a-662af81dc386" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2067.791049] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c94e74ba-f9b2-469b-be3d-b77fc48e94ea tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquired lock "f4d63a93-23af-470c-b36a-662af81dc386" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2067.791882] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54bf3ac2-ffed-4a59-bd45-e454ccf74589 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2067.817501] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-412b34f3-9d49-4a14-9c94-a09a2374c6f8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2067.860782] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c94e74ba-f9b2-469b-be3d-b77fc48e94ea tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Reconfiguring VM to detach interface {{(pid=62824) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 2067.862296] env[62824]: DEBUG nova.objects.instance [None req-70aefc57-09da-40ff-bfce-730a89b771bc tempest-VolumesAssistedSnapshotsTest-163918615 tempest-VolumesAssistedSnapshotsTest-163918615-project-admin] Lazy-loading 'flavor' on Instance uuid 29b8dd5f-1855-490c-a01b-54840073a753 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2067.863805] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-65af8030-a282-4d9d-b10e-76892192c512 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2067.886857] env[62824]: DEBUG oslo_vmware.api [None req-c94e74ba-f9b2-469b-be3d-b77fc48e94ea tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Waiting for the task: (returnval){ [ 2067.886857] env[62824]: value = "task-2146191" [ 2067.886857] env[62824]: _type = "Task" [ 2067.886857] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2067.906044] env[62824]: DEBUG oslo_vmware.api [None req-c94e74ba-f9b2-469b-be3d-b77fc48e94ea tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146191, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2067.932326] env[62824]: DEBUG oslo_concurrency.lockutils [None req-03bb66d7-d558-4473-adb2-183a408a723b tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Releasing lock "refresh_cache-b0456b8a-348a-4503-a92c-58e5ab455d1c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2067.941424] env[62824]: DEBUG oslo_vmware.api [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2146186, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.527726} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2067.942022] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] bbfcb2e3-9326-4548-b15b-e054cbfd192e/bbfcb2e3-9326-4548-b15b-e054cbfd192e.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2067.942022] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2067.943169] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-82c7f798-f4f8-4dfa-a56b-b92da21ed10d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2067.950559] env[62824]: DEBUG oslo_vmware.api [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for the task: (returnval){ [ 2067.950559] env[62824]: value = "task-2146192" [ 2067.950559] env[62824]: _type = "Task" [ 2067.950559] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2067.962431] env[62824]: DEBUG oslo_vmware.api [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2146192, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2068.018171] env[62824]: DEBUG oslo_vmware.api [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5284887c-44e8-bba2-ff0b-3e2f59861300, 'name': SearchDatastore_Task, 'duration_secs': 0.021077} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2068.018455] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2068.018714] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] ed7d3e95-f26e-40fd-a370-e17922bbff8e/ed7d3e95-f26e-40fd-a370-e17922bbff8e.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2068.018979] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-82bd65a5-7781-44e6-a064-103fbd333e85 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2068.026811] env[62824]: DEBUG oslo_vmware.api [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Waiting for the task: (returnval){ [ 2068.026811] env[62824]: value = "task-2146193" [ 2068.026811] env[62824]: _type = "Task" [ 2068.026811] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2068.036348] env[62824]: DEBUG oslo_vmware.api [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Task: {'id': task-2146193, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2068.050463] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146190, 'name': CreateVM_Task, 'duration_secs': 0.513218} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2068.050630] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2068.051673] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'device_type': None, 'boot_index': 0, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-438591', 'volume_id': '8d462122-e7ae-4257-8ce4-0a79a4acae13', 'name': 'volume-8d462122-e7ae-4257-8ce4-0a79a4acae13', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '2f315f9e-94e4-47ef-9503-ee92b59e5452', 'attached_at': '', 'detached_at': '', 'volume_id': '8d462122-e7ae-4257-8ce4-0a79a4acae13', 'serial': '8d462122-e7ae-4257-8ce4-0a79a4acae13'}, 'disk_bus': None, 'mount_device': '/dev/sda', 'delete_on_termination': True, 'attachment_id': 'adcebb43-d174-446e-98ea-edc5509fce77', 'guest_format': None, 'volume_type': None}], 'swap': None} {{(pid=62824) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 2068.051877] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Root volume attach. Driver type: vmdk {{(pid=62824) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 2068.053069] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5298df0-47e0-4d0f-819b-a7303e7a6642 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2068.063296] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54ac4936-3758-4084-a949-fc8f3d85298a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2068.071138] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30bdd99b-4510-4644-807e-c1ae9189e4fb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2068.079063] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-6a3b7aa1-a2b7-44e7-8c7e-6b4b345782e5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2068.087991] env[62824]: DEBUG oslo_vmware.api [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Waiting for the task: (returnval){ [ 2068.087991] env[62824]: value = "task-2146194" [ 2068.087991] env[62824]: _type = "Task" [ 2068.087991] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2068.102023] env[62824]: DEBUG oslo_vmware.api [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Task: {'id': task-2146194, 'name': RelocateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2068.244096] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e99f51a7-e71c-443f-812d-3cdaa01ba099 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Releasing lock "refresh_cache-21c05725-3dad-4965-98d0-0622ebcc8ece" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2068.244521] env[62824]: DEBUG nova.compute.manager [None req-e99f51a7-e71c-443f-812d-3cdaa01ba099 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 21c05725-3dad-4965-98d0-0622ebcc8ece] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2068.244655] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e99f51a7-e71c-443f-812d-3cdaa01ba099 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 21c05725-3dad-4965-98d0-0622ebcc8ece] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2068.245610] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0171477f-8f64-4fdb-b689-65158f80173d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2068.258215] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-e99f51a7-e71c-443f-812d-3cdaa01ba099 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 21c05725-3dad-4965-98d0-0622ebcc8ece] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2068.258689] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-43b52c41-a18e-4ed1-b967-2e91a27683cd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2068.268056] env[62824]: DEBUG oslo_vmware.api [None req-e99f51a7-e71c-443f-812d-3cdaa01ba099 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Waiting for the task: (returnval){ [ 2068.268056] env[62824]: value = "task-2146195" [ 2068.268056] env[62824]: _type = "Task" [ 2068.268056] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2068.273195] env[62824]: DEBUG nova.compute.utils [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2068.277080] env[62824]: DEBUG nova.compute.manager [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 15479062-af75-4925-99b3-77d6a49751ad] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2068.277080] env[62824]: DEBUG nova.network.neutron [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 15479062-af75-4925-99b3-77d6a49751ad] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2068.287099] env[62824]: DEBUG oslo_vmware.api [None req-e99f51a7-e71c-443f-812d-3cdaa01ba099 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Task: {'id': task-2146195, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2068.405868] env[62824]: DEBUG oslo_vmware.api [None req-c94e74ba-f9b2-469b-be3d-b77fc48e94ea tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146191, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2068.459842] env[62824]: DEBUG nova.policy [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0c461e862de44bb8b66545d6ec3b1b0c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd041a8209e3848eba8e47251e013ca17', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2068.470806] env[62824]: DEBUG oslo_vmware.api [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2146192, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.089633} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2068.471279] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2068.472463] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48145f29-a463-4e1e-8a59-e9defdce92eb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2068.511169] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Reconfiguring VM instance instance-00000039 to attach disk [datastore2] bbfcb2e3-9326-4548-b15b-e054cbfd192e/bbfcb2e3-9326-4548-b15b-e054cbfd192e.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2068.515357] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8a01e919-114f-4284-8da0-cd91baec325c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2068.555962] env[62824]: DEBUG oslo_vmware.api [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Task: {'id': task-2146193, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.521557} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2068.555962] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] ed7d3e95-f26e-40fd-a370-e17922bbff8e/ed7d3e95-f26e-40fd-a370-e17922bbff8e.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2068.555962] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] [instance: ed7d3e95-f26e-40fd-a370-e17922bbff8e] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2068.555962] env[62824]: DEBUG oslo_vmware.api [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for the task: (returnval){ [ 2068.555962] env[62824]: value = "task-2146196" [ 2068.555962] env[62824]: _type = "Task" [ 2068.555962] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2068.556303] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-30c4bdd2-8c6a-45cb-a011-d293fc3e68d2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2068.574784] env[62824]: DEBUG oslo_vmware.api [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2146196, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2068.577051] env[62824]: DEBUG oslo_vmware.api [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Waiting for the task: (returnval){ [ 2068.577051] env[62824]: value = "task-2146197" [ 2068.577051] env[62824]: _type = "Task" [ 2068.577051] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2068.591274] env[62824]: DEBUG oslo_vmware.api [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Task: {'id': task-2146197, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2068.611028] env[62824]: DEBUG oslo_vmware.api [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Task: {'id': task-2146194, 'name': RelocateVM_Task} progress is 35%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2068.705605] env[62824]: DEBUG nova.network.neutron [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Updating instance_info_cache with network_info: [{"id": "60a4fce0-0e63-45af-890c-46ca44ea9a0c", "address": "fa:16:3e:cf:18:eb", "network": {"id": "db233b0c-27f3-4a76-ae9c-2b5889bf3d3b", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1825671448-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.231", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3240be4f32254f54aa87cb045e3d8f31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7041d198-66a3-40de-bf7d-cfc036e6ed69", "external-id": "nsx-vlan-transportzone-278", "segmentation_id": 278, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap60a4fce0-0e", "ovs_interfaceid": "60a4fce0-0e63-45af-890c-46ca44ea9a0c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2068.781757] env[62824]: DEBUG nova.compute.utils [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2068.783353] env[62824]: DEBUG oslo_vmware.api [None req-e99f51a7-e71c-443f-812d-3cdaa01ba099 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Task: {'id': task-2146195, 'name': PowerOffVM_Task} progress is 100%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2068.871274] env[62824]: DEBUG nova.network.neutron [req-d9d0bdfd-3f51-44f3-bc13-375198d3d49d req-23492b4b-2ec3-4e43-9d25-d4b9e9f6dfdc service nova] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Updated VIF entry in instance network info cache for port d358ab3c-f4fb-41f3-954b-19de8b22ef15. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2068.871274] env[62824]: DEBUG nova.network.neutron [req-d9d0bdfd-3f51-44f3-bc13-375198d3d49d req-23492b4b-2ec3-4e43-9d25-d4b9e9f6dfdc service nova] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Updating instance_info_cache with network_info: [{"id": "d358ab3c-f4fb-41f3-954b-19de8b22ef15", "address": "fa:16:3e:12:72:e1", "network": {"id": "2ed8a86a-9c14-47b4-8cb4-e8496d289a6e", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1207171730-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "87239f55d6d44606b23ca0c25d9fb4d9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7514a465-f1a4-4a8b-b76b-726b1a9d7e2f", "external-id": "nsx-vlan-transportzone-36", "segmentation_id": 36, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd358ab3c-f4", "ovs_interfaceid": "d358ab3c-f4fb-41f3-954b-19de8b22ef15", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2068.887948] env[62824]: DEBUG oslo_concurrency.lockutils [None req-70aefc57-09da-40ff-bfce-730a89b771bc tempest-VolumesAssistedSnapshotsTest-163918615 tempest-VolumesAssistedSnapshotsTest-163918615-project-admin] Lock "29b8dd5f-1855-490c-a01b-54840073a753" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.449s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2068.914301] env[62824]: DEBUG oslo_vmware.api [None req-c94e74ba-f9b2-469b-be3d-b77fc48e94ea tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146191, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2069.043486] env[62824]: DEBUG nova.network.neutron [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 919e8a45-7810-4a8d-a1aa-5046b5ab059c] Successfully updated port: ec8bab9e-6a51-4cc3-9fe8-4764dfd97d45 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2069.082255] env[62824]: DEBUG oslo_vmware.api [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2146196, 'name': ReconfigVM_Task, 'duration_secs': 0.421939} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2069.086308] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Reconfigured VM instance instance-00000039 to attach disk [datastore2] bbfcb2e3-9326-4548-b15b-e054cbfd192e/bbfcb2e3-9326-4548-b15b-e054cbfd192e.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2069.089652] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-593ee09d-664a-451e-8310-5b4e50ee7b20 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2069.102898] env[62824]: DEBUG oslo_vmware.api [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Task: {'id': task-2146197, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.084051} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2069.102898] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] [instance: ed7d3e95-f26e-40fd-a370-e17922bbff8e] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2069.103527] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-935db92c-1798-4cf2-854c-7f8db0f8f834 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2069.108846] env[62824]: DEBUG oslo_vmware.api [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for the task: (returnval){ [ 2069.108846] env[62824]: value = "task-2146199" [ 2069.108846] env[62824]: _type = "Task" [ 2069.108846] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2069.116115] env[62824]: DEBUG oslo_vmware.api [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Task: {'id': task-2146194, 'name': RelocateVM_Task} progress is 51%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2069.138393] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] [instance: ed7d3e95-f26e-40fd-a370-e17922bbff8e] Reconfiguring VM instance instance-00000037 to attach disk [datastore2] ed7d3e95-f26e-40fd-a370-e17922bbff8e/ed7d3e95-f26e-40fd-a370-e17922bbff8e.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2069.141963] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0819731c-6af3-4f37-8136-fdae4d6c79a5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2069.160874] env[62824]: DEBUG oslo_vmware.api [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2146199, 'name': Rename_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2069.161650] env[62824]: DEBUG nova.network.neutron [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 15479062-af75-4925-99b3-77d6a49751ad] Successfully created port: 6b8b3a2d-f55e-4aa8-b995-4880c1adea60 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2069.172712] env[62824]: DEBUG oslo_vmware.api [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Waiting for the task: (returnval){ [ 2069.172712] env[62824]: value = "task-2146200" [ 2069.172712] env[62824]: _type = "Task" [ 2069.172712] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2069.188552] env[62824]: DEBUG oslo_vmware.api [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Task: {'id': task-2146200, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2069.213292] env[62824]: DEBUG oslo_concurrency.lockutils [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Releasing lock "refresh_cache-b8cc8cd3-ea03-40bf-b867-7ad193365552" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2069.291517] env[62824]: DEBUG nova.compute.manager [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 15479062-af75-4925-99b3-77d6a49751ad] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2069.294896] env[62824]: DEBUG oslo_vmware.api [None req-e99f51a7-e71c-443f-812d-3cdaa01ba099 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Task: {'id': task-2146195, 'name': PowerOffVM_Task} progress is 100%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2069.334748] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d515df3c-54b5-4dd7-8598-40172b6be750 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Acquiring lock "88bad2dd-dce7-41df-b56c-93a5d054c11e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2069.334895] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d515df3c-54b5-4dd7-8598-40172b6be750 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Lock "88bad2dd-dce7-41df-b56c-93a5d054c11e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2069.335011] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d515df3c-54b5-4dd7-8598-40172b6be750 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Acquiring lock "88bad2dd-dce7-41df-b56c-93a5d054c11e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2069.335223] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d515df3c-54b5-4dd7-8598-40172b6be750 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Lock "88bad2dd-dce7-41df-b56c-93a5d054c11e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2069.335401] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d515df3c-54b5-4dd7-8598-40172b6be750 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Lock "88bad2dd-dce7-41df-b56c-93a5d054c11e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2069.337905] env[62824]: INFO nova.compute.manager [None req-d515df3c-54b5-4dd7-8598-40172b6be750 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 88bad2dd-dce7-41df-b56c-93a5d054c11e] Terminating instance [ 2069.379280] env[62824]: DEBUG oslo_concurrency.lockutils [req-d9d0bdfd-3f51-44f3-bc13-375198d3d49d req-23492b4b-2ec3-4e43-9d25-d4b9e9f6dfdc service nova] Releasing lock "refresh_cache-2f315f9e-94e4-47ef-9503-ee92b59e5452" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2069.410189] env[62824]: DEBUG oslo_vmware.api [None req-c94e74ba-f9b2-469b-be3d-b77fc48e94ea tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146191, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2069.493339] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-03bb66d7-d558-4473-adb2-183a408a723b tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: b0456b8a-348a-4503-a92c-58e5ab455d1c] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2069.495384] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cd83f36e-8940-47e1-862b-17ae39cb8c7e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2069.508973] env[62824]: DEBUG oslo_vmware.api [None req-03bb66d7-d558-4473-adb2-183a408a723b tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Waiting for the task: (returnval){ [ 2069.508973] env[62824]: value = "task-2146201" [ 2069.508973] env[62824]: _type = "Task" [ 2069.508973] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2069.514200] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26c82096-7cda-41ea-8c4d-dede51d28eea {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2069.527531] env[62824]: DEBUG oslo_vmware.api [None req-03bb66d7-d558-4473-adb2-183a408a723b tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146201, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2069.534399] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b5b6668-76c9-4053-b0cb-96a61e0edb19 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2069.580455] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquiring lock "refresh_cache-919e8a45-7810-4a8d-a1aa-5046b5ab059c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2069.580455] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquired lock "refresh_cache-919e8a45-7810-4a8d-a1aa-5046b5ab059c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2069.580455] env[62824]: DEBUG nova.network.neutron [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 919e8a45-7810-4a8d-a1aa-5046b5ab059c] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2069.581915] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ec14248-bc19-40d2-8ba2-c3b68d4e99f4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2069.594436] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7226bf79-dbfe-47ee-9640-c3e8e99600f8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2069.620671] env[62824]: DEBUG nova.compute.provider_tree [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2069.629594] env[62824]: DEBUG oslo_vmware.api [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Task: {'id': task-2146194, 'name': RelocateVM_Task} progress is 63%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2069.639455] env[62824]: DEBUG oslo_vmware.api [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2146199, 'name': Rename_Task, 'duration_secs': 0.197328} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2069.639523] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2069.639764] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-53e75840-3ba2-48a8-8f19-ecd426580a2e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2069.652469] env[62824]: DEBUG oslo_vmware.api [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for the task: (returnval){ [ 2069.652469] env[62824]: value = "task-2146202" [ 2069.652469] env[62824]: _type = "Task" [ 2069.652469] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2069.667037] env[62824]: DEBUG oslo_vmware.api [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2146202, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2069.698047] env[62824]: DEBUG oslo_vmware.api [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Task: {'id': task-2146200, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2069.791576] env[62824]: DEBUG oslo_vmware.api [None req-e99f51a7-e71c-443f-812d-3cdaa01ba099 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Task: {'id': task-2146195, 'name': PowerOffVM_Task, 'duration_secs': 1.180782} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2069.791895] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-e99f51a7-e71c-443f-812d-3cdaa01ba099 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 21c05725-3dad-4965-98d0-0622ebcc8ece] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2069.792100] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e99f51a7-e71c-443f-812d-3cdaa01ba099 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 21c05725-3dad-4965-98d0-0622ebcc8ece] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2069.792516] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-748d6baa-6ed4-4d3c-a013-60b308676987 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2069.835660] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e99f51a7-e71c-443f-812d-3cdaa01ba099 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 21c05725-3dad-4965-98d0-0622ebcc8ece] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2069.835945] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e99f51a7-e71c-443f-812d-3cdaa01ba099 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 21c05725-3dad-4965-98d0-0622ebcc8ece] Deleting contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2069.836187] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-e99f51a7-e71c-443f-812d-3cdaa01ba099 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Deleting the datastore file [datastore1] 21c05725-3dad-4965-98d0-0622ebcc8ece {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2069.837034] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e49e9fca-409d-4008-9c01-d777eecd8226 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2069.842768] env[62824]: DEBUG nova.compute.manager [None req-d515df3c-54b5-4dd7-8598-40172b6be750 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 88bad2dd-dce7-41df-b56c-93a5d054c11e] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2069.842768] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-d515df3c-54b5-4dd7-8598-40172b6be750 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 88bad2dd-dce7-41df-b56c-93a5d054c11e] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2069.844491] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93915116-e65f-4de0-b817-562999fcc220 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2069.849991] env[62824]: DEBUG oslo_vmware.api [None req-e99f51a7-e71c-443f-812d-3cdaa01ba099 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Waiting for the task: (returnval){ [ 2069.849991] env[62824]: value = "task-2146204" [ 2069.849991] env[62824]: _type = "Task" [ 2069.849991] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2069.861235] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-d515df3c-54b5-4dd7-8598-40172b6be750 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 88bad2dd-dce7-41df-b56c-93a5d054c11e] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2069.865974] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f01411ee-f3dc-4613-8d5c-8f1488f04f5b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2069.868099] env[62824]: DEBUG oslo_vmware.api [None req-e99f51a7-e71c-443f-812d-3cdaa01ba099 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Task: {'id': task-2146204, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2069.915041] env[62824]: DEBUG oslo_vmware.api [None req-c94e74ba-f9b2-469b-be3d-b77fc48e94ea tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146191, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2070.023466] env[62824]: DEBUG oslo_vmware.api [None req-03bb66d7-d558-4473-adb2-183a408a723b tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146201, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2070.110608] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-d515df3c-54b5-4dd7-8598-40172b6be750 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 88bad2dd-dce7-41df-b56c-93a5d054c11e] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2070.114452] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-d515df3c-54b5-4dd7-8598-40172b6be750 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 88bad2dd-dce7-41df-b56c-93a5d054c11e] Deleting contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2070.115204] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-d515df3c-54b5-4dd7-8598-40172b6be750 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Deleting the datastore file [datastore1] 88bad2dd-dce7-41df-b56c-93a5d054c11e {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2070.116471] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ee6bfefb-633d-49c0-8e86-1ea9d31b11d4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2070.124115] env[62824]: DEBUG oslo_vmware.api [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Task: {'id': task-2146194, 'name': RelocateVM_Task} progress is 76%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2070.127025] env[62824]: DEBUG oslo_vmware.api [None req-d515df3c-54b5-4dd7-8598-40172b6be750 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the task: (returnval){ [ 2070.127025] env[62824]: value = "task-2146206" [ 2070.127025] env[62824]: _type = "Task" [ 2070.127025] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2070.131310] env[62824]: DEBUG nova.scheduler.client.report [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2070.146035] env[62824]: DEBUG oslo_vmware.api [None req-d515df3c-54b5-4dd7-8598-40172b6be750 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146206, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2070.169759] env[62824]: DEBUG oslo_vmware.api [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2146202, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2070.192200] env[62824]: DEBUG oslo_vmware.api [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Task: {'id': task-2146200, 'name': ReconfigVM_Task, 'duration_secs': 0.766783} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2070.192732] env[62824]: DEBUG nova.network.neutron [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 919e8a45-7810-4a8d-a1aa-5046b5ab059c] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2070.195025] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] [instance: ed7d3e95-f26e-40fd-a370-e17922bbff8e] Reconfigured VM instance instance-00000037 to attach disk [datastore2] ed7d3e95-f26e-40fd-a370-e17922bbff8e/ed7d3e95-f26e-40fd-a370-e17922bbff8e.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2070.196018] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-44b02d65-f50e-45ae-a817-009f4682bcdb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2070.209052] env[62824]: DEBUG oslo_vmware.api [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Waiting for the task: (returnval){ [ 2070.209052] env[62824]: value = "task-2146207" [ 2070.209052] env[62824]: _type = "Task" [ 2070.209052] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2070.223906] env[62824]: DEBUG oslo_vmware.api [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Task: {'id': task-2146207, 'name': Rename_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2070.250397] env[62824]: DEBUG nova.compute.manager [req-f58671c5-455f-4331-ab01-d66ec427f700 req-ccd2002d-ddad-481f-a4ab-95a827fecf47 service nova] [instance: 919e8a45-7810-4a8d-a1aa-5046b5ab059c] Received event network-vif-plugged-ec8bab9e-6a51-4cc3-9fe8-4764dfd97d45 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2070.250397] env[62824]: DEBUG oslo_concurrency.lockutils [req-f58671c5-455f-4331-ab01-d66ec427f700 req-ccd2002d-ddad-481f-a4ab-95a827fecf47 service nova] Acquiring lock "919e8a45-7810-4a8d-a1aa-5046b5ab059c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2070.250635] env[62824]: DEBUG oslo_concurrency.lockutils [req-f58671c5-455f-4331-ab01-d66ec427f700 req-ccd2002d-ddad-481f-a4ab-95a827fecf47 service nova] Lock "919e8a45-7810-4a8d-a1aa-5046b5ab059c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2070.250792] env[62824]: DEBUG oslo_concurrency.lockutils [req-f58671c5-455f-4331-ab01-d66ec427f700 req-ccd2002d-ddad-481f-a4ab-95a827fecf47 service nova] Lock "919e8a45-7810-4a8d-a1aa-5046b5ab059c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2070.251345] env[62824]: DEBUG nova.compute.manager [req-f58671c5-455f-4331-ab01-d66ec427f700 req-ccd2002d-ddad-481f-a4ab-95a827fecf47 service nova] [instance: 919e8a45-7810-4a8d-a1aa-5046b5ab059c] No waiting events found dispatching network-vif-plugged-ec8bab9e-6a51-4cc3-9fe8-4764dfd97d45 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2070.251345] env[62824]: WARNING nova.compute.manager [req-f58671c5-455f-4331-ab01-d66ec427f700 req-ccd2002d-ddad-481f-a4ab-95a827fecf47 service nova] [instance: 919e8a45-7810-4a8d-a1aa-5046b5ab059c] Received unexpected event network-vif-plugged-ec8bab9e-6a51-4cc3-9fe8-4764dfd97d45 for instance with vm_state building and task_state spawning. [ 2070.251492] env[62824]: DEBUG nova.compute.manager [req-f58671c5-455f-4331-ab01-d66ec427f700 req-ccd2002d-ddad-481f-a4ab-95a827fecf47 service nova] [instance: 919e8a45-7810-4a8d-a1aa-5046b5ab059c] Received event network-changed-ec8bab9e-6a51-4cc3-9fe8-4764dfd97d45 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2070.251695] env[62824]: DEBUG nova.compute.manager [req-f58671c5-455f-4331-ab01-d66ec427f700 req-ccd2002d-ddad-481f-a4ab-95a827fecf47 service nova] [instance: 919e8a45-7810-4a8d-a1aa-5046b5ab059c] Refreshing instance network info cache due to event network-changed-ec8bab9e-6a51-4cc3-9fe8-4764dfd97d45. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2070.251907] env[62824]: DEBUG oslo_concurrency.lockutils [req-f58671c5-455f-4331-ab01-d66ec427f700 req-ccd2002d-ddad-481f-a4ab-95a827fecf47 service nova] Acquiring lock "refresh_cache-919e8a45-7810-4a8d-a1aa-5046b5ab059c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2070.303263] env[62824]: DEBUG nova.compute.manager [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 15479062-af75-4925-99b3-77d6a49751ad] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2070.356057] env[62824]: DEBUG nova.virt.hardware [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:50:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=1,extra_specs={hw_rng:allowed='True'},flavorid='315095566',id=20,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_1-21118905',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2070.356057] env[62824]: DEBUG nova.virt.hardware [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2070.356057] env[62824]: DEBUG nova.virt.hardware [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2070.356057] env[62824]: DEBUG nova.virt.hardware [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2070.356057] env[62824]: DEBUG nova.virt.hardware [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2070.356057] env[62824]: DEBUG nova.virt.hardware [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2070.356057] env[62824]: DEBUG nova.virt.hardware [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2070.356436] env[62824]: DEBUG nova.virt.hardware [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2070.357587] env[62824]: DEBUG nova.virt.hardware [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2070.357587] env[62824]: DEBUG nova.virt.hardware [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2070.357587] env[62824]: DEBUG nova.virt.hardware [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2070.358791] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f05f369b-96de-44ea-857f-aed5ebc81b31 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2070.379510] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9f57daf-02f2-4b66-be21-e46ecddf4523 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2070.385276] env[62824]: DEBUG oslo_vmware.api [None req-e99f51a7-e71c-443f-812d-3cdaa01ba099 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Task: {'id': task-2146204, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.232564} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2070.387761] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-e99f51a7-e71c-443f-812d-3cdaa01ba099 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2070.388053] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e99f51a7-e71c-443f-812d-3cdaa01ba099 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 21c05725-3dad-4965-98d0-0622ebcc8ece] Deleted contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2070.388217] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e99f51a7-e71c-443f-812d-3cdaa01ba099 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 21c05725-3dad-4965-98d0-0622ebcc8ece] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2070.388388] env[62824]: INFO nova.compute.manager [None req-e99f51a7-e71c-443f-812d-3cdaa01ba099 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 21c05725-3dad-4965-98d0-0622ebcc8ece] Took 2.14 seconds to destroy the instance on the hypervisor. [ 2070.388718] env[62824]: DEBUG oslo.service.loopingcall [None req-e99f51a7-e71c-443f-812d-3cdaa01ba099 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2070.389477] env[62824]: DEBUG nova.compute.manager [-] [instance: 21c05725-3dad-4965-98d0-0622ebcc8ece] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2070.389651] env[62824]: DEBUG nova.network.neutron [-] [instance: 21c05725-3dad-4965-98d0-0622ebcc8ece] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2070.422071] env[62824]: DEBUG oslo_vmware.api [None req-c94e74ba-f9b2-469b-be3d-b77fc48e94ea tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146191, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2070.429404] env[62824]: DEBUG nova.network.neutron [-] [instance: 21c05725-3dad-4965-98d0-0622ebcc8ece] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2070.527042] env[62824]: DEBUG oslo_vmware.api [None req-03bb66d7-d558-4473-adb2-183a408a723b tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146201, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2070.620167] env[62824]: DEBUG oslo_vmware.api [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Task: {'id': task-2146194, 'name': RelocateVM_Task} progress is 89%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2070.641132] env[62824]: DEBUG oslo_concurrency.lockutils [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.875s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2070.641679] env[62824]: DEBUG nova.compute.manager [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: a6cd2032-de60-4f78-bf1e-79801d049df0] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2070.644432] env[62824]: DEBUG oslo_vmware.api [None req-d515df3c-54b5-4dd7-8598-40172b6be750 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146206, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.244302} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2070.645032] env[62824]: DEBUG oslo_concurrency.lockutils [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.702s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2070.646517] env[62824]: INFO nova.compute.claims [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02f34565-6675-4c79-ac47-b131ceba9df8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2070.650939] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-d515df3c-54b5-4dd7-8598-40172b6be750 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2070.650939] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-d515df3c-54b5-4dd7-8598-40172b6be750 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 88bad2dd-dce7-41df-b56c-93a5d054c11e] Deleted contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2070.650939] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-d515df3c-54b5-4dd7-8598-40172b6be750 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 88bad2dd-dce7-41df-b56c-93a5d054c11e] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2070.650939] env[62824]: INFO nova.compute.manager [None req-d515df3c-54b5-4dd7-8598-40172b6be750 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 88bad2dd-dce7-41df-b56c-93a5d054c11e] Took 0.81 seconds to destroy the instance on the hypervisor. [ 2070.650939] env[62824]: DEBUG oslo.service.loopingcall [None req-d515df3c-54b5-4dd7-8598-40172b6be750 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2070.650939] env[62824]: DEBUG nova.compute.manager [-] [instance: 88bad2dd-dce7-41df-b56c-93a5d054c11e] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2070.650939] env[62824]: DEBUG nova.network.neutron [-] [instance: 88bad2dd-dce7-41df-b56c-93a5d054c11e] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2070.666031] env[62824]: DEBUG oslo_vmware.api [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2146202, 'name': PowerOnVM_Task, 'duration_secs': 0.585242} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2070.666031] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2070.666253] env[62824]: INFO nova.compute.manager [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Took 10.77 seconds to spawn the instance on the hypervisor. [ 2070.666292] env[62824]: DEBUG nova.compute.manager [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2070.667477] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c355121-6a2b-4f0e-8dd5-e330553072c9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2070.725016] env[62824]: DEBUG oslo_vmware.api [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Task: {'id': task-2146207, 'name': Rename_Task, 'duration_secs': 0.298056} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2070.727182] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] [instance: ed7d3e95-f26e-40fd-a370-e17922bbff8e] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2070.727463] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-97f5a7a3-34af-4c19-9d51-38b5cbbc1b25 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2070.737449] env[62824]: DEBUG oslo_vmware.api [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Waiting for the task: (returnval){ [ 2070.737449] env[62824]: value = "task-2146208" [ 2070.737449] env[62824]: _type = "Task" [ 2070.737449] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2070.748208] env[62824]: DEBUG oslo_vmware.api [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Task: {'id': task-2146208, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2070.799749] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2070.800676] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-819fde45-4e98-4ac2-bf00-84e1018bf10d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2070.810183] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2070.810517] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-545800ff-def2-4f05-969f-561693becc94 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2070.831565] env[62824]: DEBUG nova.network.neutron [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 919e8a45-7810-4a8d-a1aa-5046b5ab059c] Updating instance_info_cache with network_info: [{"id": "ec8bab9e-6a51-4cc3-9fe8-4764dfd97d45", "address": "fa:16:3e:bf:ed:44", "network": {"id": "b995934c-4934-4469-884c-73cb4de1796f", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1892307918-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c11aef7fb6247cb9b5272a6a063cd12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15165046-2de9-4ada-9e99-0126e20854a9", "external-id": "nsx-vlan-transportzone-974", "segmentation_id": 974, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapec8bab9e-6a", "ovs_interfaceid": "ec8bab9e-6a51-4cc3-9fe8-4764dfd97d45", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2070.855745] env[62824]: DEBUG oslo_concurrency.lockutils [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Acquiring lock "1870b82a-9783-44ac-8de2-7b9ffc2a1bc8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2070.855993] env[62824]: DEBUG oslo_concurrency.lockutils [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Lock "1870b82a-9783-44ac-8de2-7b9ffc2a1bc8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2070.917113] env[62824]: DEBUG oslo_vmware.api [None req-c94e74ba-f9b2-469b-be3d-b77fc48e94ea tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146191, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2070.934645] env[62824]: DEBUG nova.network.neutron [-] [instance: 21c05725-3dad-4965-98d0-0622ebcc8ece] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2070.959324] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2070.959324] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Deleting contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2070.959530] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Deleting the datastore file [datastore1] b8cc8cd3-ea03-40bf-b867-7ad193365552 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2070.959882] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f251fb37-63a0-40dd-a71f-b61f2875e04c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2070.969903] env[62824]: DEBUG oslo_vmware.api [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Waiting for the task: (returnval){ [ 2070.969903] env[62824]: value = "task-2146210" [ 2070.969903] env[62824]: _type = "Task" [ 2070.969903] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2070.989249] env[62824]: DEBUG oslo_vmware.api [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2146210, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2071.024724] env[62824]: DEBUG oslo_vmware.api [None req-03bb66d7-d558-4473-adb2-183a408a723b tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146201, 'name': PowerOffVM_Task} progress is 100%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2071.115812] env[62824]: DEBUG oslo_vmware.api [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Task: {'id': task-2146194, 'name': RelocateVM_Task} progress is 97%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2071.146705] env[62824]: DEBUG nova.compute.utils [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2071.148160] env[62824]: DEBUG nova.compute.manager [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: a6cd2032-de60-4f78-bf1e-79801d049df0] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2071.148335] env[62824]: DEBUG nova.network.neutron [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: a6cd2032-de60-4f78-bf1e-79801d049df0] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2071.190901] env[62824]: INFO nova.compute.manager [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Took 38.88 seconds to build instance. [ 2071.220774] env[62824]: DEBUG nova.policy [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '959d127a7d144b33a0cae94db5c11846', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dfff08982dad4790bf4d555e2b4db5e4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2071.256045] env[62824]: DEBUG oslo_vmware.api [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Task: {'id': task-2146208, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2071.334987] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Releasing lock "refresh_cache-919e8a45-7810-4a8d-a1aa-5046b5ab059c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2071.335428] env[62824]: DEBUG nova.compute.manager [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 919e8a45-7810-4a8d-a1aa-5046b5ab059c] Instance network_info: |[{"id": "ec8bab9e-6a51-4cc3-9fe8-4764dfd97d45", "address": "fa:16:3e:bf:ed:44", "network": {"id": "b995934c-4934-4469-884c-73cb4de1796f", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1892307918-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c11aef7fb6247cb9b5272a6a063cd12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15165046-2de9-4ada-9e99-0126e20854a9", "external-id": "nsx-vlan-transportzone-974", "segmentation_id": 974, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapec8bab9e-6a", "ovs_interfaceid": "ec8bab9e-6a51-4cc3-9fe8-4764dfd97d45", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2071.335976] env[62824]: DEBUG oslo_concurrency.lockutils [req-f58671c5-455f-4331-ab01-d66ec427f700 req-ccd2002d-ddad-481f-a4ab-95a827fecf47 service nova] Acquired lock "refresh_cache-919e8a45-7810-4a8d-a1aa-5046b5ab059c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2071.335976] env[62824]: DEBUG nova.network.neutron [req-f58671c5-455f-4331-ab01-d66ec427f700 req-ccd2002d-ddad-481f-a4ab-95a827fecf47 service nova] [instance: 919e8a45-7810-4a8d-a1aa-5046b5ab059c] Refreshing network info cache for port ec8bab9e-6a51-4cc3-9fe8-4764dfd97d45 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2071.337229] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 919e8a45-7810-4a8d-a1aa-5046b5ab059c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:bf:ed:44', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '15165046-2de9-4ada-9e99-0126e20854a9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ec8bab9e-6a51-4cc3-9fe8-4764dfd97d45', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2071.345956] env[62824]: DEBUG oslo.service.loopingcall [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2071.348301] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 919e8a45-7810-4a8d-a1aa-5046b5ab059c] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2071.348982] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ecaf35ec-fe98-4e0b-8662-825fc0655f64 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2071.369057] env[62824]: DEBUG nova.compute.manager [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2071.378044] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2071.378044] env[62824]: value = "task-2146211" [ 2071.378044] env[62824]: _type = "Task" [ 2071.378044] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2071.388820] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146211, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2071.419015] env[62824]: DEBUG oslo_vmware.api [None req-c94e74ba-f9b2-469b-be3d-b77fc48e94ea tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146191, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2071.437253] env[62824]: INFO nova.compute.manager [-] [instance: 21c05725-3dad-4965-98d0-0622ebcc8ece] Took 1.05 seconds to deallocate network for instance. [ 2071.483376] env[62824]: DEBUG oslo_vmware.api [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2146210, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.132868} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2071.483657] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2071.483807] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Deleted contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2071.483983] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2071.539614] env[62824]: DEBUG oslo_vmware.api [None req-03bb66d7-d558-4473-adb2-183a408a723b tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146201, 'name': PowerOffVM_Task} progress is 100%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2071.620697] env[62824]: DEBUG oslo_vmware.api [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Task: {'id': task-2146194, 'name': RelocateVM_Task} progress is 97%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2071.653159] env[62824]: DEBUG nova.compute.manager [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: a6cd2032-de60-4f78-bf1e-79801d049df0] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2071.665604] env[62824]: DEBUG nova.network.neutron [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 15479062-af75-4925-99b3-77d6a49751ad] Successfully updated port: 6b8b3a2d-f55e-4aa8-b995-4880c1adea60 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2071.692658] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c29ad508-d524-49e1-9140-bef9f6c77bad tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Lock "bbfcb2e3-9326-4548-b15b-e054cbfd192e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 80.503s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2071.698388] env[62824]: INFO nova.scheduler.client.report [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Deleted allocations for instance b8cc8cd3-ea03-40bf-b867-7ad193365552 [ 2071.749099] env[62824]: DEBUG nova.network.neutron [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: a6cd2032-de60-4f78-bf1e-79801d049df0] Successfully created port: 018929f6-4f6c-4148-b1eb-9cf11f5cc746 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2071.763634] env[62824]: DEBUG oslo_vmware.api [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Task: {'id': task-2146208, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2071.878677] env[62824]: DEBUG nova.network.neutron [req-f58671c5-455f-4331-ab01-d66ec427f700 req-ccd2002d-ddad-481f-a4ab-95a827fecf47 service nova] [instance: 919e8a45-7810-4a8d-a1aa-5046b5ab059c] Updated VIF entry in instance network info cache for port ec8bab9e-6a51-4cc3-9fe8-4764dfd97d45. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2071.879329] env[62824]: DEBUG nova.network.neutron [req-f58671c5-455f-4331-ab01-d66ec427f700 req-ccd2002d-ddad-481f-a4ab-95a827fecf47 service nova] [instance: 919e8a45-7810-4a8d-a1aa-5046b5ab059c] Updating instance_info_cache with network_info: [{"id": "ec8bab9e-6a51-4cc3-9fe8-4764dfd97d45", "address": "fa:16:3e:bf:ed:44", "network": {"id": "b995934c-4934-4469-884c-73cb4de1796f", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1892307918-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c11aef7fb6247cb9b5272a6a063cd12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15165046-2de9-4ada-9e99-0126e20854a9", "external-id": "nsx-vlan-transportzone-974", "segmentation_id": 974, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapec8bab9e-6a", "ovs_interfaceid": "ec8bab9e-6a51-4cc3-9fe8-4764dfd97d45", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2071.892630] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146211, 'name': CreateVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2071.894090] env[62824]: DEBUG oslo_concurrency.lockutils [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2071.924648] env[62824]: DEBUG nova.network.neutron [-] [instance: 88bad2dd-dce7-41df-b56c-93a5d054c11e] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2071.925811] env[62824]: DEBUG oslo_vmware.api [None req-c94e74ba-f9b2-469b-be3d-b77fc48e94ea tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146191, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2071.943575] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e99f51a7-e71c-443f-812d-3cdaa01ba099 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2072.030431] env[62824]: DEBUG oslo_vmware.api [None req-03bb66d7-d558-4473-adb2-183a408a723b tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146201, 'name': PowerOffVM_Task, 'duration_secs': 2.091191} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2072.033554] env[62824]: DEBUG oslo_concurrency.lockutils [None req-4715e6dc-1d1a-49a2-9d97-4f4b36f634e1 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Acquiring lock "387b7f14-f3c1-43ed-8c4e-6de3ce9822e2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2072.033691] env[62824]: DEBUG oslo_concurrency.lockutils [None req-4715e6dc-1d1a-49a2-9d97-4f4b36f634e1 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Lock "387b7f14-f3c1-43ed-8c4e-6de3ce9822e2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2072.033920] env[62824]: DEBUG oslo_concurrency.lockutils [None req-4715e6dc-1d1a-49a2-9d97-4f4b36f634e1 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Acquiring lock "387b7f14-f3c1-43ed-8c4e-6de3ce9822e2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2072.034457] env[62824]: DEBUG oslo_concurrency.lockutils [None req-4715e6dc-1d1a-49a2-9d97-4f4b36f634e1 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Lock "387b7f14-f3c1-43ed-8c4e-6de3ce9822e2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2072.034661] env[62824]: DEBUG oslo_concurrency.lockutils [None req-4715e6dc-1d1a-49a2-9d97-4f4b36f634e1 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Lock "387b7f14-f3c1-43ed-8c4e-6de3ce9822e2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2072.036355] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-03bb66d7-d558-4473-adb2-183a408a723b tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: b0456b8a-348a-4503-a92c-58e5ab455d1c] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2072.037908] env[62824]: INFO nova.compute.manager [None req-4715e6dc-1d1a-49a2-9d97-4f4b36f634e1 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2] Terminating instance [ 2072.039293] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d2c778c-322b-4fa9-a7c5-7094a859e05f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2072.065512] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6afdc2b-7691-4e53-922c-160de5f25727 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2072.107138] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-03bb66d7-d558-4473-adb2-183a408a723b tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: b0456b8a-348a-4503-a92c-58e5ab455d1c] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2072.107801] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0bc0b497-971a-4a4c-8a0f-2f327cf979c9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2072.124454] env[62824]: DEBUG oslo_vmware.api [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Task: {'id': task-2146194, 'name': RelocateVM_Task} progress is 98%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2072.126141] env[62824]: DEBUG oslo_vmware.api [None req-03bb66d7-d558-4473-adb2-183a408a723b tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Waiting for the task: (returnval){ [ 2072.126141] env[62824]: value = "task-2146212" [ 2072.126141] env[62824]: _type = "Task" [ 2072.126141] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2072.140417] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-03bb66d7-d558-4473-adb2-183a408a723b tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: b0456b8a-348a-4503-a92c-58e5ab455d1c] VM already powered off {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 2072.140516] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-03bb66d7-d558-4473-adb2-183a408a723b tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: b0456b8a-348a-4503-a92c-58e5ab455d1c] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2072.141014] env[62824]: DEBUG oslo_concurrency.lockutils [None req-03bb66d7-d558-4473-adb2-183a408a723b tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2072.141178] env[62824]: DEBUG oslo_concurrency.lockutils [None req-03bb66d7-d558-4473-adb2-183a408a723b tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2072.141362] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-03bb66d7-d558-4473-adb2-183a408a723b tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2072.142998] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8268a539-cb73-4c5a-a64b-d5638af57f06 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2072.152273] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-03bb66d7-d558-4473-adb2-183a408a723b tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2072.152792] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-03bb66d7-d558-4473-adb2-183a408a723b tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2072.153296] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4f443134-6bd9-47d4-91db-53c6600582ca {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2072.159977] env[62824]: DEBUG oslo_vmware.api [None req-03bb66d7-d558-4473-adb2-183a408a723b tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Waiting for the task: (returnval){ [ 2072.159977] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52ffd791-4c55-0966-13f2-2316f8d39111" [ 2072.159977] env[62824]: _type = "Task" [ 2072.159977] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2072.172689] env[62824]: DEBUG oslo_vmware.api [None req-03bb66d7-d558-4473-adb2-183a408a723b tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52ffd791-4c55-0966-13f2-2316f8d39111, 'name': SearchDatastore_Task, 'duration_secs': 0.010001} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2072.176121] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b56b3d5b-6c27-4f04-bad5-526d18649948 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2072.180302] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Acquiring lock "refresh_cache-15479062-af75-4925-99b3-77d6a49751ad" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2072.180495] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Acquired lock "refresh_cache-15479062-af75-4925-99b3-77d6a49751ad" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2072.180703] env[62824]: DEBUG nova.network.neutron [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 15479062-af75-4925-99b3-77d6a49751ad] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2072.187116] env[62824]: DEBUG oslo_vmware.api [None req-03bb66d7-d558-4473-adb2-183a408a723b tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Waiting for the task: (returnval){ [ 2072.187116] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52794953-ce43-020c-9d66-ec9211f4ed51" [ 2072.187116] env[62824]: _type = "Task" [ 2072.187116] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2072.200774] env[62824]: DEBUG oslo_vmware.api [None req-03bb66d7-d558-4473-adb2-183a408a723b tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52794953-ce43-020c-9d66-ec9211f4ed51, 'name': SearchDatastore_Task, 'duration_secs': 0.010622} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2072.200774] env[62824]: DEBUG oslo_concurrency.lockutils [None req-03bb66d7-d558-4473-adb2-183a408a723b tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2072.200774] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-03bb66d7-d558-4473-adb2-183a408a723b tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] b0456b8a-348a-4503-a92c-58e5ab455d1c/9e2a7d30-212d-4ab8-9606-c5c6d52629e8-rescue.vmdk. {{(pid=62824) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 2072.200774] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-132abadb-d05c-4e64-9573-21609cb2b931 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2072.208758] env[62824]: DEBUG oslo_concurrency.lockutils [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2072.208758] env[62824]: DEBUG oslo_vmware.api [None req-03bb66d7-d558-4473-adb2-183a408a723b tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Waiting for the task: (returnval){ [ 2072.208758] env[62824]: value = "task-2146213" [ 2072.208758] env[62824]: _type = "Task" [ 2072.208758] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2072.217281] env[62824]: DEBUG oslo_vmware.api [None req-03bb66d7-d558-4473-adb2-183a408a723b tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146213, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2072.251147] env[62824]: DEBUG oslo_vmware.api [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Task: {'id': task-2146208, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2072.300986] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bb9cbe6-48bb-4db1-8d05-f005e9f43003 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2072.309421] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2621d60-85f7-4a02-8406-a395209951ba {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2072.346340] env[62824]: DEBUG oslo_concurrency.lockutils [None req-423ea96c-8206-47c1-8b7e-a945775631af tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Acquiring lock "29b8dd5f-1855-490c-a01b-54840073a753" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2072.346555] env[62824]: DEBUG oslo_concurrency.lockutils [None req-423ea96c-8206-47c1-8b7e-a945775631af tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Lock "29b8dd5f-1855-490c-a01b-54840073a753" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2072.346755] env[62824]: DEBUG oslo_concurrency.lockutils [None req-423ea96c-8206-47c1-8b7e-a945775631af tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Acquiring lock "29b8dd5f-1855-490c-a01b-54840073a753-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2072.346910] env[62824]: DEBUG oslo_concurrency.lockutils [None req-423ea96c-8206-47c1-8b7e-a945775631af tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Lock "29b8dd5f-1855-490c-a01b-54840073a753-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2072.347117] env[62824]: DEBUG oslo_concurrency.lockutils [None req-423ea96c-8206-47c1-8b7e-a945775631af tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Lock "29b8dd5f-1855-490c-a01b-54840073a753-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2072.349891] env[62824]: INFO nova.compute.manager [None req-423ea96c-8206-47c1-8b7e-a945775631af tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Terminating instance [ 2072.351914] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-565014b8-2b52-436e-9738-295610dd7595 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2072.357102] env[62824]: DEBUG nova.compute.manager [req-d79fa943-8b7e-4aa6-a137-72af43c818de req-91874a69-4e1c-4335-a595-5bf1ad573753 service nova] [instance: 88bad2dd-dce7-41df-b56c-93a5d054c11e] Received event network-vif-deleted-aaeb6e9d-ce3d-4e4c-b2a1-319b383fdf1d {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2072.367294] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c05566c7-6be0-4e5b-93c2-223a7f874c8b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2072.383502] env[62824]: DEBUG nova.compute.provider_tree [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2072.385398] env[62824]: DEBUG oslo_concurrency.lockutils [req-f58671c5-455f-4331-ab01-d66ec427f700 req-ccd2002d-ddad-481f-a4ab-95a827fecf47 service nova] Releasing lock "refresh_cache-919e8a45-7810-4a8d-a1aa-5046b5ab059c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2072.395547] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146211, 'name': CreateVM_Task, 'duration_secs': 0.623377} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2072.395547] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 919e8a45-7810-4a8d-a1aa-5046b5ab059c] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2072.396363] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2072.396533] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2072.396932] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2072.397205] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-68ef537e-e89a-47b5-b9d6-aaf8b1a049f4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2072.403063] env[62824]: DEBUG oslo_vmware.api [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 2072.403063] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52c85c42-4acd-fc6c-f8ae-a32ef90d4a41" [ 2072.403063] env[62824]: _type = "Task" [ 2072.403063] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2072.418096] env[62824]: DEBUG oslo_vmware.api [None req-c94e74ba-f9b2-469b-be3d-b77fc48e94ea tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146191, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2072.421419] env[62824]: DEBUG oslo_vmware.api [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52c85c42-4acd-fc6c-f8ae-a32ef90d4a41, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2072.431961] env[62824]: INFO nova.compute.manager [-] [instance: 88bad2dd-dce7-41df-b56c-93a5d054c11e] Took 1.78 seconds to deallocate network for instance. [ 2072.546134] env[62824]: DEBUG nova.compute.manager [None req-4715e6dc-1d1a-49a2-9d97-4f4b36f634e1 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2072.546404] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-4715e6dc-1d1a-49a2-9d97-4f4b36f634e1 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2072.547346] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-885a79cd-7249-42d3-8773-80e55dcc072e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2072.560758] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-4715e6dc-1d1a-49a2-9d97-4f4b36f634e1 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2072.561677] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4b24a645-7471-4185-abdd-18f531b077e5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2072.573032] env[62824]: DEBUG oslo_vmware.api [None req-4715e6dc-1d1a-49a2-9d97-4f4b36f634e1 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Waiting for the task: (returnval){ [ 2072.573032] env[62824]: value = "task-2146214" [ 2072.573032] env[62824]: _type = "Task" [ 2072.573032] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2072.584947] env[62824]: DEBUG oslo_vmware.api [None req-4715e6dc-1d1a-49a2-9d97-4f4b36f634e1 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2146214, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2072.618819] env[62824]: DEBUG oslo_vmware.api [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Task: {'id': task-2146194, 'name': RelocateVM_Task, 'duration_secs': 4.291928} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2072.619339] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Volume attach. Driver type: vmdk {{(pid=62824) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 2072.619433] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-438591', 'volume_id': '8d462122-e7ae-4257-8ce4-0a79a4acae13', 'name': 'volume-8d462122-e7ae-4257-8ce4-0a79a4acae13', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '2f315f9e-94e4-47ef-9503-ee92b59e5452', 'attached_at': '', 'detached_at': '', 'volume_id': '8d462122-e7ae-4257-8ce4-0a79a4acae13', 'serial': '8d462122-e7ae-4257-8ce4-0a79a4acae13'} {{(pid=62824) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 2072.620364] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-158cc3d7-f92c-469f-a409-a01ceb3ca511 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2072.643122] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdc6ffb3-9cb6-4530-9369-7e64d0f19840 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2072.668931] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Reconfiguring VM instance instance-0000003a to attach disk [datastore2] volume-8d462122-e7ae-4257-8ce4-0a79a4acae13/volume-8d462122-e7ae-4257-8ce4-0a79a4acae13.vmdk or device None with type thin {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2072.668931] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7813d090-3e96-4e53-8727-a8a7920935a0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2072.692423] env[62824]: DEBUG nova.compute.manager [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: a6cd2032-de60-4f78-bf1e-79801d049df0] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2072.696240] env[62824]: DEBUG oslo_vmware.api [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Waiting for the task: (returnval){ [ 2072.696240] env[62824]: value = "task-2146215" [ 2072.696240] env[62824]: _type = "Task" [ 2072.696240] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2072.705256] env[62824]: DEBUG oslo_vmware.api [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Task: {'id': task-2146215, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2072.722519] env[62824]: DEBUG oslo_vmware.api [None req-03bb66d7-d558-4473-adb2-183a408a723b tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146213, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.473951} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2072.722519] env[62824]: INFO nova.virt.vmwareapi.ds_util [None req-03bb66d7-d558-4473-adb2-183a408a723b tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] b0456b8a-348a-4503-a92c-58e5ab455d1c/9e2a7d30-212d-4ab8-9606-c5c6d52629e8-rescue.vmdk. [ 2072.722519] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f6092d2-be5d-41b3-8d7e-f458d41b7809 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2072.727304] env[62824]: DEBUG nova.virt.hardware [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2072.727505] env[62824]: DEBUG nova.virt.hardware [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2072.727980] env[62824]: DEBUG nova.virt.hardware [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2072.727980] env[62824]: DEBUG nova.virt.hardware [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2072.727980] env[62824]: DEBUG nova.virt.hardware [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2072.728171] env[62824]: DEBUG nova.virt.hardware [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2072.728347] env[62824]: DEBUG nova.virt.hardware [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2072.728475] env[62824]: DEBUG nova.virt.hardware [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2072.728631] env[62824]: DEBUG nova.virt.hardware [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2072.729063] env[62824]: DEBUG nova.virt.hardware [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2072.729063] env[62824]: DEBUG nova.virt.hardware [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2072.729934] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66342c2a-1ea8-46d0-9fc2-fe1bdaba2382 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2072.754618] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c34d871-0ea4-47a6-8db3-2c36d8e7b947 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2072.773170] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-03bb66d7-d558-4473-adb2-183a408a723b tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: b0456b8a-348a-4503-a92c-58e5ab455d1c] Reconfiguring VM instance instance-00000038 to attach disk [datastore1] b0456b8a-348a-4503-a92c-58e5ab455d1c/9e2a7d30-212d-4ab8-9606-c5c6d52629e8-rescue.vmdk or device None with type thin {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2072.773170] env[62824]: DEBUG nova.network.neutron [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 15479062-af75-4925-99b3-77d6a49751ad] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2072.777129] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eee2ab79-1fc7-4e8a-805e-f3789d144259 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2072.800606] env[62824]: DEBUG oslo_vmware.api [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Task: {'id': task-2146208, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2072.811414] env[62824]: DEBUG oslo_vmware.api [None req-03bb66d7-d558-4473-adb2-183a408a723b tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Waiting for the task: (returnval){ [ 2072.811414] env[62824]: value = "task-2146216" [ 2072.811414] env[62824]: _type = "Task" [ 2072.811414] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2072.827159] env[62824]: DEBUG oslo_vmware.api [None req-03bb66d7-d558-4473-adb2-183a408a723b tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146216, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2072.861403] env[62824]: DEBUG nova.compute.manager [None req-423ea96c-8206-47c1-8b7e-a945775631af tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2072.861494] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-423ea96c-8206-47c1-8b7e-a945775631af tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2072.864956] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e878b2b-9f7e-4044-9a44-13ac10d51ae5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2072.874248] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-423ea96c-8206-47c1-8b7e-a945775631af tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2072.874544] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-029edf5d-acc3-4dce-b999-90ebd5a990b0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2072.885030] env[62824]: DEBUG oslo_vmware.api [None req-423ea96c-8206-47c1-8b7e-a945775631af tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Waiting for the task: (returnval){ [ 2072.885030] env[62824]: value = "task-2146217" [ 2072.885030] env[62824]: _type = "Task" [ 2072.885030] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2072.893881] env[62824]: DEBUG nova.scheduler.client.report [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2072.897903] env[62824]: DEBUG oslo_vmware.api [None req-423ea96c-8206-47c1-8b7e-a945775631af tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Task: {'id': task-2146217, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2072.919440] env[62824]: DEBUG oslo_vmware.api [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52c85c42-4acd-fc6c-f8ae-a32ef90d4a41, 'name': SearchDatastore_Task, 'duration_secs': 0.024633} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2072.920761] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2072.924201] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 919e8a45-7810-4a8d-a1aa-5046b5ab059c] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2072.924201] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2072.924201] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2072.924201] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2072.924201] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2fa3ceeb-20d1-41eb-8c94-c14bf9b89e9e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2072.931228] env[62824]: DEBUG oslo_vmware.api [None req-c94e74ba-f9b2-469b-be3d-b77fc48e94ea tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146191, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2072.940427] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d515df3c-54b5-4dd7-8598-40172b6be750 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2072.943214] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2072.943444] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2072.944326] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3fe9581f-1c11-4b97-b442-83b91cdc9ae3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2072.951352] env[62824]: DEBUG oslo_vmware.api [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 2072.951352] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52b2cd5d-7a48-5783-202e-1cacd02cc940" [ 2072.951352] env[62824]: _type = "Task" [ 2072.951352] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2072.963520] env[62824]: DEBUG oslo_vmware.api [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52b2cd5d-7a48-5783-202e-1cacd02cc940, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2073.083332] env[62824]: DEBUG oslo_vmware.api [None req-4715e6dc-1d1a-49a2-9d97-4f4b36f634e1 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2146214, 'name': PowerOffVM_Task, 'duration_secs': 0.293925} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2073.083611] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-4715e6dc-1d1a-49a2-9d97-4f4b36f634e1 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2073.083771] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-4715e6dc-1d1a-49a2-9d97-4f4b36f634e1 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2073.084043] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a1a6f227-b13f-40d7-bd79-c1d782b64fc4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2073.105883] env[62824]: DEBUG nova.network.neutron [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 15479062-af75-4925-99b3-77d6a49751ad] Updating instance_info_cache with network_info: [{"id": "6b8b3a2d-f55e-4aa8-b995-4880c1adea60", "address": "fa:16:3e:77:1e:5b", "network": {"id": "6f888a0c-8089-4dd4-aace-14b020e11bee", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-137042394-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d041a8209e3848eba8e47251e013ca17", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5f60c972-a72d-4c5f-a250-faadfd6eafbe", "external-id": "nsx-vlan-transportzone-932", "segmentation_id": 932, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6b8b3a2d-f5", "ovs_interfaceid": "6b8b3a2d-f55e-4aa8-b995-4880c1adea60", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2073.209049] env[62824]: DEBUG oslo_vmware.api [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Task: {'id': task-2146215, 'name': ReconfigVM_Task, 'duration_secs': 0.376943} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2073.209815] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Reconfigured VM instance instance-0000003a to attach disk [datastore2] volume-8d462122-e7ae-4257-8ce4-0a79a4acae13/volume-8d462122-e7ae-4257-8ce4-0a79a4acae13.vmdk or device None with type thin {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2073.216098] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-61331914-40ce-4330-8e23-d4a4c0c0d5fb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2073.237077] env[62824]: DEBUG oslo_vmware.api [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Waiting for the task: (returnval){ [ 2073.237077] env[62824]: value = "task-2146219" [ 2073.237077] env[62824]: _type = "Task" [ 2073.237077] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2073.248649] env[62824]: DEBUG oslo_vmware.api [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Task: {'id': task-2146219, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2073.275059] env[62824]: DEBUG oslo_vmware.api [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Task: {'id': task-2146208, 'name': PowerOnVM_Task, 'duration_secs': 2.344773} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2073.277456] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] [instance: ed7d3e95-f26e-40fd-a370-e17922bbff8e] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2073.277694] env[62824]: DEBUG nova.compute.manager [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] [instance: ed7d3e95-f26e-40fd-a370-e17922bbff8e] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2073.278635] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2689009-1fbb-4760-a3ac-b7432f2260c1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2073.327333] env[62824]: DEBUG oslo_vmware.api [None req-03bb66d7-d558-4473-adb2-183a408a723b tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146216, 'name': ReconfigVM_Task, 'duration_secs': 0.49631} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2073.327879] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-03bb66d7-d558-4473-adb2-183a408a723b tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: b0456b8a-348a-4503-a92c-58e5ab455d1c] Reconfigured VM instance instance-00000038 to attach disk [datastore1] b0456b8a-348a-4503-a92c-58e5ab455d1c/9e2a7d30-212d-4ab8-9606-c5c6d52629e8-rescue.vmdk or device None with type thin {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2073.328645] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5f821e8-75cf-4155-a48c-977904400c49 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2073.361654] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6a022504-72c4-40c2-8cf1-19d136db02be {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2073.380738] env[62824]: DEBUG oslo_vmware.api [None req-03bb66d7-d558-4473-adb2-183a408a723b tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Waiting for the task: (returnval){ [ 2073.380738] env[62824]: value = "task-2146220" [ 2073.380738] env[62824]: _type = "Task" [ 2073.380738] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2073.396894] env[62824]: DEBUG oslo_vmware.api [None req-03bb66d7-d558-4473-adb2-183a408a723b tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146220, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2073.400633] env[62824]: DEBUG oslo_concurrency.lockutils [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.756s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2073.401170] env[62824]: DEBUG nova.compute.manager [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02f34565-6675-4c79-ac47-b131ceba9df8] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2073.405967] env[62824]: DEBUG oslo_vmware.api [None req-423ea96c-8206-47c1-8b7e-a945775631af tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Task: {'id': task-2146217, 'name': PowerOffVM_Task, 'duration_secs': 0.270671} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2073.407412] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9a50fcc7-25bf-46e5-b936-5cbc1f3c424b tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.790s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2073.407412] env[62824]: DEBUG nova.objects.instance [None req-9a50fcc7-25bf-46e5-b936-5cbc1f3c424b tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Lazy-loading 'resources' on Instance uuid c63b05b7-1dbf-4637-8728-33673be0d3ea {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2073.407956] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-423ea96c-8206-47c1-8b7e-a945775631af tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2073.408141] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-423ea96c-8206-47c1-8b7e-a945775631af tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2073.409205] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cf3e06de-61d9-4a52-b753-6deb4471ee09 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2073.423677] env[62824]: DEBUG oslo_vmware.api [None req-c94e74ba-f9b2-469b-be3d-b77fc48e94ea tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146191, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2073.460221] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-4715e6dc-1d1a-49a2-9d97-4f4b36f634e1 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2073.460494] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-4715e6dc-1d1a-49a2-9d97-4f4b36f634e1 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2] Deleting contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2073.460846] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-4715e6dc-1d1a-49a2-9d97-4f4b36f634e1 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Deleting the datastore file [datastore2] 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2073.461199] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d58526cc-4abe-4d61-ab42-0c5dbe0c52cf {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2073.467650] env[62824]: DEBUG oslo_vmware.api [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52b2cd5d-7a48-5783-202e-1cacd02cc940, 'name': SearchDatastore_Task, 'duration_secs': 0.01197} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2073.468836] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-506004fe-6660-43fb-9ad9-abca62547b29 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2073.473556] env[62824]: DEBUG oslo_vmware.api [None req-4715e6dc-1d1a-49a2-9d97-4f4b36f634e1 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Waiting for the task: (returnval){ [ 2073.473556] env[62824]: value = "task-2146222" [ 2073.473556] env[62824]: _type = "Task" [ 2073.473556] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2073.474741] env[62824]: DEBUG oslo_vmware.api [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 2073.474741] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5221d4c3-c333-f2a3-99b2-569eb490e82e" [ 2073.474741] env[62824]: _type = "Task" [ 2073.474741] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2073.490482] env[62824]: DEBUG oslo_vmware.api [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5221d4c3-c333-f2a3-99b2-569eb490e82e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2073.493156] env[62824]: DEBUG oslo_vmware.api [None req-4715e6dc-1d1a-49a2-9d97-4f4b36f634e1 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2146222, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2073.569462] env[62824]: DEBUG nova.network.neutron [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: a6cd2032-de60-4f78-bf1e-79801d049df0] Successfully updated port: 018929f6-4f6c-4148-b1eb-9cf11f5cc746 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2073.609332] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Releasing lock "refresh_cache-15479062-af75-4925-99b3-77d6a49751ad" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2073.609677] env[62824]: DEBUG nova.compute.manager [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 15479062-af75-4925-99b3-77d6a49751ad] Instance network_info: |[{"id": "6b8b3a2d-f55e-4aa8-b995-4880c1adea60", "address": "fa:16:3e:77:1e:5b", "network": {"id": "6f888a0c-8089-4dd4-aace-14b020e11bee", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-137042394-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d041a8209e3848eba8e47251e013ca17", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5f60c972-a72d-4c5f-a250-faadfd6eafbe", "external-id": "nsx-vlan-transportzone-932", "segmentation_id": 932, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6b8b3a2d-f5", "ovs_interfaceid": "6b8b3a2d-f55e-4aa8-b995-4880c1adea60", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2073.610158] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 15479062-af75-4925-99b3-77d6a49751ad] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:77:1e:5b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5f60c972-a72d-4c5f-a250-faadfd6eafbe', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6b8b3a2d-f55e-4aa8-b995-4880c1adea60', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2073.619214] env[62824]: DEBUG oslo.service.loopingcall [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2073.619830] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 15479062-af75-4925-99b3-77d6a49751ad] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2073.620109] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7dd33acf-adfc-4bd0-8988-a573339e9551 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2073.646805] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2073.646805] env[62824]: value = "task-2146223" [ 2073.646805] env[62824]: _type = "Task" [ 2073.646805] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2073.656847] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146223, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2073.753458] env[62824]: DEBUG oslo_vmware.api [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Task: {'id': task-2146219, 'name': ReconfigVM_Task, 'duration_secs': 0.145594} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2073.753845] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-438591', 'volume_id': '8d462122-e7ae-4257-8ce4-0a79a4acae13', 'name': 'volume-8d462122-e7ae-4257-8ce4-0a79a4acae13', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '2f315f9e-94e4-47ef-9503-ee92b59e5452', 'attached_at': '', 'detached_at': '', 'volume_id': '8d462122-e7ae-4257-8ce4-0a79a4acae13', 'serial': '8d462122-e7ae-4257-8ce4-0a79a4acae13'} {{(pid=62824) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 2073.754412] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-14ffaeba-3743-44db-8713-19edc5cb9123 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2073.763260] env[62824]: DEBUG oslo_vmware.api [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Waiting for the task: (returnval){ [ 2073.763260] env[62824]: value = "task-2146224" [ 2073.763260] env[62824]: _type = "Task" [ 2073.763260] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2073.776249] env[62824]: DEBUG oslo_vmware.api [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Task: {'id': task-2146224, 'name': Rename_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2073.805057] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2073.895674] env[62824]: DEBUG oslo_vmware.api [None req-03bb66d7-d558-4473-adb2-183a408a723b tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146220, 'name': ReconfigVM_Task, 'duration_secs': 0.399799} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2073.895986] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-03bb66d7-d558-4473-adb2-183a408a723b tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: b0456b8a-348a-4503-a92c-58e5ab455d1c] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2073.896306] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f3cb027f-4341-4788-ba37-45e020d49d2b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2073.903817] env[62824]: DEBUG oslo_vmware.api [None req-03bb66d7-d558-4473-adb2-183a408a723b tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Waiting for the task: (returnval){ [ 2073.903817] env[62824]: value = "task-2146225" [ 2073.903817] env[62824]: _type = "Task" [ 2073.903817] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2073.909076] env[62824]: DEBUG nova.compute.utils [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2073.910483] env[62824]: DEBUG nova.compute.manager [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02f34565-6675-4c79-ac47-b131ceba9df8] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2073.910654] env[62824]: DEBUG nova.network.neutron [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02f34565-6675-4c79-ac47-b131ceba9df8] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2073.927416] env[62824]: DEBUG oslo_vmware.api [None req-03bb66d7-d558-4473-adb2-183a408a723b tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146225, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2073.934960] env[62824]: DEBUG oslo_vmware.api [None req-c94e74ba-f9b2-469b-be3d-b77fc48e94ea tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146191, 'name': ReconfigVM_Task, 'duration_secs': 5.862341} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2073.937448] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c94e74ba-f9b2-469b-be3d-b77fc48e94ea tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Releasing lock "f4d63a93-23af-470c-b36a-662af81dc386" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2073.937448] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c94e74ba-f9b2-469b-be3d-b77fc48e94ea tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Reconfigured VM to detach interface {{(pid=62824) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 2073.990269] env[62824]: DEBUG nova.policy [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a6a63328f3534c7c912384254ccda1f0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '248af2bc3e7b41d2a74b0c6074b359e1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2073.999308] env[62824]: DEBUG oslo_vmware.api [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5221d4c3-c333-f2a3-99b2-569eb490e82e, 'name': SearchDatastore_Task, 'duration_secs': 0.025697} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2073.999415] env[62824]: DEBUG oslo_vmware.api [None req-4715e6dc-1d1a-49a2-9d97-4f4b36f634e1 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2146222, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.174209} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2074.003140] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2074.003372] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 919e8a45-7810-4a8d-a1aa-5046b5ab059c/919e8a45-7810-4a8d-a1aa-5046b5ab059c.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2074.003736] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-4715e6dc-1d1a-49a2-9d97-4f4b36f634e1 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2074.006141] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-4715e6dc-1d1a-49a2-9d97-4f4b36f634e1 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2] Deleted contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2074.006141] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-4715e6dc-1d1a-49a2-9d97-4f4b36f634e1 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2074.006141] env[62824]: INFO nova.compute.manager [None req-4715e6dc-1d1a-49a2-9d97-4f4b36f634e1 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2] Took 1.46 seconds to destroy the instance on the hypervisor. [ 2074.006141] env[62824]: DEBUG oslo.service.loopingcall [None req-4715e6dc-1d1a-49a2-9d97-4f4b36f634e1 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2074.006141] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e51f79bf-ff27-4e64-91ae-489f0ee55c6a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2074.007420] env[62824]: DEBUG nova.compute.manager [-] [instance: 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2074.007566] env[62824]: DEBUG nova.network.neutron [-] [instance: 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2074.027736] env[62824]: DEBUG oslo_vmware.api [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 2074.027736] env[62824]: value = "task-2146226" [ 2074.027736] env[62824]: _type = "Task" [ 2074.027736] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2074.034195] env[62824]: DEBUG nova.compute.manager [req-10c1d576-2089-4a57-a277-7f27157fd2d5 req-1632e8b8-4036-4299-80ac-d65a8f3bff48 service nova] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Received event network-vif-unplugged-60a4fce0-0e63-45af-890c-46ca44ea9a0c {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2074.034558] env[62824]: DEBUG oslo_concurrency.lockutils [req-10c1d576-2089-4a57-a277-7f27157fd2d5 req-1632e8b8-4036-4299-80ac-d65a8f3bff48 service nova] Acquiring lock "b8cc8cd3-ea03-40bf-b867-7ad193365552-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2074.034802] env[62824]: DEBUG oslo_concurrency.lockutils [req-10c1d576-2089-4a57-a277-7f27157fd2d5 req-1632e8b8-4036-4299-80ac-d65a8f3bff48 service nova] Lock "b8cc8cd3-ea03-40bf-b867-7ad193365552-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2074.034969] env[62824]: DEBUG oslo_concurrency.lockutils [req-10c1d576-2089-4a57-a277-7f27157fd2d5 req-1632e8b8-4036-4299-80ac-d65a8f3bff48 service nova] Lock "b8cc8cd3-ea03-40bf-b867-7ad193365552-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2074.035150] env[62824]: DEBUG nova.compute.manager [req-10c1d576-2089-4a57-a277-7f27157fd2d5 req-1632e8b8-4036-4299-80ac-d65a8f3bff48 service nova] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] No waiting events found dispatching network-vif-unplugged-60a4fce0-0e63-45af-890c-46ca44ea9a0c {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2074.035318] env[62824]: WARNING nova.compute.manager [req-10c1d576-2089-4a57-a277-7f27157fd2d5 req-1632e8b8-4036-4299-80ac-d65a8f3bff48 service nova] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Received unexpected event network-vif-unplugged-60a4fce0-0e63-45af-890c-46ca44ea9a0c for instance with vm_state shelved_offloaded and task_state None. [ 2074.035473] env[62824]: DEBUG nova.compute.manager [req-10c1d576-2089-4a57-a277-7f27157fd2d5 req-1632e8b8-4036-4299-80ac-d65a8f3bff48 service nova] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Received event network-changed-60a4fce0-0e63-45af-890c-46ca44ea9a0c {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2074.035622] env[62824]: DEBUG nova.compute.manager [req-10c1d576-2089-4a57-a277-7f27157fd2d5 req-1632e8b8-4036-4299-80ac-d65a8f3bff48 service nova] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Refreshing instance network info cache due to event network-changed-60a4fce0-0e63-45af-890c-46ca44ea9a0c. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2074.035825] env[62824]: DEBUG oslo_concurrency.lockutils [req-10c1d576-2089-4a57-a277-7f27157fd2d5 req-1632e8b8-4036-4299-80ac-d65a8f3bff48 service nova] Acquiring lock "refresh_cache-b8cc8cd3-ea03-40bf-b867-7ad193365552" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2074.035959] env[62824]: DEBUG oslo_concurrency.lockutils [req-10c1d576-2089-4a57-a277-7f27157fd2d5 req-1632e8b8-4036-4299-80ac-d65a8f3bff48 service nova] Acquired lock "refresh_cache-b8cc8cd3-ea03-40bf-b867-7ad193365552" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2074.036209] env[62824]: DEBUG nova.network.neutron [req-10c1d576-2089-4a57-a277-7f27157fd2d5 req-1632e8b8-4036-4299-80ac-d65a8f3bff48 service nova] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Refreshing network info cache for port 60a4fce0-0e63-45af-890c-46ca44ea9a0c {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2074.044246] env[62824]: DEBUG oslo_vmware.api [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146226, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2074.072559] env[62824]: DEBUG oslo_concurrency.lockutils [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquiring lock "refresh_cache-a6cd2032-de60-4f78-bf1e-79801d049df0" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2074.072804] env[62824]: DEBUG oslo_concurrency.lockutils [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquired lock "refresh_cache-a6cd2032-de60-4f78-bf1e-79801d049df0" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2074.073038] env[62824]: DEBUG nova.network.neutron [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: a6cd2032-de60-4f78-bf1e-79801d049df0] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2074.163884] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146223, 'name': CreateVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2074.197158] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-423ea96c-8206-47c1-8b7e-a945775631af tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2074.197393] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-423ea96c-8206-47c1-8b7e-a945775631af tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Deleting contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2074.197624] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-423ea96c-8206-47c1-8b7e-a945775631af tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Deleting the datastore file [datastore1] 29b8dd5f-1855-490c-a01b-54840073a753 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2074.197967] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0e5ec416-2e1e-46ab-8136-cfeb8b463e38 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2074.212915] env[62824]: DEBUG oslo_vmware.api [None req-423ea96c-8206-47c1-8b7e-a945775631af tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Waiting for the task: (returnval){ [ 2074.212915] env[62824]: value = "task-2146227" [ 2074.212915] env[62824]: _type = "Task" [ 2074.212915] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2074.273532] env[62824]: DEBUG oslo_vmware.api [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Task: {'id': task-2146224, 'name': Rename_Task, 'duration_secs': 0.149412} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2074.273809] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2074.274076] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-658ceee8-db2e-423f-bf54-f3f7a45fcd81 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2074.284073] env[62824]: DEBUG oslo_vmware.api [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Waiting for the task: (returnval){ [ 2074.284073] env[62824]: value = "task-2146228" [ 2074.284073] env[62824]: _type = "Task" [ 2074.284073] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2074.294689] env[62824]: DEBUG oslo_vmware.api [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Task: {'id': task-2146228, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2074.418886] env[62824]: DEBUG nova.compute.manager [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02f34565-6675-4c79-ac47-b131ceba9df8] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2074.422350] env[62824]: DEBUG oslo_vmware.api [None req-03bb66d7-d558-4473-adb2-183a408a723b tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146225, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2074.489877] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ee2d53f-9734-4f81-8bc0-d6bf1ef1242b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2074.505465] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8543c7e6-54b1-4b7f-8d07-607086174754 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2074.559481] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23cb2aa0-ea77-48b1-9479-d39233243818 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2074.577371] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2f21a9c-c5a0-4ace-9abc-71c60e5d4bfe {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2074.586388] env[62824]: DEBUG oslo_vmware.api [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146226, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2074.602038] env[62824]: DEBUG nova.compute.provider_tree [None req-9a50fcc7-25bf-46e5-b936-5cbc1f3c424b tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2074.649055] env[62824]: DEBUG nova.network.neutron [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02f34565-6675-4c79-ac47-b131ceba9df8] Successfully created port: be36fef6-17a7-4c25-bf33-98cc6a037b53 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2074.663990] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146223, 'name': CreateVM_Task, 'duration_secs': 0.558549} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2074.664211] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 15479062-af75-4925-99b3-77d6a49751ad] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2074.664936] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2074.665389] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2074.665660] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2074.665958] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c2207372-5f80-4405-93a5-00013fd21c36 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2074.676605] env[62824]: DEBUG oslo_vmware.api [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Waiting for the task: (returnval){ [ 2074.676605] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5228c1e1-0666-c55f-55c9-d35ee813f5ff" [ 2074.676605] env[62824]: _type = "Task" [ 2074.676605] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2074.681600] env[62824]: DEBUG nova.network.neutron [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: a6cd2032-de60-4f78-bf1e-79801d049df0] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2074.690459] env[62824]: DEBUG oslo_vmware.api [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5228c1e1-0666-c55f-55c9-d35ee813f5ff, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2074.728594] env[62824]: DEBUG oslo_vmware.api [None req-423ea96c-8206-47c1-8b7e-a945775631af tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Task: {'id': task-2146227, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.367737} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2074.728746] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-423ea96c-8206-47c1-8b7e-a945775631af tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2074.728937] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-423ea96c-8206-47c1-8b7e-a945775631af tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Deleted contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2074.729134] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-423ea96c-8206-47c1-8b7e-a945775631af tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2074.729310] env[62824]: INFO nova.compute.manager [None req-423ea96c-8206-47c1-8b7e-a945775631af tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Took 1.87 seconds to destroy the instance on the hypervisor. [ 2074.729555] env[62824]: DEBUG oslo.service.loopingcall [None req-423ea96c-8206-47c1-8b7e-a945775631af tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2074.729750] env[62824]: DEBUG nova.compute.manager [-] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2074.729848] env[62824]: DEBUG nova.network.neutron [-] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2074.798104] env[62824]: DEBUG oslo_vmware.api [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Task: {'id': task-2146228, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2074.918416] env[62824]: DEBUG oslo_vmware.api [None req-03bb66d7-d558-4473-adb2-183a408a723b tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146225, 'name': PowerOnVM_Task, 'duration_secs': 0.780424} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2074.918416] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-03bb66d7-d558-4473-adb2-183a408a723b tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: b0456b8a-348a-4503-a92c-58e5ab455d1c] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2074.922621] env[62824]: DEBUG nova.compute.manager [None req-03bb66d7-d558-4473-adb2-183a408a723b tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: b0456b8a-348a-4503-a92c-58e5ab455d1c] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2074.923525] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fa5fe35-294f-493c-adb4-f559f9305b3d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2075.061981] env[62824]: DEBUG oslo_vmware.api [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146226, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.681925} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2075.062294] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 919e8a45-7810-4a8d-a1aa-5046b5ab059c/919e8a45-7810-4a8d-a1aa-5046b5ab059c.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2075.062662] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 919e8a45-7810-4a8d-a1aa-5046b5ab059c] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2075.062906] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a43183c9-0646-4d0e-bd49-5aa2076bc2e3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2075.077170] env[62824]: DEBUG oslo_vmware.api [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 2075.077170] env[62824]: value = "task-2146229" [ 2075.077170] env[62824]: _type = "Task" [ 2075.077170] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2075.087839] env[62824]: DEBUG oslo_vmware.api [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146229, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2075.109020] env[62824]: DEBUG nova.scheduler.client.report [None req-9a50fcc7-25bf-46e5-b936-5cbc1f3c424b tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2075.132805] env[62824]: DEBUG nova.network.neutron [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: a6cd2032-de60-4f78-bf1e-79801d049df0] Updating instance_info_cache with network_info: [{"id": "018929f6-4f6c-4148-b1eb-9cf11f5cc746", "address": "fa:16:3e:38:48:15", "network": {"id": "0f512108-670f-4a85-9381-108e18b3b5ed", "bridge": "br-int", "label": "tempest-ServersTestJSON-766397782-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dfff08982dad4790bf4d555e2b4db5e4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f3a2eb5-353f-45c5-a73b-869626f4bb13", "external-id": "nsx-vlan-transportzone-411", "segmentation_id": 411, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap018929f6-4f", "ovs_interfaceid": "018929f6-4f6c-4148-b1eb-9cf11f5cc746", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2075.190228] env[62824]: DEBUG oslo_vmware.api [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5228c1e1-0666-c55f-55c9-d35ee813f5ff, 'name': SearchDatastore_Task, 'duration_secs': 0.018288} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2075.193043] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2075.193392] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 15479062-af75-4925-99b3-77d6a49751ad] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2075.193677] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2075.193870] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2075.194101] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2075.194438] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-eff99cb7-470c-4c36-a467-efba3579fed0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2075.205434] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2075.205625] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2075.206434] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d680421f-0898-4782-b5ca-34c8bbab5944 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2075.214347] env[62824]: DEBUG oslo_vmware.api [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Waiting for the task: (returnval){ [ 2075.214347] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52a0e0af-cb6c-78ea-ec7c-07e861c7bb32" [ 2075.214347] env[62824]: _type = "Task" [ 2075.214347] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2075.226597] env[62824]: DEBUG oslo_vmware.api [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52a0e0af-cb6c-78ea-ec7c-07e861c7bb32, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2075.292823] env[62824]: DEBUG nova.network.neutron [req-10c1d576-2089-4a57-a277-7f27157fd2d5 req-1632e8b8-4036-4299-80ac-d65a8f3bff48 service nova] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Updated VIF entry in instance network info cache for port 60a4fce0-0e63-45af-890c-46ca44ea9a0c. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2075.293191] env[62824]: DEBUG nova.network.neutron [req-10c1d576-2089-4a57-a277-7f27157fd2d5 req-1632e8b8-4036-4299-80ac-d65a8f3bff48 service nova] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Updating instance_info_cache with network_info: [{"id": "60a4fce0-0e63-45af-890c-46ca44ea9a0c", "address": "fa:16:3e:cf:18:eb", "network": {"id": "db233b0c-27f3-4a76-ae9c-2b5889bf3d3b", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-1825671448-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.231", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3240be4f32254f54aa87cb045e3d8f31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap60a4fce0-0e", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2075.301708] env[62824]: DEBUG oslo_vmware.api [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Task: {'id': task-2146228, 'name': PowerOnVM_Task, 'duration_secs': 0.878216} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2075.301708] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2075.301935] env[62824]: INFO nova.compute.manager [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Took 8.67 seconds to spawn the instance on the hypervisor. [ 2075.302158] env[62824]: DEBUG nova.compute.manager [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2075.302993] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d62d3f35-1c3f-4d72-b16e-edad42cff835 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2075.440329] env[62824]: DEBUG nova.compute.manager [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02f34565-6675-4c79-ac47-b131ceba9df8] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2075.442664] env[62824]: DEBUG nova.network.neutron [-] [instance: 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2075.479088] env[62824]: DEBUG nova.virt.hardware [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2075.479088] env[62824]: DEBUG nova.virt.hardware [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2075.479088] env[62824]: DEBUG nova.virt.hardware [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2075.479088] env[62824]: DEBUG nova.virt.hardware [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2075.479326] env[62824]: DEBUG nova.virt.hardware [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2075.479579] env[62824]: DEBUG nova.virt.hardware [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2075.479801] env[62824]: DEBUG nova.virt.hardware [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2075.479957] env[62824]: DEBUG nova.virt.hardware [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2075.480137] env[62824]: DEBUG nova.virt.hardware [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2075.480836] env[62824]: DEBUG nova.virt.hardware [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2075.480836] env[62824]: DEBUG nova.virt.hardware [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2075.482090] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08a63683-3c55-4d4a-b6d0-624999cf4a89 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2075.491688] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8936352e-a981-4099-9ca5-da972d7e38aa {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2075.569078] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c94e74ba-f9b2-469b-be3d-b77fc48e94ea tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquiring lock "refresh_cache-f4d63a93-23af-470c-b36a-662af81dc386" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2075.569316] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c94e74ba-f9b2-469b-be3d-b77fc48e94ea tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquired lock "refresh_cache-f4d63a93-23af-470c-b36a-662af81dc386" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2075.569554] env[62824]: DEBUG nova.network.neutron [None req-c94e74ba-f9b2-469b-be3d-b77fc48e94ea tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2075.592103] env[62824]: DEBUG oslo_vmware.api [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146229, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.095358} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2075.592403] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 919e8a45-7810-4a8d-a1aa-5046b5ab059c] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2075.594798] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de9db174-55c8-4b53-93d6-f158b059f67b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2075.601819] env[62824]: DEBUG nova.compute.manager [req-c4eff0be-169f-4025-a8d0-b01f2eab4019 req-8034a27a-a152-44ba-b39a-4d40bd999a6e service nova] [instance: 15479062-af75-4925-99b3-77d6a49751ad] Received event network-vif-plugged-6b8b3a2d-f55e-4aa8-b995-4880c1adea60 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2075.601819] env[62824]: DEBUG oslo_concurrency.lockutils [req-c4eff0be-169f-4025-a8d0-b01f2eab4019 req-8034a27a-a152-44ba-b39a-4d40bd999a6e service nova] Acquiring lock "15479062-af75-4925-99b3-77d6a49751ad-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2075.601819] env[62824]: DEBUG oslo_concurrency.lockutils [req-c4eff0be-169f-4025-a8d0-b01f2eab4019 req-8034a27a-a152-44ba-b39a-4d40bd999a6e service nova] Lock "15479062-af75-4925-99b3-77d6a49751ad-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2075.601819] env[62824]: DEBUG oslo_concurrency.lockutils [req-c4eff0be-169f-4025-a8d0-b01f2eab4019 req-8034a27a-a152-44ba-b39a-4d40bd999a6e service nova] Lock "15479062-af75-4925-99b3-77d6a49751ad-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2075.601819] env[62824]: DEBUG nova.compute.manager [req-c4eff0be-169f-4025-a8d0-b01f2eab4019 req-8034a27a-a152-44ba-b39a-4d40bd999a6e service nova] [instance: 15479062-af75-4925-99b3-77d6a49751ad] No waiting events found dispatching network-vif-plugged-6b8b3a2d-f55e-4aa8-b995-4880c1adea60 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2075.601819] env[62824]: WARNING nova.compute.manager [req-c4eff0be-169f-4025-a8d0-b01f2eab4019 req-8034a27a-a152-44ba-b39a-4d40bd999a6e service nova] [instance: 15479062-af75-4925-99b3-77d6a49751ad] Received unexpected event network-vif-plugged-6b8b3a2d-f55e-4aa8-b995-4880c1adea60 for instance with vm_state building and task_state spawning. [ 2075.601819] env[62824]: DEBUG nova.compute.manager [req-c4eff0be-169f-4025-a8d0-b01f2eab4019 req-8034a27a-a152-44ba-b39a-4d40bd999a6e service nova] [instance: 15479062-af75-4925-99b3-77d6a49751ad] Received event network-changed-6b8b3a2d-f55e-4aa8-b995-4880c1adea60 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2075.601819] env[62824]: DEBUG nova.compute.manager [req-c4eff0be-169f-4025-a8d0-b01f2eab4019 req-8034a27a-a152-44ba-b39a-4d40bd999a6e service nova] [instance: 15479062-af75-4925-99b3-77d6a49751ad] Refreshing instance network info cache due to event network-changed-6b8b3a2d-f55e-4aa8-b995-4880c1adea60. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2075.601819] env[62824]: DEBUG oslo_concurrency.lockutils [req-c4eff0be-169f-4025-a8d0-b01f2eab4019 req-8034a27a-a152-44ba-b39a-4d40bd999a6e service nova] Acquiring lock "refresh_cache-15479062-af75-4925-99b3-77d6a49751ad" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2075.601819] env[62824]: DEBUG oslo_concurrency.lockutils [req-c4eff0be-169f-4025-a8d0-b01f2eab4019 req-8034a27a-a152-44ba-b39a-4d40bd999a6e service nova] Acquired lock "refresh_cache-15479062-af75-4925-99b3-77d6a49751ad" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2075.605499] env[62824]: DEBUG nova.network.neutron [req-c4eff0be-169f-4025-a8d0-b01f2eab4019 req-8034a27a-a152-44ba-b39a-4d40bd999a6e service nova] [instance: 15479062-af75-4925-99b3-77d6a49751ad] Refreshing network info cache for port 6b8b3a2d-f55e-4aa8-b995-4880c1adea60 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2075.620277] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9a50fcc7-25bf-46e5-b936-5cbc1f3c424b tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.214s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2075.632787] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 919e8a45-7810-4a8d-a1aa-5046b5ab059c] Reconfiguring VM instance instance-0000003b to attach disk [datastore2] 919e8a45-7810-4a8d-a1aa-5046b5ab059c/919e8a45-7810-4a8d-a1aa-5046b5ab059c.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2075.639338] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 20.446s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2075.642276] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-90eaf1ae-efbd-48e0-8633-f7ac765703c0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2075.660689] env[62824]: DEBUG oslo_concurrency.lockutils [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Releasing lock "refresh_cache-a6cd2032-de60-4f78-bf1e-79801d049df0" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2075.661364] env[62824]: DEBUG nova.compute.manager [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: a6cd2032-de60-4f78-bf1e-79801d049df0] Instance network_info: |[{"id": "018929f6-4f6c-4148-b1eb-9cf11f5cc746", "address": "fa:16:3e:38:48:15", "network": {"id": "0f512108-670f-4a85-9381-108e18b3b5ed", "bridge": "br-int", "label": "tempest-ServersTestJSON-766397782-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dfff08982dad4790bf4d555e2b4db5e4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f3a2eb5-353f-45c5-a73b-869626f4bb13", "external-id": "nsx-vlan-transportzone-411", "segmentation_id": 411, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap018929f6-4f", "ovs_interfaceid": "018929f6-4f6c-4148-b1eb-9cf11f5cc746", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2075.662808] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: a6cd2032-de60-4f78-bf1e-79801d049df0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:38:48:15', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9f3a2eb5-353f-45c5-a73b-869626f4bb13', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '018929f6-4f6c-4148-b1eb-9cf11f5cc746', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2075.672185] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Creating folder: Project (dfff08982dad4790bf4d555e2b4db5e4). Parent ref: group-v438503. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 2075.673760] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bd24f673-1c60-430d-8ce6-6604c740de23 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2075.676716] env[62824]: INFO nova.scheduler.client.report [None req-9a50fcc7-25bf-46e5-b936-5cbc1f3c424b tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Deleted allocations for instance c63b05b7-1dbf-4637-8728-33673be0d3ea [ 2075.682284] env[62824]: DEBUG oslo_vmware.api [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 2075.682284] env[62824]: value = "task-2146230" [ 2075.682284] env[62824]: _type = "Task" [ 2075.682284] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2075.691403] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Created folder: Project (dfff08982dad4790bf4d555e2b4db5e4) in parent group-v438503. [ 2075.691594] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Creating folder: Instances. Parent ref: group-v438660. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 2075.692256] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e3e27370-77a8-46dd-bfb4-f6045651e33e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2075.698579] env[62824]: DEBUG oslo_vmware.api [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146230, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2075.703878] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Created folder: Instances in parent group-v438660. [ 2075.704398] env[62824]: DEBUG oslo.service.loopingcall [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2075.704398] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a6cd2032-de60-4f78-bf1e-79801d049df0] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2075.704557] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c973d290-5369-4446-ade5-f587971ffa3b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2075.736936] env[62824]: DEBUG oslo_vmware.api [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52a0e0af-cb6c-78ea-ec7c-07e861c7bb32, 'name': SearchDatastore_Task, 'duration_secs': 0.037874} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2075.738841] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2075.738841] env[62824]: value = "task-2146233" [ 2075.738841] env[62824]: _type = "Task" [ 2075.738841] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2075.740685] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-88da5d5c-47a2-4b58-9b60-4ee977a6d947 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2075.753515] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146233, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2075.753701] env[62824]: DEBUG oslo_vmware.api [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Waiting for the task: (returnval){ [ 2075.753701] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52b0a486-9875-e7f0-6708-606a8b283133" [ 2075.753701] env[62824]: _type = "Task" [ 2075.753701] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2075.765159] env[62824]: DEBUG oslo_vmware.api [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52b0a486-9875-e7f0-6708-606a8b283133, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2075.797555] env[62824]: DEBUG oslo_concurrency.lockutils [req-10c1d576-2089-4a57-a277-7f27157fd2d5 req-1632e8b8-4036-4299-80ac-d65a8f3bff48 service nova] Releasing lock "refresh_cache-b8cc8cd3-ea03-40bf-b867-7ad193365552" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2075.831821] env[62824]: INFO nova.compute.manager [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Took 40.41 seconds to build instance. [ 2075.949298] env[62824]: INFO nova.compute.manager [-] [instance: 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2] Took 1.94 seconds to deallocate network for instance. [ 2076.198023] env[62824]: DEBUG oslo_vmware.api [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146230, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2076.198023] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9a50fcc7-25bf-46e5-b936-5cbc1f3c424b tempest-InstanceActionsNegativeTestJSON-1374647901 tempest-InstanceActionsNegativeTestJSON-1374647901-project-member] Lock "c63b05b7-1dbf-4637-8728-33673be0d3ea" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.364s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2076.258269] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146233, 'name': CreateVM_Task, 'duration_secs': 0.451066} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2076.259871] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a6cd2032-de60-4f78-bf1e-79801d049df0] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2076.260762] env[62824]: DEBUG oslo_concurrency.lockutils [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2076.260880] env[62824]: DEBUG oslo_concurrency.lockutils [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2076.263171] env[62824]: DEBUG oslo_concurrency.lockutils [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2076.263171] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f3e71c1e-90e7-4402-9937-786886311ba6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2076.270944] env[62824]: DEBUG oslo_vmware.api [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52b0a486-9875-e7f0-6708-606a8b283133, 'name': SearchDatastore_Task, 'duration_secs': 0.020073} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2076.274886] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2076.275362] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 15479062-af75-4925-99b3-77d6a49751ad/15479062-af75-4925-99b3-77d6a49751ad.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2076.275834] env[62824]: DEBUG oslo_vmware.api [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2076.275834] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52edcf83-c550-b116-4dde-832e7d691359" [ 2076.275834] env[62824]: _type = "Task" [ 2076.275834] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2076.276141] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c5b4a999-18e2-4fee-bfb7-77ff7d9fa957 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2076.289589] env[62824]: DEBUG oslo_vmware.api [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52edcf83-c550-b116-4dde-832e7d691359, 'name': SearchDatastore_Task, 'duration_secs': 0.015699} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2076.291192] env[62824]: DEBUG oslo_concurrency.lockutils [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2076.291442] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: a6cd2032-de60-4f78-bf1e-79801d049df0] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2076.291680] env[62824]: DEBUG oslo_concurrency.lockutils [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2076.291850] env[62824]: DEBUG oslo_concurrency.lockutils [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2076.292124] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2076.292424] env[62824]: DEBUG oslo_vmware.api [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Waiting for the task: (returnval){ [ 2076.292424] env[62824]: value = "task-2146234" [ 2076.292424] env[62824]: _type = "Task" [ 2076.292424] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2076.292651] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-35420972-5e12-4570-88aa-2f3616be6258 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2076.308143] env[62824]: DEBUG oslo_vmware.api [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Task: {'id': task-2146234, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2076.309769] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2076.309946] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2076.310754] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f68cf49f-52eb-4402-a700-1023794067b9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2076.317458] env[62824]: DEBUG oslo_vmware.api [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2076.317458] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52ba44c1-06c2-32d1-bb8c-744a92ce0e62" [ 2076.317458] env[62824]: _type = "Task" [ 2076.317458] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2076.333991] env[62824]: DEBUG oslo_concurrency.lockutils [None req-51ba8a2e-4c65-43ad-91a2-3cbe6d407dfe tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Lock "2f315f9e-94e4-47ef-9503-ee92b59e5452" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 84.012s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2076.335755] env[62824]: DEBUG oslo_vmware.api [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52ba44c1-06c2-32d1-bb8c-744a92ce0e62, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2076.357840] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Acquiring lock "350cfd12-a8d5-4a5f-b3de-d795fa179dfd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2076.360640] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Lock "350cfd12-a8d5-4a5f-b3de-d795fa179dfd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2076.459537] env[62824]: DEBUG oslo_concurrency.lockutils [None req-4715e6dc-1d1a-49a2-9d97-4f4b36f634e1 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2076.495820] env[62824]: DEBUG nova.network.neutron [-] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2076.589689] env[62824]: INFO nova.network.neutron [None req-c94e74ba-f9b2-469b-be3d-b77fc48e94ea tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Port c34f030d-46da-4a7a-9703-5acb2e526d50 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 2076.589689] env[62824]: DEBUG nova.network.neutron [None req-c94e74ba-f9b2-469b-be3d-b77fc48e94ea tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Updating instance_info_cache with network_info: [{"id": "5a010977-7d05-47e0-8688-bbf9a0e70900", "address": "fa:16:3e:1e:14:06", "network": {"id": "aa1c09e5-8d72-43ad-a903-a0f5e711da80", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1643616572-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.251", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c987c5f844be42c99cc9d57a9ca84c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9bc2632-36f9-4912-8782-8bbb789f909d", "external-id": "nsx-vlan-transportzone-897", "segmentation_id": 897, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5a010977-7d", "ovs_interfaceid": "5a010977-7d05-47e0-8688-bbf9a0e70900", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2076.659869] env[62824]: DEBUG nova.network.neutron [req-c4eff0be-169f-4025-a8d0-b01f2eab4019 req-8034a27a-a152-44ba-b39a-4d40bd999a6e service nova] [instance: 15479062-af75-4925-99b3-77d6a49751ad] Updated VIF entry in instance network info cache for port 6b8b3a2d-f55e-4aa8-b995-4880c1adea60. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2076.660194] env[62824]: DEBUG nova.network.neutron [req-c4eff0be-169f-4025-a8d0-b01f2eab4019 req-8034a27a-a152-44ba-b39a-4d40bd999a6e service nova] [instance: 15479062-af75-4925-99b3-77d6a49751ad] Updating instance_info_cache with network_info: [{"id": "6b8b3a2d-f55e-4aa8-b995-4880c1adea60", "address": "fa:16:3e:77:1e:5b", "network": {"id": "6f888a0c-8089-4dd4-aace-14b020e11bee", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-137042394-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d041a8209e3848eba8e47251e013ca17", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5f60c972-a72d-4c5f-a250-faadfd6eafbe", "external-id": "nsx-vlan-transportzone-932", "segmentation_id": 932, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6b8b3a2d-f5", "ovs_interfaceid": "6b8b3a2d-f55e-4aa8-b995-4880c1adea60", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2076.697177] env[62824]: DEBUG oslo_vmware.api [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146230, 'name': ReconfigVM_Task, 'duration_secs': 0.618881} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2076.697533] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 919e8a45-7810-4a8d-a1aa-5046b5ab059c] Reconfigured VM instance instance-0000003b to attach disk [datastore2] 919e8a45-7810-4a8d-a1aa-5046b5ab059c/919e8a45-7810-4a8d-a1aa-5046b5ab059c.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2076.698673] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e3f8cdda-bace-4405-a73d-c7098053c0be {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2076.712696] env[62824]: DEBUG oslo_vmware.api [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 2076.712696] env[62824]: value = "task-2146235" [ 2076.712696] env[62824]: _type = "Task" [ 2076.712696] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2076.724186] env[62824]: DEBUG oslo_vmware.api [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146235, 'name': Rename_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2076.725480] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 29b8dd5f-1855-490c-a01b-54840073a753 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2076.725694] env[62824]: WARNING nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 82382932-7302-4441-a6f8-9aa2300ec0f6 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 2076.726010] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance f4d63a93-23af-470c-b36a-662af81dc386 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2076.726191] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2076.726359] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 07ce931f-45ef-409b-b714-9f1cd47a3a88 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2076.726983] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance f3740d9a-b21a-4a79-9e28-2a89ecd08bb6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2076.726983] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance c36ece43-3d70-4e67-a740-9057f413c722 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2076.726983] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance b3003c4b-ae5a-48df-8c12-a915a76253f4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2076.727203] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 6d80ec06-8559-4964-8577-a2512aa366ed actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2076.727251] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 59b5b883-4188-471c-8862-444f3ce08cb0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2076.727326] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance bbf1440b-8681-48c8-a178-9a83b925c695 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2076.727459] env[62824]: WARNING nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 7febb8b4-f984-4d79-a888-b2829f2a9df6 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 2076.728525] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2076.728525] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 8bb90a4d-93c9-4f54-b15e-48fe966e6c97 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2076.728525] env[62824]: WARNING nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 2afecec5-763d-4616-b690-41d3101cfc47 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 2076.728525] env[62824]: WARNING nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 35f0435c-fe46-46ab-bc71-c21bb0ad3d55 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 2076.728525] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance b9f30f87-3594-4468-9d29-70890d8761e3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2076.728713] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 02405b17-7022-4e2c-a357-415de88f63d0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2076.728713] env[62824]: WARNING nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance c65ccd20-b682-420a-9c1a-47a45959197e is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 2076.728894] env[62824]: WARNING nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 4423d172-acdd-4c69-b3b8-ff166e1b8548 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 2076.728958] env[62824]: WARNING nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 88bad2dd-dce7-41df-b56c-93a5d054c11e is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 2076.729083] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 6a724332-a165-4a2b-9dd8-85f27e7b7637 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2076.729212] env[62824]: WARNING nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 21c05725-3dad-4965-98d0-0622ebcc8ece is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 2076.729356] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance ed7d3e95-f26e-40fd-a370-e17922bbff8e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2076.729617] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance b0456b8a-348a-4503-a92c-58e5ab455d1c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2076.729617] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance bbfcb2e3-9326-4548-b15b-e054cbfd192e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2076.729723] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 2f315f9e-94e4-47ef-9503-ee92b59e5452 actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2076.729797] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 919e8a45-7810-4a8d-a1aa-5046b5ab059c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2076.729904] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 15479062-af75-4925-99b3-77d6a49751ad actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 2, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2076.730160] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance a6cd2032-de60-4f78-bf1e-79801d049df0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2076.730160] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 02f34565-6675-4c79-ac47-b131ceba9df8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2076.805881] env[62824]: DEBUG oslo_vmware.api [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Task: {'id': task-2146234, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2076.832696] env[62824]: DEBUG oslo_vmware.api [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52ba44c1-06c2-32d1-bb8c-744a92ce0e62, 'name': SearchDatastore_Task, 'duration_secs': 0.010668} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2076.834275] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0850e148-62c0-4b79-bfd0-39feb3fe7d42 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2076.844753] env[62824]: DEBUG oslo_vmware.api [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2076.844753] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5276888c-6094-c245-8429-d4b1b7d86ba1" [ 2076.844753] env[62824]: _type = "Task" [ 2076.844753] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2076.856322] env[62824]: DEBUG oslo_vmware.api [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5276888c-6094-c245-8429-d4b1b7d86ba1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2076.862634] env[62824]: DEBUG nova.compute.manager [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 350cfd12-a8d5-4a5f-b3de-d795fa179dfd] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2076.928513] env[62824]: DEBUG nova.network.neutron [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02f34565-6675-4c79-ac47-b131ceba9df8] Successfully updated port: be36fef6-17a7-4c25-bf33-98cc6a037b53 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2076.999364] env[62824]: INFO nova.compute.manager [-] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Took 2.27 seconds to deallocate network for instance. [ 2077.092808] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c94e74ba-f9b2-469b-be3d-b77fc48e94ea tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Releasing lock "refresh_cache-f4d63a93-23af-470c-b36a-662af81dc386" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2077.163056] env[62824]: DEBUG oslo_concurrency.lockutils [req-c4eff0be-169f-4025-a8d0-b01f2eab4019 req-8034a27a-a152-44ba-b39a-4d40bd999a6e service nova] Releasing lock "refresh_cache-15479062-af75-4925-99b3-77d6a49751ad" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2077.225078] env[62824]: DEBUG oslo_vmware.api [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146235, 'name': Rename_Task, 'duration_secs': 0.332636} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2077.225078] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 919e8a45-7810-4a8d-a1aa-5046b5ab059c] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2077.225078] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-584e1455-6631-492d-8a7b-7cb2560ebb4e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2077.236370] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 2077.237673] env[62824]: DEBUG oslo_vmware.api [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 2077.237673] env[62824]: value = "task-2146236" [ 2077.237673] env[62824]: _type = "Task" [ 2077.237673] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2077.238960] env[62824]: DEBUG nova.compute.manager [req-1d0f0919-8440-406f-9a64-216cf7a1cf5d req-d57e62e6-a6cf-4d08-a872-fa236ed0a943 service nova] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Received event network-changed-d85a0acc-fed6-4797-9f81-2aafa5bb6967 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2077.239846] env[62824]: DEBUG nova.compute.manager [req-1d0f0919-8440-406f-9a64-216cf7a1cf5d req-d57e62e6-a6cf-4d08-a872-fa236ed0a943 service nova] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Refreshing instance network info cache due to event network-changed-d85a0acc-fed6-4797-9f81-2aafa5bb6967. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2077.239846] env[62824]: DEBUG oslo_concurrency.lockutils [req-1d0f0919-8440-406f-9a64-216cf7a1cf5d req-d57e62e6-a6cf-4d08-a872-fa236ed0a943 service nova] Acquiring lock "refresh_cache-bbfcb2e3-9326-4548-b15b-e054cbfd192e" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2077.239846] env[62824]: DEBUG oslo_concurrency.lockutils [req-1d0f0919-8440-406f-9a64-216cf7a1cf5d req-d57e62e6-a6cf-4d08-a872-fa236ed0a943 service nova] Acquired lock "refresh_cache-bbfcb2e3-9326-4548-b15b-e054cbfd192e" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2077.239846] env[62824]: DEBUG nova.network.neutron [req-1d0f0919-8440-406f-9a64-216cf7a1cf5d req-d57e62e6-a6cf-4d08-a872-fa236ed0a943 service nova] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Refreshing network info cache for port d85a0acc-fed6-4797-9f81-2aafa5bb6967 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2077.251959] env[62824]: DEBUG oslo_vmware.api [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146236, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2077.305204] env[62824]: DEBUG oslo_vmware.api [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Task: {'id': task-2146234, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.754098} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2077.305527] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 15479062-af75-4925-99b3-77d6a49751ad/15479062-af75-4925-99b3-77d6a49751ad.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2077.306310] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 15479062-af75-4925-99b3-77d6a49751ad] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2077.306310] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-810c6805-1423-4a76-8c03-ff21fce8cd66 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2077.315842] env[62824]: DEBUG oslo_vmware.api [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Waiting for the task: (returnval){ [ 2077.315842] env[62824]: value = "task-2146237" [ 2077.315842] env[62824]: _type = "Task" [ 2077.315842] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2077.329502] env[62824]: DEBUG oslo_vmware.api [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Task: {'id': task-2146237, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2077.359493] env[62824]: DEBUG oslo_vmware.api [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5276888c-6094-c245-8429-d4b1b7d86ba1, 'name': SearchDatastore_Task, 'duration_secs': 0.025275} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2077.360008] env[62824]: DEBUG oslo_concurrency.lockutils [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2077.360147] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] a6cd2032-de60-4f78-bf1e-79801d049df0/a6cd2032-de60-4f78-bf1e-79801d049df0.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2077.360573] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1d628ff0-5208-4f91-adc5-a48c0c0d6166 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2077.366594] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3aa78573-24a8-42b5-8896-df119ccd8889 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquiring lock "f4d63a93-23af-470c-b36a-662af81dc386" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2077.366926] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3aa78573-24a8-42b5-8896-df119ccd8889 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lock "f4d63a93-23af-470c-b36a-662af81dc386" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2077.367130] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3aa78573-24a8-42b5-8896-df119ccd8889 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquiring lock "f4d63a93-23af-470c-b36a-662af81dc386-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2077.367349] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3aa78573-24a8-42b5-8896-df119ccd8889 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lock "f4d63a93-23af-470c-b36a-662af81dc386-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2077.367512] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3aa78573-24a8-42b5-8896-df119ccd8889 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lock "f4d63a93-23af-470c-b36a-662af81dc386-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2077.372772] env[62824]: DEBUG oslo_vmware.api [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2077.372772] env[62824]: value = "task-2146238" [ 2077.372772] env[62824]: _type = "Task" [ 2077.372772] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2077.373257] env[62824]: INFO nova.compute.manager [None req-3aa78573-24a8-42b5-8896-df119ccd8889 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Terminating instance [ 2077.389662] env[62824]: DEBUG oslo_vmware.api [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146238, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2077.399222] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2077.434623] env[62824]: DEBUG oslo_concurrency.lockutils [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Acquiring lock "refresh_cache-02f34565-6675-4c79-ac47-b131ceba9df8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2077.434669] env[62824]: DEBUG oslo_concurrency.lockutils [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Acquired lock "refresh_cache-02f34565-6675-4c79-ac47-b131ceba9df8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2077.434940] env[62824]: DEBUG nova.network.neutron [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02f34565-6675-4c79-ac47-b131ceba9df8] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2077.487431] env[62824]: DEBUG oslo_concurrency.lockutils [None req-84ba9f19-2925-48b1-85d5-00fe075f71bc tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Acquiring lock "ed7d3e95-f26e-40fd-a370-e17922bbff8e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2077.487689] env[62824]: DEBUG oslo_concurrency.lockutils [None req-84ba9f19-2925-48b1-85d5-00fe075f71bc tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Lock "ed7d3e95-f26e-40fd-a370-e17922bbff8e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2077.487895] env[62824]: DEBUG oslo_concurrency.lockutils [None req-84ba9f19-2925-48b1-85d5-00fe075f71bc tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Acquiring lock "ed7d3e95-f26e-40fd-a370-e17922bbff8e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2077.488091] env[62824]: DEBUG oslo_concurrency.lockutils [None req-84ba9f19-2925-48b1-85d5-00fe075f71bc tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Lock "ed7d3e95-f26e-40fd-a370-e17922bbff8e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2077.488262] env[62824]: DEBUG oslo_concurrency.lockutils [None req-84ba9f19-2925-48b1-85d5-00fe075f71bc tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Lock "ed7d3e95-f26e-40fd-a370-e17922bbff8e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2077.491289] env[62824]: INFO nova.compute.manager [None req-84ba9f19-2925-48b1-85d5-00fe075f71bc tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] [instance: ed7d3e95-f26e-40fd-a370-e17922bbff8e] Terminating instance [ 2077.512030] env[62824]: DEBUG oslo_concurrency.lockutils [None req-423ea96c-8206-47c1-8b7e-a945775631af tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2077.602015] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c94e74ba-f9b2-469b-be3d-b77fc48e94ea tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lock "interface-f4d63a93-23af-470c-b36a-662af81dc386-c34f030d-46da-4a7a-9703-5acb2e526d50" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.313s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2077.698208] env[62824]: DEBUG oslo_concurrency.lockutils [None req-123477fc-8c3e-4f2a-83de-cced4f011020 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Acquiring lock "b8cc8cd3-ea03-40bf-b867-7ad193365552" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2077.743509] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 350cfd12-a8d5-4a5f-b3de-d795fa179dfd has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 2077.743890] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Total usable vcpus: 48, total allocated vcpus: 23 {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 2077.744057] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=4928MB phys_disk=200GB used_disk=23GB total_vcpus=48 used_vcpus=23 pci_stats=[] {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 2077.758909] env[62824]: DEBUG oslo_vmware.api [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146236, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2077.834179] env[62824]: DEBUG oslo_vmware.api [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Task: {'id': task-2146237, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.305495} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2077.834483] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 15479062-af75-4925-99b3-77d6a49751ad] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2077.835369] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fc18d31-a4b7-4975-a631-a23e5fad5b4b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2077.859605] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 15479062-af75-4925-99b3-77d6a49751ad] Reconfiguring VM instance instance-0000003c to attach disk [datastore2] 15479062-af75-4925-99b3-77d6a49751ad/15479062-af75-4925-99b3-77d6a49751ad.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2077.865070] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b54c38d6-54e5-4750-b4e1-cf8debe5c416 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2077.885599] env[62824]: DEBUG nova.compute.manager [None req-3aa78573-24a8-42b5-8896-df119ccd8889 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2077.885889] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-3aa78573-24a8-42b5-8896-df119ccd8889 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2077.891264] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-299c91d3-f9d4-4509-97e4-18f4f5583d70 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2077.901611] env[62824]: DEBUG oslo_vmware.api [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146238, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2077.906939] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-3aa78573-24a8-42b5-8896-df119ccd8889 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2077.907384] env[62824]: DEBUG oslo_vmware.api [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Waiting for the task: (returnval){ [ 2077.907384] env[62824]: value = "task-2146239" [ 2077.907384] env[62824]: _type = "Task" [ 2077.907384] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2077.912336] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f4fd8dbb-268c-43db-b7a2-9501dba881d9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2077.924103] env[62824]: DEBUG oslo_vmware.api [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Task: {'id': task-2146239, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2077.925904] env[62824]: DEBUG oslo_vmware.api [None req-3aa78573-24a8-42b5-8896-df119ccd8889 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Waiting for the task: (returnval){ [ 2077.925904] env[62824]: value = "task-2146240" [ 2077.925904] env[62824]: _type = "Task" [ 2077.925904] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2077.939124] env[62824]: DEBUG oslo_vmware.api [None req-3aa78573-24a8-42b5-8896-df119ccd8889 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146240, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2077.997743] env[62824]: DEBUG oslo_concurrency.lockutils [None req-84ba9f19-2925-48b1-85d5-00fe075f71bc tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Acquiring lock "refresh_cache-ed7d3e95-f26e-40fd-a370-e17922bbff8e" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2077.998068] env[62824]: DEBUG oslo_concurrency.lockutils [None req-84ba9f19-2925-48b1-85d5-00fe075f71bc tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Acquired lock "refresh_cache-ed7d3e95-f26e-40fd-a370-e17922bbff8e" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2077.998068] env[62824]: DEBUG nova.network.neutron [None req-84ba9f19-2925-48b1-85d5-00fe075f71bc tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] [instance: ed7d3e95-f26e-40fd-a370-e17922bbff8e] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2078.042801] env[62824]: DEBUG nova.network.neutron [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02f34565-6675-4c79-ac47-b131ceba9df8] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2078.259275] env[62824]: DEBUG oslo_vmware.api [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146236, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2078.300115] env[62824]: DEBUG nova.network.neutron [req-1d0f0919-8440-406f-9a64-216cf7a1cf5d req-d57e62e6-a6cf-4d08-a872-fa236ed0a943 service nova] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Updated VIF entry in instance network info cache for port d85a0acc-fed6-4797-9f81-2aafa5bb6967. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2078.300115] env[62824]: DEBUG nova.network.neutron [req-1d0f0919-8440-406f-9a64-216cf7a1cf5d req-d57e62e6-a6cf-4d08-a872-fa236ed0a943 service nova] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Updating instance_info_cache with network_info: [{"id": "d85a0acc-fed6-4797-9f81-2aafa5bb6967", "address": "fa:16:3e:d8:c8:ab", "network": {"id": "a410f4f3-d072-4034-b797-8210e1a79093", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1482574391-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.211", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e10208040df34eb5adfe8dcbc76043d5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f4399275-8e92-4448-be9e-d4984e93e89c", "external-id": "nsx-vlan-transportzone-192", "segmentation_id": 192, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd85a0acc-fe", "ovs_interfaceid": "d85a0acc-fed6-4797-9f81-2aafa5bb6967", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2078.406386] env[62824]: DEBUG oslo_vmware.api [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146238, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.724834} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2078.407536] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] a6cd2032-de60-4f78-bf1e-79801d049df0/a6cd2032-de60-4f78-bf1e-79801d049df0.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2078.407536] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: a6cd2032-de60-4f78-bf1e-79801d049df0] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2078.407756] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-27c7f575-7495-4aab-828d-f04e81875fe0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2078.427262] env[62824]: DEBUG oslo_vmware.api [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2078.427262] env[62824]: value = "task-2146241" [ 2078.427262] env[62824]: _type = "Task" [ 2078.427262] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2078.438314] env[62824]: DEBUG oslo_vmware.api [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Task: {'id': task-2146239, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2078.446957] env[62824]: DEBUG oslo_vmware.api [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146241, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2078.450636] env[62824]: DEBUG oslo_vmware.api [None req-3aa78573-24a8-42b5-8896-df119ccd8889 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146240, 'name': PowerOffVM_Task, 'duration_secs': 0.355641} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2078.453957] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-3aa78573-24a8-42b5-8896-df119ccd8889 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2078.453957] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-3aa78573-24a8-42b5-8896-df119ccd8889 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2078.454552] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-174956fc-c6f3-4fce-98f8-a5b53057c0ea {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2078.459399] env[62824]: DEBUG nova.network.neutron [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02f34565-6675-4c79-ac47-b131ceba9df8] Updating instance_info_cache with network_info: [{"id": "be36fef6-17a7-4c25-bf33-98cc6a037b53", "address": "fa:16:3e:c1:60:d0", "network": {"id": "fc381ca5-198f-4afc-9ebf-bc6e3de68c35", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-425455451-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "248af2bc3e7b41d2a74b0c6074b359e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4c9a12d2-469f-4199-bfaa-f791d765deac", "external-id": "nsx-vlan-transportzone-96", "segmentation_id": 96, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbe36fef6-17", "ovs_interfaceid": "be36fef6-17a7-4c25-bf33-98cc6a037b53", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2078.505160] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e98f5d6f-76f0-4c21-8014-96f97db9ecff {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2078.516721] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66b9b2e4-57d1-4f9d-a81a-adcd160c0533 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2078.558665] env[62824]: DEBUG nova.network.neutron [None req-84ba9f19-2925-48b1-85d5-00fe075f71bc tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] [instance: ed7d3e95-f26e-40fd-a370-e17922bbff8e] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2078.561579] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c42530ee-7345-4a5e-89a0-2c799c1fe289 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2078.572431] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22ec00bf-6a46-4630-b014-59aeb4a99653 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2078.591250] env[62824]: DEBUG nova.compute.provider_tree [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2078.662669] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-3aa78573-24a8-42b5-8896-df119ccd8889 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2078.662985] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-3aa78573-24a8-42b5-8896-df119ccd8889 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Deleting contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2078.663247] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-3aa78573-24a8-42b5-8896-df119ccd8889 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Deleting the datastore file [datastore2] f4d63a93-23af-470c-b36a-662af81dc386 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2078.663572] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7560e411-8f9d-47a7-bc25-27a662ed8008 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2078.673311] env[62824]: DEBUG oslo_vmware.api [None req-3aa78573-24a8-42b5-8896-df119ccd8889 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Waiting for the task: (returnval){ [ 2078.673311] env[62824]: value = "task-2146243" [ 2078.673311] env[62824]: _type = "Task" [ 2078.673311] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2078.687872] env[62824]: DEBUG oslo_vmware.api [None req-3aa78573-24a8-42b5-8896-df119ccd8889 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146243, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2078.734298] env[62824]: DEBUG nova.network.neutron [None req-84ba9f19-2925-48b1-85d5-00fe075f71bc tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] [instance: ed7d3e95-f26e-40fd-a370-e17922bbff8e] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2078.758454] env[62824]: DEBUG oslo_vmware.api [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146236, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2078.806692] env[62824]: DEBUG oslo_concurrency.lockutils [req-1d0f0919-8440-406f-9a64-216cf7a1cf5d req-d57e62e6-a6cf-4d08-a872-fa236ed0a943 service nova] Releasing lock "refresh_cache-bbfcb2e3-9326-4548-b15b-e054cbfd192e" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2078.807090] env[62824]: DEBUG nova.compute.manager [req-1d0f0919-8440-406f-9a64-216cf7a1cf5d req-d57e62e6-a6cf-4d08-a872-fa236ed0a943 service nova] [instance: a6cd2032-de60-4f78-bf1e-79801d049df0] Received event network-vif-plugged-018929f6-4f6c-4148-b1eb-9cf11f5cc746 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2078.807410] env[62824]: DEBUG oslo_concurrency.lockutils [req-1d0f0919-8440-406f-9a64-216cf7a1cf5d req-d57e62e6-a6cf-4d08-a872-fa236ed0a943 service nova] Acquiring lock "a6cd2032-de60-4f78-bf1e-79801d049df0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2078.807755] env[62824]: DEBUG oslo_concurrency.lockutils [req-1d0f0919-8440-406f-9a64-216cf7a1cf5d req-d57e62e6-a6cf-4d08-a872-fa236ed0a943 service nova] Lock "a6cd2032-de60-4f78-bf1e-79801d049df0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2078.808048] env[62824]: DEBUG oslo_concurrency.lockutils [req-1d0f0919-8440-406f-9a64-216cf7a1cf5d req-d57e62e6-a6cf-4d08-a872-fa236ed0a943 service nova] Lock "a6cd2032-de60-4f78-bf1e-79801d049df0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2078.808338] env[62824]: DEBUG nova.compute.manager [req-1d0f0919-8440-406f-9a64-216cf7a1cf5d req-d57e62e6-a6cf-4d08-a872-fa236ed0a943 service nova] [instance: a6cd2032-de60-4f78-bf1e-79801d049df0] No waiting events found dispatching network-vif-plugged-018929f6-4f6c-4148-b1eb-9cf11f5cc746 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2078.808621] env[62824]: WARNING nova.compute.manager [req-1d0f0919-8440-406f-9a64-216cf7a1cf5d req-d57e62e6-a6cf-4d08-a872-fa236ed0a943 service nova] [instance: a6cd2032-de60-4f78-bf1e-79801d049df0] Received unexpected event network-vif-plugged-018929f6-4f6c-4148-b1eb-9cf11f5cc746 for instance with vm_state building and task_state spawning. [ 2078.808895] env[62824]: DEBUG nova.compute.manager [req-1d0f0919-8440-406f-9a64-216cf7a1cf5d req-d57e62e6-a6cf-4d08-a872-fa236ed0a943 service nova] [instance: a6cd2032-de60-4f78-bf1e-79801d049df0] Received event network-changed-018929f6-4f6c-4148-b1eb-9cf11f5cc746 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2078.809177] env[62824]: DEBUG nova.compute.manager [req-1d0f0919-8440-406f-9a64-216cf7a1cf5d req-d57e62e6-a6cf-4d08-a872-fa236ed0a943 service nova] [instance: a6cd2032-de60-4f78-bf1e-79801d049df0] Refreshing instance network info cache due to event network-changed-018929f6-4f6c-4148-b1eb-9cf11f5cc746. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2078.809496] env[62824]: DEBUG oslo_concurrency.lockutils [req-1d0f0919-8440-406f-9a64-216cf7a1cf5d req-d57e62e6-a6cf-4d08-a872-fa236ed0a943 service nova] Acquiring lock "refresh_cache-a6cd2032-de60-4f78-bf1e-79801d049df0" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2078.809754] env[62824]: DEBUG oslo_concurrency.lockutils [req-1d0f0919-8440-406f-9a64-216cf7a1cf5d req-d57e62e6-a6cf-4d08-a872-fa236ed0a943 service nova] Acquired lock "refresh_cache-a6cd2032-de60-4f78-bf1e-79801d049df0" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2078.810125] env[62824]: DEBUG nova.network.neutron [req-1d0f0919-8440-406f-9a64-216cf7a1cf5d req-d57e62e6-a6cf-4d08-a872-fa236ed0a943 service nova] [instance: a6cd2032-de60-4f78-bf1e-79801d049df0] Refreshing network info cache for port 018929f6-4f6c-4148-b1eb-9cf11f5cc746 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2078.908275] env[62824]: DEBUG nova.compute.manager [req-29fc9032-dc44-4610-bf70-5c0fa8bc1d30 req-e0006438-9e3a-4086-b480-7456ca61fbe3 service nova] [instance: 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2] Received event network-vif-deleted-b90aaedc-aa4e-4f70-9365-1ee7a5bd49ee {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2078.926255] env[62824]: DEBUG oslo_vmware.api [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Task: {'id': task-2146239, 'name': ReconfigVM_Task, 'duration_secs': 0.743439} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2078.926851] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 15479062-af75-4925-99b3-77d6a49751ad] Reconfigured VM instance instance-0000003c to attach disk [datastore2] 15479062-af75-4925-99b3-77d6a49751ad/15479062-af75-4925-99b3-77d6a49751ad.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2078.927224] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Creating Virtual Disk of size 1048576 KB and adapter type paraVirtual on the data store {{(pid=62824) create_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1383}} [ 2078.928677] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CreateVirtualDisk_Task with opID=oslo.vmware-87322085-2e6c-4f27-9a7e-5c42b35536d8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2078.942707] env[62824]: DEBUG oslo_vmware.api [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146241, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.181058} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2078.944657] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: a6cd2032-de60-4f78-bf1e-79801d049df0] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2078.945305] env[62824]: DEBUG oslo_vmware.api [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Waiting for the task: (returnval){ [ 2078.945305] env[62824]: value = "task-2146244" [ 2078.945305] env[62824]: _type = "Task" [ 2078.945305] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2078.946494] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0c326b6-7b75-402d-8bcd-9076b0462cf5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2078.978124] env[62824]: DEBUG oslo_concurrency.lockutils [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Releasing lock "refresh_cache-02f34565-6675-4c79-ac47-b131ceba9df8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2078.978745] env[62824]: DEBUG nova.compute.manager [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02f34565-6675-4c79-ac47-b131ceba9df8] Instance network_info: |[{"id": "be36fef6-17a7-4c25-bf33-98cc6a037b53", "address": "fa:16:3e:c1:60:d0", "network": {"id": "fc381ca5-198f-4afc-9ebf-bc6e3de68c35", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-425455451-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "248af2bc3e7b41d2a74b0c6074b359e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4c9a12d2-469f-4199-bfaa-f791d765deac", "external-id": "nsx-vlan-transportzone-96", "segmentation_id": 96, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbe36fef6-17", "ovs_interfaceid": "be36fef6-17a7-4c25-bf33-98cc6a037b53", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2078.990575] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: a6cd2032-de60-4f78-bf1e-79801d049df0] Reconfiguring VM instance instance-0000003d to attach disk [datastore1] a6cd2032-de60-4f78-bf1e-79801d049df0/a6cd2032-de60-4f78-bf1e-79801d049df0.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2078.994600] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02f34565-6675-4c79-ac47-b131ceba9df8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c1:60:d0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4c9a12d2-469f-4199-bfaa-f791d765deac', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'be36fef6-17a7-4c25-bf33-98cc6a037b53', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2079.006791] env[62824]: DEBUG oslo.service.loopingcall [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2079.007153] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-abdf195e-9939-4780-a729-0f6df05ff448 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2079.031268] env[62824]: DEBUG oslo_vmware.api [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Task: {'id': task-2146244, 'name': CreateVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2079.031613] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 02f34565-6675-4c79-ac47-b131ceba9df8] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2079.032151] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1fe278ac-6b2d-40b7-9996-31dc91977e26 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2079.063805] env[62824]: DEBUG oslo_concurrency.lockutils [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Acquiring lock "363177c0-dc40-429a-a74b-e690da133edb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2079.065144] env[62824]: DEBUG oslo_concurrency.lockutils [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Lock "363177c0-dc40-429a-a74b-e690da133edb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2079.065848] env[62824]: DEBUG oslo_vmware.api [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2079.065848] env[62824]: value = "task-2146245" [ 2079.065848] env[62824]: _type = "Task" [ 2079.065848] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2079.073151] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2079.073151] env[62824]: value = "task-2146246" [ 2079.073151] env[62824]: _type = "Task" [ 2079.073151] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2079.083963] env[62824]: DEBUG oslo_vmware.api [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146245, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2079.092960] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146246, 'name': CreateVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2079.095194] env[62824]: DEBUG nova.scheduler.client.report [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2079.186801] env[62824]: DEBUG oslo_vmware.api [None req-3aa78573-24a8-42b5-8896-df119ccd8889 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146243, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2079.237217] env[62824]: DEBUG oslo_concurrency.lockutils [None req-84ba9f19-2925-48b1-85d5-00fe075f71bc tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Releasing lock "refresh_cache-ed7d3e95-f26e-40fd-a370-e17922bbff8e" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2079.237917] env[62824]: DEBUG nova.compute.manager [None req-84ba9f19-2925-48b1-85d5-00fe075f71bc tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] [instance: ed7d3e95-f26e-40fd-a370-e17922bbff8e] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2079.237917] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-84ba9f19-2925-48b1-85d5-00fe075f71bc tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] [instance: ed7d3e95-f26e-40fd-a370-e17922bbff8e] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2079.238843] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff46dfcf-2244-44b8-bc5d-7a9cc6cdc67f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2079.248409] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-84ba9f19-2925-48b1-85d5-00fe075f71bc tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] [instance: ed7d3e95-f26e-40fd-a370-e17922bbff8e] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2079.252051] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2b3593a7-077b-4430-bfd5-8b4eaa373a7a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2079.261093] env[62824]: DEBUG oslo_vmware.api [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146236, 'name': PowerOnVM_Task, 'duration_secs': 1.625274} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2079.262472] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 919e8a45-7810-4a8d-a1aa-5046b5ab059c] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2079.262686] env[62824]: INFO nova.compute.manager [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 919e8a45-7810-4a8d-a1aa-5046b5ab059c] Took 12.30 seconds to spawn the instance on the hypervisor. [ 2079.262867] env[62824]: DEBUG nova.compute.manager [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 919e8a45-7810-4a8d-a1aa-5046b5ab059c] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2079.263224] env[62824]: DEBUG oslo_vmware.api [None req-84ba9f19-2925-48b1-85d5-00fe075f71bc tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Waiting for the task: (returnval){ [ 2079.263224] env[62824]: value = "task-2146247" [ 2079.263224] env[62824]: _type = "Task" [ 2079.263224] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2079.264039] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66f80c5b-0b0f-48dc-bd9e-6cd400f50448 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2079.276277] env[62824]: DEBUG oslo_vmware.api [None req-84ba9f19-2925-48b1-85d5-00fe075f71bc tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Task: {'id': task-2146247, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2079.463606] env[62824]: DEBUG oslo_vmware.api [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Task: {'id': task-2146244, 'name': CreateVirtualDisk_Task, 'duration_secs': 0.085525} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2079.464026] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Created Virtual Disk of size 1048576 KB and type thin {{(pid=62824) create_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1404}} [ 2079.464757] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e98d6b8d-fdae-43b0-a9f8-56daf29bdf96 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2079.517831] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 15479062-af75-4925-99b3-77d6a49751ad] Reconfiguring VM instance instance-0000003c to attach disk [datastore2] 15479062-af75-4925-99b3-77d6a49751ad/ephemeral_0.vmdk or device None with type thin {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2079.521858] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-716a35ed-3af3-4e44-9f2a-d3f574506450 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2079.545750] env[62824]: DEBUG oslo_vmware.api [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Waiting for the task: (returnval){ [ 2079.545750] env[62824]: value = "task-2146248" [ 2079.545750] env[62824]: _type = "Task" [ 2079.545750] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2079.565181] env[62824]: DEBUG oslo_vmware.api [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Task: {'id': task-2146248, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2079.566814] env[62824]: DEBUG nova.compute.manager [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2079.595941] env[62824]: DEBUG oslo_vmware.api [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146245, 'name': ReconfigVM_Task, 'duration_secs': 0.367131} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2079.599898] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: a6cd2032-de60-4f78-bf1e-79801d049df0] Reconfigured VM instance instance-0000003d to attach disk [datastore1] a6cd2032-de60-4f78-bf1e-79801d049df0/a6cd2032-de60-4f78-bf1e-79801d049df0.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2079.600552] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146246, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2079.603581] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62824) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 2079.603581] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.962s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2079.603581] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-411cb58c-d0d7-4c59-9a88-1a068b16f81c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2079.603971] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bf6ca20b-2439-4b4d-ba93-061898ce061b tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.687s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2079.604191] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bf6ca20b-2439-4b4d-ba93-061898ce061b tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2079.607021] env[62824]: DEBUG oslo_concurrency.lockutils [None req-02170dc2-e93c-4f62-a015-5cb11263272d tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.632s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2079.607021] env[62824]: DEBUG oslo_concurrency.lockutils [None req-02170dc2-e93c-4f62-a015-5cb11263272d tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2079.609120] env[62824]: DEBUG oslo_concurrency.lockutils [None req-24b1fd6d-11b0-414f-a359-57219816e958 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.081s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2079.609358] env[62824]: DEBUG oslo_concurrency.lockutils [None req-24b1fd6d-11b0-414f-a359-57219816e958 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2079.611126] env[62824]: DEBUG oslo_concurrency.lockutils [None req-abb349ec-c188-4834-ae78-a5dd5c6e4af7 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.282s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2079.611311] env[62824]: DEBUG oslo_concurrency.lockutils [None req-abb349ec-c188-4834-ae78-a5dd5c6e4af7 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2079.613638] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 20.526s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2079.613830] env[62824]: DEBUG nova.objects.instance [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 21c05725-3dad-4965-98d0-0622ebcc8ece] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62824) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 2079.620395] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2079.620395] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Cleaning up deleted instances {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11745}} [ 2079.628660] env[62824]: DEBUG oslo_vmware.api [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2079.628660] env[62824]: value = "task-2146249" [ 2079.628660] env[62824]: _type = "Task" [ 2079.628660] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2079.641162] env[62824]: DEBUG oslo_vmware.api [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146249, 'name': Rename_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2079.658563] env[62824]: INFO nova.scheduler.client.report [None req-02170dc2-e93c-4f62-a015-5cb11263272d tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Deleted allocations for instance c65ccd20-b682-420a-9c1a-47a45959197e [ 2079.667266] env[62824]: INFO nova.scheduler.client.report [None req-24b1fd6d-11b0-414f-a359-57219816e958 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Deleted allocations for instance 7febb8b4-f984-4d79-a888-b2829f2a9df6 [ 2079.689948] env[62824]: DEBUG oslo_vmware.api [None req-3aa78573-24a8-42b5-8896-df119ccd8889 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146243, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.606175} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2079.691978] env[62824]: INFO nova.scheduler.client.report [None req-abb349ec-c188-4834-ae78-a5dd5c6e4af7 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Deleted allocations for instance 4423d172-acdd-4c69-b3b8-ff166e1b8548 [ 2079.693242] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-3aa78573-24a8-42b5-8896-df119ccd8889 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2079.694200] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-3aa78573-24a8-42b5-8896-df119ccd8889 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Deleted contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2079.694200] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-3aa78573-24a8-42b5-8896-df119ccd8889 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2079.694200] env[62824]: INFO nova.compute.manager [None req-3aa78573-24a8-42b5-8896-df119ccd8889 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Took 1.81 seconds to destroy the instance on the hypervisor. [ 2079.694434] env[62824]: DEBUG oslo.service.loopingcall [None req-3aa78573-24a8-42b5-8896-df119ccd8889 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2079.695372] env[62824]: INFO nova.scheduler.client.report [None req-bf6ca20b-2439-4b4d-ba93-061898ce061b tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Deleted allocations for instance 35f0435c-fe46-46ab-bc71-c21bb0ad3d55 [ 2079.696374] env[62824]: DEBUG nova.compute.manager [-] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2079.696478] env[62824]: DEBUG nova.network.neutron [-] [instance: f4d63a93-23af-470c-b36a-662af81dc386] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2079.780760] env[62824]: DEBUG oslo_vmware.api [None req-84ba9f19-2925-48b1-85d5-00fe075f71bc tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Task: {'id': task-2146247, 'name': PowerOffVM_Task, 'duration_secs': 0.275929} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2079.781087] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-84ba9f19-2925-48b1-85d5-00fe075f71bc tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] [instance: ed7d3e95-f26e-40fd-a370-e17922bbff8e] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2079.781207] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-84ba9f19-2925-48b1-85d5-00fe075f71bc tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] [instance: ed7d3e95-f26e-40fd-a370-e17922bbff8e] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2079.781472] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-59bdf711-d588-4e53-ba2b-e5c86285b888 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2079.795305] env[62824]: INFO nova.compute.manager [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 919e8a45-7810-4a8d-a1aa-5046b5ab059c] Took 41.93 seconds to build instance. [ 2079.824629] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-84ba9f19-2925-48b1-85d5-00fe075f71bc tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] [instance: ed7d3e95-f26e-40fd-a370-e17922bbff8e] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2079.824884] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-84ba9f19-2925-48b1-85d5-00fe075f71bc tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] [instance: ed7d3e95-f26e-40fd-a370-e17922bbff8e] Deleting contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2079.825112] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-84ba9f19-2925-48b1-85d5-00fe075f71bc tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Deleting the datastore file [datastore2] ed7d3e95-f26e-40fd-a370-e17922bbff8e {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2079.825397] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-47122495-0705-4f31-98c5-ade3b76b5e93 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2079.837889] env[62824]: DEBUG oslo_vmware.api [None req-84ba9f19-2925-48b1-85d5-00fe075f71bc tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Waiting for the task: (returnval){ [ 2079.837889] env[62824]: value = "task-2146251" [ 2079.837889] env[62824]: _type = "Task" [ 2079.837889] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2079.848707] env[62824]: DEBUG oslo_vmware.api [None req-84ba9f19-2925-48b1-85d5-00fe075f71bc tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Task: {'id': task-2146251, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2080.055840] env[62824]: DEBUG oslo_vmware.api [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Task: {'id': task-2146248, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2080.089930] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146246, 'name': CreateVM_Task, 'duration_secs': 0.691316} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2080.090254] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 02f34565-6675-4c79-ac47-b131ceba9df8] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2080.091106] env[62824]: DEBUG oslo_concurrency.lockutils [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2080.091418] env[62824]: DEBUG oslo_concurrency.lockutils [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2080.091838] env[62824]: DEBUG oslo_concurrency.lockutils [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2080.092283] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aee4fe07-7cd9-4902-930b-09c86715e2a8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2080.096359] env[62824]: DEBUG oslo_concurrency.lockutils [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2080.098477] env[62824]: DEBUG oslo_vmware.api [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for the task: (returnval){ [ 2080.098477] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52dd0c94-7892-2929-7f97-ae1a7ccd3d44" [ 2080.098477] env[62824]: _type = "Task" [ 2080.098477] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2080.112211] env[62824]: DEBUG oslo_vmware.api [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52dd0c94-7892-2929-7f97-ae1a7ccd3d44, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2080.129698] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] There are 19 instances to clean {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11754}} [ 2080.130183] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: c63b05b7-1dbf-4637-8728-33673be0d3ea] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2080.144428] env[62824]: DEBUG oslo_vmware.api [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146249, 'name': Rename_Task, 'duration_secs': 0.261963} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2080.144816] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: a6cd2032-de60-4f78-bf1e-79801d049df0] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2080.145164] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a94a6dbc-f0b4-48a8-b415-4a23f963bc12 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2080.152829] env[62824]: DEBUG oslo_vmware.api [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2080.152829] env[62824]: value = "task-2146252" [ 2080.152829] env[62824]: _type = "Task" [ 2080.152829] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2080.167347] env[62824]: DEBUG oslo_vmware.api [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146252, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2080.171431] env[62824]: DEBUG oslo_concurrency.lockutils [None req-02170dc2-e93c-4f62-a015-5cb11263272d tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Lock "c65ccd20-b682-420a-9c1a-47a45959197e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.501s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2080.184986] env[62824]: DEBUG oslo_concurrency.lockutils [None req-24b1fd6d-11b0-414f-a359-57219816e958 tempest-AttachInterfacesUnderV243Test-59186889 tempest-AttachInterfacesUnderV243Test-59186889-project-member] Lock "7febb8b4-f984-4d79-a888-b2829f2a9df6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.882s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2080.207852] env[62824]: DEBUG oslo_concurrency.lockutils [None req-abb349ec-c188-4834-ae78-a5dd5c6e4af7 tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Lock "4423d172-acdd-4c69-b3b8-ff166e1b8548" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.960s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2080.211265] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bf6ca20b-2439-4b4d-ba93-061898ce061b tempest-ImagesOneServerTestJSON-231149901 tempest-ImagesOneServerTestJSON-231149901-project-member] Lock "35f0435c-fe46-46ab-bc71-c21bb0ad3d55" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.434s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2080.297182] env[62824]: DEBUG nova.network.neutron [req-1d0f0919-8440-406f-9a64-216cf7a1cf5d req-d57e62e6-a6cf-4d08-a872-fa236ed0a943 service nova] [instance: a6cd2032-de60-4f78-bf1e-79801d049df0] Updated VIF entry in instance network info cache for port 018929f6-4f6c-4148-b1eb-9cf11f5cc746. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2080.297622] env[62824]: DEBUG nova.network.neutron [req-1d0f0919-8440-406f-9a64-216cf7a1cf5d req-d57e62e6-a6cf-4d08-a872-fa236ed0a943 service nova] [instance: a6cd2032-de60-4f78-bf1e-79801d049df0] Updating instance_info_cache with network_info: [{"id": "018929f6-4f6c-4148-b1eb-9cf11f5cc746", "address": "fa:16:3e:38:48:15", "network": {"id": "0f512108-670f-4a85-9381-108e18b3b5ed", "bridge": "br-int", "label": "tempest-ServersTestJSON-766397782-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dfff08982dad4790bf4d555e2b4db5e4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f3a2eb5-353f-45c5-a73b-869626f4bb13", "external-id": "nsx-vlan-transportzone-411", "segmentation_id": 411, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap018929f6-4f", "ovs_interfaceid": "018929f6-4f6c-4148-b1eb-9cf11f5cc746", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2080.299458] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5841cc07-af23-4c73-9239-8773239e8024 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "919e8a45-7810-4a8d-a1aa-5046b5ab059c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 76.189s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2080.357424] env[62824]: DEBUG oslo_vmware.api [None req-84ba9f19-2925-48b1-85d5-00fe075f71bc tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Task: {'id': task-2146251, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.135484} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2080.359231] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-84ba9f19-2925-48b1-85d5-00fe075f71bc tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2080.359451] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-84ba9f19-2925-48b1-85d5-00fe075f71bc tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] [instance: ed7d3e95-f26e-40fd-a370-e17922bbff8e] Deleted contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2080.359641] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-84ba9f19-2925-48b1-85d5-00fe075f71bc tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] [instance: ed7d3e95-f26e-40fd-a370-e17922bbff8e] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2080.359816] env[62824]: INFO nova.compute.manager [None req-84ba9f19-2925-48b1-85d5-00fe075f71bc tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] [instance: ed7d3e95-f26e-40fd-a370-e17922bbff8e] Took 1.12 seconds to destroy the instance on the hypervisor. [ 2080.360073] env[62824]: DEBUG oslo.service.loopingcall [None req-84ba9f19-2925-48b1-85d5-00fe075f71bc tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2080.361305] env[62824]: DEBUG nova.compute.manager [-] [instance: ed7d3e95-f26e-40fd-a370-e17922bbff8e] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2080.361402] env[62824]: DEBUG nova.network.neutron [-] [instance: ed7d3e95-f26e-40fd-a370-e17922bbff8e] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2080.364287] env[62824]: DEBUG nova.compute.manager [req-914bea17-6ebc-439c-bbf2-9541bb48d04b req-10bf24fd-83c0-436e-a1f0-7b139b8841ab service nova] [instance: 02f34565-6675-4c79-ac47-b131ceba9df8] Received event network-vif-plugged-be36fef6-17a7-4c25-bf33-98cc6a037b53 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2080.364482] env[62824]: DEBUG oslo_concurrency.lockutils [req-914bea17-6ebc-439c-bbf2-9541bb48d04b req-10bf24fd-83c0-436e-a1f0-7b139b8841ab service nova] Acquiring lock "02f34565-6675-4c79-ac47-b131ceba9df8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2080.364682] env[62824]: DEBUG oslo_concurrency.lockutils [req-914bea17-6ebc-439c-bbf2-9541bb48d04b req-10bf24fd-83c0-436e-a1f0-7b139b8841ab service nova] Lock "02f34565-6675-4c79-ac47-b131ceba9df8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2080.364848] env[62824]: DEBUG oslo_concurrency.lockutils [req-914bea17-6ebc-439c-bbf2-9541bb48d04b req-10bf24fd-83c0-436e-a1f0-7b139b8841ab service nova] Lock "02f34565-6675-4c79-ac47-b131ceba9df8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2080.365018] env[62824]: DEBUG nova.compute.manager [req-914bea17-6ebc-439c-bbf2-9541bb48d04b req-10bf24fd-83c0-436e-a1f0-7b139b8841ab service nova] [instance: 02f34565-6675-4c79-ac47-b131ceba9df8] No waiting events found dispatching network-vif-plugged-be36fef6-17a7-4c25-bf33-98cc6a037b53 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2080.365182] env[62824]: WARNING nova.compute.manager [req-914bea17-6ebc-439c-bbf2-9541bb48d04b req-10bf24fd-83c0-436e-a1f0-7b139b8841ab service nova] [instance: 02f34565-6675-4c79-ac47-b131ceba9df8] Received unexpected event network-vif-plugged-be36fef6-17a7-4c25-bf33-98cc6a037b53 for instance with vm_state building and task_state spawning. [ 2080.365338] env[62824]: DEBUG nova.compute.manager [req-914bea17-6ebc-439c-bbf2-9541bb48d04b req-10bf24fd-83c0-436e-a1f0-7b139b8841ab service nova] [instance: 02f34565-6675-4c79-ac47-b131ceba9df8] Received event network-changed-be36fef6-17a7-4c25-bf33-98cc6a037b53 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2080.365484] env[62824]: DEBUG nova.compute.manager [req-914bea17-6ebc-439c-bbf2-9541bb48d04b req-10bf24fd-83c0-436e-a1f0-7b139b8841ab service nova] [instance: 02f34565-6675-4c79-ac47-b131ceba9df8] Refreshing instance network info cache due to event network-changed-be36fef6-17a7-4c25-bf33-98cc6a037b53. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2080.365661] env[62824]: DEBUG oslo_concurrency.lockutils [req-914bea17-6ebc-439c-bbf2-9541bb48d04b req-10bf24fd-83c0-436e-a1f0-7b139b8841ab service nova] Acquiring lock "refresh_cache-02f34565-6675-4c79-ac47-b131ceba9df8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2080.365858] env[62824]: DEBUG oslo_concurrency.lockutils [req-914bea17-6ebc-439c-bbf2-9541bb48d04b req-10bf24fd-83c0-436e-a1f0-7b139b8841ab service nova] Acquired lock "refresh_cache-02f34565-6675-4c79-ac47-b131ceba9df8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2080.366018] env[62824]: DEBUG nova.network.neutron [req-914bea17-6ebc-439c-bbf2-9541bb48d04b req-10bf24fd-83c0-436e-a1f0-7b139b8841ab service nova] [instance: 02f34565-6675-4c79-ac47-b131ceba9df8] Refreshing network info cache for port be36fef6-17a7-4c25-bf33-98cc6a037b53 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2080.391852] env[62824]: DEBUG nova.network.neutron [-] [instance: ed7d3e95-f26e-40fd-a370-e17922bbff8e] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2080.561233] env[62824]: DEBUG oslo_vmware.api [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Task: {'id': task-2146248, 'name': ReconfigVM_Task, 'duration_secs': 0.541587} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2080.561560] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 15479062-af75-4925-99b3-77d6a49751ad] Reconfigured VM instance instance-0000003c to attach disk [datastore2] 15479062-af75-4925-99b3-77d6a49751ad/ephemeral_0.vmdk or device None with type thin {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2080.562198] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6d6a19b2-f014-4275-95e7-78c4fa966e12 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2080.570118] env[62824]: DEBUG oslo_vmware.api [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Waiting for the task: (returnval){ [ 2080.570118] env[62824]: value = "task-2146253" [ 2080.570118] env[62824]: _type = "Task" [ 2080.570118] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2080.579741] env[62824]: DEBUG oslo_vmware.api [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Task: {'id': task-2146253, 'name': Rename_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2080.608759] env[62824]: DEBUG oslo_vmware.api [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52dd0c94-7892-2929-7f97-ae1a7ccd3d44, 'name': SearchDatastore_Task, 'duration_secs': 0.014504} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2080.610184] env[62824]: DEBUG oslo_concurrency.lockutils [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2080.610184] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02f34565-6675-4c79-ac47-b131ceba9df8] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2080.610184] env[62824]: DEBUG oslo_concurrency.lockutils [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2080.610184] env[62824]: DEBUG oslo_concurrency.lockutils [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2080.610184] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2080.610395] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f9733734-9b47-4a6a-9f0c-8c1a6bf21986 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2080.623507] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5235cec8-f7d5-4de0-b94a-cfb615a0ae19 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.009s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2080.624802] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8b5da446-9f53-4569-8936-1a55fec9a1e7 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.313s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2080.624802] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8b5da446-9f53-4569-8936-1a55fec9a1e7 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2080.626717] env[62824]: DEBUG oslo_concurrency.lockutils [None req-23d5dc9c-2a7f-430b-a467-cfa30a8b4ca0 tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.260s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2080.626781] env[62824]: DEBUG oslo_concurrency.lockutils [None req-23d5dc9c-2a7f-430b-a467-cfa30a8b4ca0 tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2080.629267] env[62824]: DEBUG oslo_concurrency.lockutils [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.735s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2080.630916] env[62824]: INFO nova.compute.claims [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2080.633314] env[62824]: WARNING oslo_messaging._drivers.amqpdriver [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Number of call queues is 11, greater than warning threshold: 10. There could be a leak. Increasing threshold to: 20 [ 2080.637563] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: db018af0-7bdf-40c5-b129-fde15dfb8ab4] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2080.640874] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2080.640874] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2080.641273] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-43259c3f-ea30-45d6-8d23-4d06833c5357 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2080.651718] env[62824]: DEBUG oslo_vmware.api [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for the task: (returnval){ [ 2080.651718] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52309cab-d72e-276c-cee2-d0c46689962d" [ 2080.651718] env[62824]: _type = "Task" [ 2080.651718] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2080.664280] env[62824]: DEBUG oslo_vmware.api [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52309cab-d72e-276c-cee2-d0c46689962d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2080.665408] env[62824]: INFO nova.scheduler.client.report [None req-23d5dc9c-2a7f-430b-a467-cfa30a8b4ca0 tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Deleted allocations for instance 82382932-7302-4441-a6f8-9aa2300ec0f6 [ 2080.671715] env[62824]: DEBUG oslo_vmware.api [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146252, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2080.671715] env[62824]: INFO nova.scheduler.client.report [None req-8b5da446-9f53-4569-8936-1a55fec9a1e7 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Deleted allocations for instance 2afecec5-763d-4616-b690-41d3101cfc47 [ 2080.805193] env[62824]: DEBUG oslo_concurrency.lockutils [req-1d0f0919-8440-406f-9a64-216cf7a1cf5d req-d57e62e6-a6cf-4d08-a872-fa236ed0a943 service nova] Releasing lock "refresh_cache-a6cd2032-de60-4f78-bf1e-79801d049df0" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2080.805193] env[62824]: DEBUG nova.compute.manager [req-1d0f0919-8440-406f-9a64-216cf7a1cf5d req-d57e62e6-a6cf-4d08-a872-fa236ed0a943 service nova] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Received event network-vif-deleted-c34f030d-46da-4a7a-9703-5acb2e526d50 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2080.805193] env[62824]: DEBUG nova.compute.manager [req-1d0f0919-8440-406f-9a64-216cf7a1cf5d req-d57e62e6-a6cf-4d08-a872-fa236ed0a943 service nova] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Received event network-vif-deleted-2d861cfc-3ad4-4d40-ad77-e4530d363421 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2080.894348] env[62824]: DEBUG nova.network.neutron [-] [instance: ed7d3e95-f26e-40fd-a370-e17922bbff8e] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2081.086952] env[62824]: DEBUG oslo_vmware.api [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Task: {'id': task-2146253, 'name': Rename_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2081.115399] env[62824]: DEBUG nova.network.neutron [-] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2081.147715] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 41f068d3-2c8d-46f0-8d84-78531d28b0dd] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2081.168170] env[62824]: DEBUG oslo_vmware.api [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52309cab-d72e-276c-cee2-d0c46689962d, 'name': SearchDatastore_Task, 'duration_secs': 0.015965} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2081.172376] env[62824]: DEBUG oslo_vmware.api [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146252, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2081.173158] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aca00f64-97eb-4aa9-88c5-e3a06f35a9cb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2081.180565] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8b5da446-9f53-4569-8936-1a55fec9a1e7 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Lock "2afecec5-763d-4616-b690-41d3101cfc47" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.409s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2081.181627] env[62824]: DEBUG oslo_concurrency.lockutils [None req-23d5dc9c-2a7f-430b-a467-cfa30a8b4ca0 tempest-ServersV294TestFqdnHostnames-812722115 tempest-ServersV294TestFqdnHostnames-812722115-project-member] Lock "82382932-7302-4441-a6f8-9aa2300ec0f6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.801s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2081.183731] env[62824]: DEBUG oslo_vmware.api [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for the task: (returnval){ [ 2081.183731] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5252e62f-5f5e-c6af-4366-c93ce34b1933" [ 2081.183731] env[62824]: _type = "Task" [ 2081.183731] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2081.200168] env[62824]: DEBUG oslo_vmware.api [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5252e62f-5f5e-c6af-4366-c93ce34b1933, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2081.272174] env[62824]: DEBUG nova.network.neutron [req-914bea17-6ebc-439c-bbf2-9541bb48d04b req-10bf24fd-83c0-436e-a1f0-7b139b8841ab service nova] [instance: 02f34565-6675-4c79-ac47-b131ceba9df8] Updated VIF entry in instance network info cache for port be36fef6-17a7-4c25-bf33-98cc6a037b53. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2081.273297] env[62824]: DEBUG nova.network.neutron [req-914bea17-6ebc-439c-bbf2-9541bb48d04b req-10bf24fd-83c0-436e-a1f0-7b139b8841ab service nova] [instance: 02f34565-6675-4c79-ac47-b131ceba9df8] Updating instance_info_cache with network_info: [{"id": "be36fef6-17a7-4c25-bf33-98cc6a037b53", "address": "fa:16:3e:c1:60:d0", "network": {"id": "fc381ca5-198f-4afc-9ebf-bc6e3de68c35", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-425455451-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "248af2bc3e7b41d2a74b0c6074b359e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4c9a12d2-469f-4199-bfaa-f791d765deac", "external-id": "nsx-vlan-transportzone-96", "segmentation_id": 96, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbe36fef6-17", "ovs_interfaceid": "be36fef6-17a7-4c25-bf33-98cc6a037b53", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2081.397940] env[62824]: INFO nova.compute.manager [-] [instance: ed7d3e95-f26e-40fd-a370-e17922bbff8e] Took 1.04 seconds to deallocate network for instance. [ 2081.451247] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquiring lock "ca42fef0-1e90-4ab6-9d60-8ef7e4997884" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2081.451588] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "ca42fef0-1e90-4ab6-9d60-8ef7e4997884" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2081.583033] env[62824]: DEBUG oslo_vmware.api [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Task: {'id': task-2146253, 'name': Rename_Task, 'duration_secs': 0.667835} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2081.583337] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 15479062-af75-4925-99b3-77d6a49751ad] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2081.583576] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6e797a57-c020-492f-a81c-6b5b07e97af2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2081.593532] env[62824]: DEBUG oslo_vmware.api [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Waiting for the task: (returnval){ [ 2081.593532] env[62824]: value = "task-2146254" [ 2081.593532] env[62824]: _type = "Task" [ 2081.593532] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2081.600246] env[62824]: DEBUG oslo_vmware.api [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Task: {'id': task-2146254, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2081.621378] env[62824]: INFO nova.compute.manager [-] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Took 1.92 seconds to deallocate network for instance. [ 2081.649906] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: dc3317e4-6958-4df6-8e39-c89b4a5d861c] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2081.679454] env[62824]: DEBUG oslo_vmware.api [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146252, 'name': PowerOnVM_Task, 'duration_secs': 1.360715} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2081.681923] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: a6cd2032-de60-4f78-bf1e-79801d049df0] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2081.682142] env[62824]: INFO nova.compute.manager [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: a6cd2032-de60-4f78-bf1e-79801d049df0] Took 8.99 seconds to spawn the instance on the hypervisor. [ 2081.682309] env[62824]: DEBUG nova.compute.manager [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: a6cd2032-de60-4f78-bf1e-79801d049df0] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2081.683517] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b783b6ef-1bd9-472e-8004-096d40ce8a50 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2081.714591] env[62824]: DEBUG oslo_vmware.api [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5252e62f-5f5e-c6af-4366-c93ce34b1933, 'name': SearchDatastore_Task, 'duration_secs': 0.018315} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2081.715359] env[62824]: DEBUG oslo_concurrency.lockutils [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2081.715771] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 02f34565-6675-4c79-ac47-b131ceba9df8/02f34565-6675-4c79-ac47-b131ceba9df8.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2081.716177] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cb7002e9-77bf-4093-8f68-3bbf89dcdc6f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2081.728493] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a851343c-36b8-4f6b-b725-ef04164a199a tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquiring lock "919e8a45-7810-4a8d-a1aa-5046b5ab059c" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2081.728780] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a851343c-36b8-4f6b-b725-ef04164a199a tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "919e8a45-7810-4a8d-a1aa-5046b5ab059c" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2081.733688] env[62824]: DEBUG oslo_vmware.api [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for the task: (returnval){ [ 2081.733688] env[62824]: value = "task-2146255" [ 2081.733688] env[62824]: _type = "Task" [ 2081.733688] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2081.752299] env[62824]: DEBUG oslo_vmware.api [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146255, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2081.779438] env[62824]: DEBUG oslo_concurrency.lockutils [req-914bea17-6ebc-439c-bbf2-9541bb48d04b req-10bf24fd-83c0-436e-a1f0-7b139b8841ab service nova] Releasing lock "refresh_cache-02f34565-6675-4c79-ac47-b131ceba9df8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2081.814611] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fb4bd00b-e6f9-4459-a9db-0d3353b82d74 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Acquiring lock "dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2081.814883] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fb4bd00b-e6f9-4459-a9db-0d3353b82d74 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Lock "dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2081.818580] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fb4bd00b-e6f9-4459-a9db-0d3353b82d74 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Acquiring lock "dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2081.818580] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fb4bd00b-e6f9-4459-a9db-0d3353b82d74 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Lock "dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2081.818580] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fb4bd00b-e6f9-4459-a9db-0d3353b82d74 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Lock "dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2081.818580] env[62824]: INFO nova.compute.manager [None req-fb4bd00b-e6f9-4459-a9db-0d3353b82d74 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] [instance: dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda] Terminating instance [ 2081.910738] env[62824]: DEBUG oslo_concurrency.lockutils [None req-84ba9f19-2925-48b1-85d5-00fe075f71bc tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2081.955303] env[62824]: DEBUG nova.compute.manager [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2082.106986] env[62824]: DEBUG oslo_vmware.api [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Task: {'id': task-2146254, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2082.129137] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3aa78573-24a8-42b5-8896-df119ccd8889 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2082.159298] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 5f1237ff-5841-41f1-8d77-df9fabaa62b6] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2082.233720] env[62824]: INFO nova.compute.manager [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: a6cd2032-de60-4f78-bf1e-79801d049df0] Took 35.78 seconds to build instance. [ 2082.239920] env[62824]: DEBUG nova.compute.utils [None req-a851343c-36b8-4f6b-b725-ef04164a199a tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2082.257382] env[62824]: DEBUG oslo_vmware.api [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146255, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2082.266703] env[62824]: DEBUG oslo_concurrency.lockutils [None req-56d14257-fc28-4784-a770-ae333154aa8d tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Acquiring lock "c36ece43-3d70-4e67-a740-9057f413c722" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2082.266913] env[62824]: DEBUG oslo_concurrency.lockutils [None req-56d14257-fc28-4784-a770-ae333154aa8d tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Lock "c36ece43-3d70-4e67-a740-9057f413c722" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2082.267010] env[62824]: DEBUG oslo_concurrency.lockutils [None req-56d14257-fc28-4784-a770-ae333154aa8d tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Acquiring lock "c36ece43-3d70-4e67-a740-9057f413c722-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2082.267262] env[62824]: DEBUG oslo_concurrency.lockutils [None req-56d14257-fc28-4784-a770-ae333154aa8d tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Lock "c36ece43-3d70-4e67-a740-9057f413c722-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2082.267451] env[62824]: DEBUG oslo_concurrency.lockutils [None req-56d14257-fc28-4784-a770-ae333154aa8d tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Lock "c36ece43-3d70-4e67-a740-9057f413c722-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2082.269907] env[62824]: INFO nova.compute.manager [None req-56d14257-fc28-4784-a770-ae333154aa8d tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] [instance: c36ece43-3d70-4e67-a740-9057f413c722] Terminating instance [ 2082.309023] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e443e021-27e4-44e0-b922-6e6be983f7d5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2082.319225] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5074442a-e511-4aed-a6dc-3b26dcefc4cb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2082.324209] env[62824]: DEBUG nova.compute.manager [None req-fb4bd00b-e6f9-4459-a9db-0d3353b82d74 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] [instance: dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2082.324410] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-fb4bd00b-e6f9-4459-a9db-0d3353b82d74 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] [instance: dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2082.325498] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84ad89a3-1097-47fb-b2cd-bf301b451a29 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2082.366230] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42763c2a-21eb-4f38-b5c3-b5a8d78064e7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2082.370719] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb4bd00b-e6f9-4459-a9db-0d3353b82d74 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] [instance: dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2082.370719] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ff4e12aa-97c8-438d-855c-8072abd1cade {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2082.378068] env[62824]: DEBUG oslo_concurrency.lockutils [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquiring lock "5b3df31f-15fe-473f-992c-ddb272661c53" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2082.378068] env[62824]: DEBUG oslo_concurrency.lockutils [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lock "5b3df31f-15fe-473f-992c-ddb272661c53" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2082.381298] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df27cbda-7899-4a67-80f8-344d94deedf0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2082.386772] env[62824]: DEBUG oslo_vmware.api [None req-fb4bd00b-e6f9-4459-a9db-0d3353b82d74 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Waiting for the task: (returnval){ [ 2082.386772] env[62824]: value = "task-2146256" [ 2082.386772] env[62824]: _type = "Task" [ 2082.386772] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2082.398807] env[62824]: DEBUG nova.compute.provider_tree [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2082.407824] env[62824]: DEBUG oslo_vmware.api [None req-fb4bd00b-e6f9-4459-a9db-0d3353b82d74 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Task: {'id': task-2146256, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2082.478527] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2082.611252] env[62824]: DEBUG oslo_vmware.api [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Task: {'id': task-2146254, 'name': PowerOnVM_Task, 'duration_secs': 0.707772} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2082.611729] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 15479062-af75-4925-99b3-77d6a49751ad] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2082.611987] env[62824]: INFO nova.compute.manager [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 15479062-af75-4925-99b3-77d6a49751ad] Took 12.31 seconds to spawn the instance on the hypervisor. [ 2082.612189] env[62824]: DEBUG nova.compute.manager [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 15479062-af75-4925-99b3-77d6a49751ad] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2082.613312] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae2b150d-9096-4722-81c0-65641ae19643 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2082.662703] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 69bcac33-d6ac-4ee7-b674-58ab25389149] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2082.742713] env[62824]: DEBUG oslo_concurrency.lockutils [None req-be34a00a-a96a-4174-bda1-4b4d09d4e789 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "a6cd2032-de60-4f78-bf1e-79801d049df0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 51.965s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2082.752775] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a851343c-36b8-4f6b-b725-ef04164a199a tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "919e8a45-7810-4a8d-a1aa-5046b5ab059c" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.024s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2082.753265] env[62824]: DEBUG oslo_vmware.api [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146255, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.663233} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2082.753543] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 02f34565-6675-4c79-ac47-b131ceba9df8/02f34565-6675-4c79-ac47-b131ceba9df8.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2082.753752] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02f34565-6675-4c79-ac47-b131ceba9df8] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2082.755244] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7c2b741b-ea3b-4a1f-b97c-8b938003ba1d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2082.763096] env[62824]: DEBUG oslo_vmware.api [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for the task: (returnval){ [ 2082.763096] env[62824]: value = "task-2146257" [ 2082.763096] env[62824]: _type = "Task" [ 2082.763096] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2082.778848] env[62824]: DEBUG oslo_vmware.api [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146257, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2082.779440] env[62824]: DEBUG nova.compute.manager [None req-56d14257-fc28-4784-a770-ae333154aa8d tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] [instance: c36ece43-3d70-4e67-a740-9057f413c722] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2082.779635] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-56d14257-fc28-4784-a770-ae333154aa8d tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] [instance: c36ece43-3d70-4e67-a740-9057f413c722] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2082.780479] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63b38fc1-7bdd-4384-a485-9e28901e8da9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2082.790448] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-56d14257-fc28-4784-a770-ae333154aa8d tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] [instance: c36ece43-3d70-4e67-a740-9057f413c722] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2082.790448] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fa249d65-406e-4ae0-b850-de041a8bc5c5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2082.800022] env[62824]: DEBUG oslo_vmware.api [None req-56d14257-fc28-4784-a770-ae333154aa8d tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Waiting for the task: (returnval){ [ 2082.800022] env[62824]: value = "task-2146258" [ 2082.800022] env[62824]: _type = "Task" [ 2082.800022] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2082.815870] env[62824]: DEBUG oslo_vmware.api [None req-56d14257-fc28-4784-a770-ae333154aa8d tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Task: {'id': task-2146258, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2082.888291] env[62824]: DEBUG nova.compute.manager [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2082.909370] env[62824]: DEBUG nova.scheduler.client.report [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2082.915263] env[62824]: DEBUG oslo_vmware.api [None req-fb4bd00b-e6f9-4459-a9db-0d3353b82d74 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Task: {'id': task-2146256, 'name': PowerOffVM_Task, 'duration_secs': 0.316825} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2082.915902] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb4bd00b-e6f9-4459-a9db-0d3353b82d74 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] [instance: dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2082.916072] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-fb4bd00b-e6f9-4459-a9db-0d3353b82d74 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] [instance: dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2082.916339] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1745e6ca-0075-4483-9eea-02b88a18943f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2083.026753] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-fb4bd00b-e6f9-4459-a9db-0d3353b82d74 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] [instance: dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2083.027162] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-fb4bd00b-e6f9-4459-a9db-0d3353b82d74 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] [instance: dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda] Deleting contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2083.027398] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-fb4bd00b-e6f9-4459-a9db-0d3353b82d74 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Deleting the datastore file [datastore2] dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2083.028245] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a804301b-5093-4433-be56-7c8cdc154cda {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2083.037471] env[62824]: DEBUG oslo_vmware.api [None req-fb4bd00b-e6f9-4459-a9db-0d3353b82d74 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Waiting for the task: (returnval){ [ 2083.037471] env[62824]: value = "task-2146260" [ 2083.037471] env[62824]: _type = "Task" [ 2083.037471] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2083.049872] env[62824]: DEBUG oslo_vmware.api [None req-fb4bd00b-e6f9-4459-a9db-0d3353b82d74 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Task: {'id': task-2146260, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2083.143950] env[62824]: INFO nova.compute.manager [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 15479062-af75-4925-99b3-77d6a49751ad] Took 40.10 seconds to build instance. [ 2083.166397] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 487c2c9d-2cd2-4912-8613-e1bfac732c40] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2083.274712] env[62824]: DEBUG oslo_vmware.api [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146257, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.128451} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2083.275093] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02f34565-6675-4c79-ac47-b131ceba9df8] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2083.275977] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b283329-3d81-4959-bf23-57c9b7f7262b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2083.303426] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02f34565-6675-4c79-ac47-b131ceba9df8] Reconfiguring VM instance instance-0000003e to attach disk [datastore1] 02f34565-6675-4c79-ac47-b131ceba9df8/02f34565-6675-4c79-ac47-b131ceba9df8.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2083.307091] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b8d3f4a2-0229-47f9-85db-a3a6b571da90 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2083.348217] env[62824]: DEBUG oslo_vmware.api [None req-56d14257-fc28-4784-a770-ae333154aa8d tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Task: {'id': task-2146258, 'name': PowerOffVM_Task, 'duration_secs': 0.287176} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2083.350940] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-56d14257-fc28-4784-a770-ae333154aa8d tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] [instance: c36ece43-3d70-4e67-a740-9057f413c722] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2083.352074] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-56d14257-fc28-4784-a770-ae333154aa8d tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] [instance: c36ece43-3d70-4e67-a740-9057f413c722] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2083.352074] env[62824]: DEBUG oslo_vmware.api [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for the task: (returnval){ [ 2083.352074] env[62824]: value = "task-2146261" [ 2083.352074] env[62824]: _type = "Task" [ 2083.352074] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2083.352343] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9f612674-92a5-40c7-841a-240ec4fa35ef {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2083.368440] env[62824]: DEBUG oslo_vmware.api [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146261, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2083.424624] env[62824]: DEBUG oslo_concurrency.lockutils [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2083.425431] env[62824]: DEBUG oslo_concurrency.lockutils [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.796s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2083.425943] env[62824]: DEBUG nova.compute.manager [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2083.429385] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e99f51a7-e71c-443f-812d-3cdaa01ba099 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.486s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2083.429633] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e99f51a7-e71c-443f-812d-3cdaa01ba099 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2083.431973] env[62824]: DEBUG oslo_concurrency.lockutils [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.224s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2083.432238] env[62824]: DEBUG oslo_concurrency.lockutils [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2083.438028] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d515df3c-54b5-4dd7-8598-40172b6be750 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.495s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2083.438028] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d515df3c-54b5-4dd7-8598-40172b6be750 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2083.440455] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 9.636s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2083.440689] env[62824]: DEBUG nova.objects.instance [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] [instance: ed7d3e95-f26e-40fd-a370-e17922bbff8e] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62824) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 2083.474356] env[62824]: INFO nova.scheduler.client.report [None req-e99f51a7-e71c-443f-812d-3cdaa01ba099 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Deleted allocations for instance 21c05725-3dad-4965-98d0-0622ebcc8ece [ 2083.484233] env[62824]: INFO nova.scheduler.client.report [None req-d515df3c-54b5-4dd7-8598-40172b6be750 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Deleted allocations for instance 88bad2dd-dce7-41df-b56c-93a5d054c11e [ 2083.491907] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-56d14257-fc28-4784-a770-ae333154aa8d tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] [instance: c36ece43-3d70-4e67-a740-9057f413c722] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2083.492633] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-56d14257-fc28-4784-a770-ae333154aa8d tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] [instance: c36ece43-3d70-4e67-a740-9057f413c722] Deleting contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2083.492633] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-56d14257-fc28-4784-a770-ae333154aa8d tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Deleting the datastore file [datastore2] c36ece43-3d70-4e67-a740-9057f413c722 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2083.493022] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-74ff7699-448b-4ccc-a8bb-1d6e327acb0a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2083.502964] env[62824]: DEBUG oslo_vmware.api [None req-56d14257-fc28-4784-a770-ae333154aa8d tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Waiting for the task: (returnval){ [ 2083.502964] env[62824]: value = "task-2146263" [ 2083.502964] env[62824]: _type = "Task" [ 2083.502964] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2083.517702] env[62824]: DEBUG nova.compute.manager [req-3674a087-4dc7-4574-a5f2-5283fb43df26 req-7e4a237e-e515-41aa-a5ab-e84061ae528b service nova] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Received event network-changed-d358ab3c-f4fb-41f3-954b-19de8b22ef15 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2083.517702] env[62824]: DEBUG nova.compute.manager [req-3674a087-4dc7-4574-a5f2-5283fb43df26 req-7e4a237e-e515-41aa-a5ab-e84061ae528b service nova] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Refreshing instance network info cache due to event network-changed-d358ab3c-f4fb-41f3-954b-19de8b22ef15. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2083.517702] env[62824]: DEBUG oslo_concurrency.lockutils [req-3674a087-4dc7-4574-a5f2-5283fb43df26 req-7e4a237e-e515-41aa-a5ab-e84061ae528b service nova] Acquiring lock "refresh_cache-2f315f9e-94e4-47ef-9503-ee92b59e5452" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2083.517702] env[62824]: DEBUG oslo_concurrency.lockutils [req-3674a087-4dc7-4574-a5f2-5283fb43df26 req-7e4a237e-e515-41aa-a5ab-e84061ae528b service nova] Acquired lock "refresh_cache-2f315f9e-94e4-47ef-9503-ee92b59e5452" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2083.517702] env[62824]: DEBUG nova.network.neutron [req-3674a087-4dc7-4574-a5f2-5283fb43df26 req-7e4a237e-e515-41aa-a5ab-e84061ae528b service nova] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Refreshing network info cache for port d358ab3c-f4fb-41f3-954b-19de8b22ef15 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2083.530463] env[62824]: DEBUG oslo_vmware.api [None req-56d14257-fc28-4784-a770-ae333154aa8d tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Task: {'id': task-2146263, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2083.553043] env[62824]: DEBUG oslo_vmware.api [None req-fb4bd00b-e6f9-4459-a9db-0d3353b82d74 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Task: {'id': task-2146260, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.35028} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2083.554083] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-fb4bd00b-e6f9-4459-a9db-0d3353b82d74 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2083.554534] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-fb4bd00b-e6f9-4459-a9db-0d3353b82d74 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] [instance: dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda] Deleted contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2083.554924] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-fb4bd00b-e6f9-4459-a9db-0d3353b82d74 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] [instance: dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2083.555201] env[62824]: INFO nova.compute.manager [None req-fb4bd00b-e6f9-4459-a9db-0d3353b82d74 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] [instance: dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda] Took 1.23 seconds to destroy the instance on the hypervisor. [ 2083.555544] env[62824]: DEBUG oslo.service.loopingcall [None req-fb4bd00b-e6f9-4459-a9db-0d3353b82d74 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2083.555843] env[62824]: DEBUG nova.compute.manager [-] [instance: dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2083.556154] env[62824]: DEBUG nova.network.neutron [-] [instance: dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2083.646832] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8d439482-8d19-4f0b-b9d5-21105a0d09b9 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Lock "15479062-af75-4925-99b3-77d6a49751ad" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 79.131s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2083.670218] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: b96d1351-f3a7-4bac-998c-a34ab2606041] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2083.834800] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a851343c-36b8-4f6b-b725-ef04164a199a tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquiring lock "919e8a45-7810-4a8d-a1aa-5046b5ab059c" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2083.835163] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a851343c-36b8-4f6b-b725-ef04164a199a tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "919e8a45-7810-4a8d-a1aa-5046b5ab059c" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2083.836033] env[62824]: INFO nova.compute.manager [None req-a851343c-36b8-4f6b-b725-ef04164a199a tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 919e8a45-7810-4a8d-a1aa-5046b5ab059c] Attaching volume 20c6ff5e-0a8b-4cf4-8bd3-92cb944dd7a3 to /dev/sdb [ 2083.837587] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b1f093ef-801c-4886-991f-869391717179 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Acquiring lock "8bb90a4d-93c9-4f54-b15e-48fe966e6c97" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2083.838110] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b1f093ef-801c-4886-991f-869391717179 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Lock "8bb90a4d-93c9-4f54-b15e-48fe966e6c97" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2083.838110] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b1f093ef-801c-4886-991f-869391717179 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Acquiring lock "8bb90a4d-93c9-4f54-b15e-48fe966e6c97-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2083.838291] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b1f093ef-801c-4886-991f-869391717179 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Lock "8bb90a4d-93c9-4f54-b15e-48fe966e6c97-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2083.838494] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b1f093ef-801c-4886-991f-869391717179 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Lock "8bb90a4d-93c9-4f54-b15e-48fe966e6c97-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2083.841506] env[62824]: INFO nova.compute.manager [None req-b1f093ef-801c-4886-991f-869391717179 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 8bb90a4d-93c9-4f54-b15e-48fe966e6c97] Terminating instance [ 2083.867312] env[62824]: DEBUG oslo_vmware.api [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146261, 'name': ReconfigVM_Task, 'duration_secs': 0.480926} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2083.869340] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02f34565-6675-4c79-ac47-b131ceba9df8] Reconfigured VM instance instance-0000003e to attach disk [datastore1] 02f34565-6675-4c79-ac47-b131ceba9df8/02f34565-6675-4c79-ac47-b131ceba9df8.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2083.872343] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7e3ef1fe-2994-456e-85d6-d7b3153223ff {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2083.883233] env[62824]: DEBUG oslo_vmware.api [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for the task: (returnval){ [ 2083.883233] env[62824]: value = "task-2146264" [ 2083.883233] env[62824]: _type = "Task" [ 2083.883233] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2083.885765] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21edf134-f4bb-4268-b4fb-e4c0af088383 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2083.899205] env[62824]: DEBUG oslo_vmware.api [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146264, 'name': Rename_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2083.901444] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e095c51-58f2-489c-848d-e9576ebeba17 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2083.919209] env[62824]: DEBUG nova.virt.block_device [None req-a851343c-36b8-4f6b-b725-ef04164a199a tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 919e8a45-7810-4a8d-a1aa-5046b5ab059c] Updating existing volume attachment record: becea960-8ede-4942-b95d-73ac4e6bac8b {{(pid=62824) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 2083.931348] env[62824]: DEBUG nova.compute.utils [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2083.932610] env[62824]: DEBUG nova.compute.manager [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2083.932781] env[62824]: DEBUG nova.network.neutron [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2083.944464] env[62824]: DEBUG oslo_concurrency.lockutils [None req-131c4deb-10bc-4e30-9be1-c5e4486f0bec tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Lock "b8cc8cd3-ea03-40bf-b867-7ad193365552" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 36.653s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2083.947121] env[62824]: DEBUG oslo_concurrency.lockutils [None req-123477fc-8c3e-4f2a-83de-cced4f011020 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Lock "b8cc8cd3-ea03-40bf-b867-7ad193365552" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 6.249s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2083.947446] env[62824]: INFO nova.compute.manager [None req-123477fc-8c3e-4f2a-83de-cced4f011020 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Unshelving [ 2083.996518] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e99f51a7-e71c-443f-812d-3cdaa01ba099 tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Lock "21c05725-3dad-4965-98d0-0622ebcc8ece" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.443s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2083.998393] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d515df3c-54b5-4dd7-8598-40172b6be750 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Lock "88bad2dd-dce7-41df-b56c-93a5d054c11e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.663s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2084.014832] env[62824]: DEBUG oslo_vmware.api [None req-56d14257-fc28-4784-a770-ae333154aa8d tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Task: {'id': task-2146263, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.267187} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2084.014938] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-56d14257-fc28-4784-a770-ae333154aa8d tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2084.015186] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-56d14257-fc28-4784-a770-ae333154aa8d tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] [instance: c36ece43-3d70-4e67-a740-9057f413c722] Deleted contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2084.015485] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-56d14257-fc28-4784-a770-ae333154aa8d tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] [instance: c36ece43-3d70-4e67-a740-9057f413c722] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2084.016830] env[62824]: INFO nova.compute.manager [None req-56d14257-fc28-4784-a770-ae333154aa8d tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] [instance: c36ece43-3d70-4e67-a740-9057f413c722] Took 1.24 seconds to destroy the instance on the hypervisor. [ 2084.017209] env[62824]: DEBUG oslo.service.loopingcall [None req-56d14257-fc28-4784-a770-ae333154aa8d tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2084.018076] env[62824]: DEBUG nova.compute.manager [-] [instance: c36ece43-3d70-4e67-a740-9057f413c722] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2084.018212] env[62824]: DEBUG nova.network.neutron [-] [instance: c36ece43-3d70-4e67-a740-9057f413c722] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2084.098585] env[62824]: DEBUG nova.policy [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4c914343569a462d9b8564f3376c2f54', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '10ee5fe93d584725ace90e76605a1b2e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2084.174234] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: b14bb74a-5e13-4d15-97d4-ee7b2d67f2e6] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2084.270339] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8854c3a5-43a5-4ba3-bce0-709d5c6fa611 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Acquiring lock "59b5b883-4188-471c-8862-444f3ce08cb0" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2084.270824] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8854c3a5-43a5-4ba3-bce0-709d5c6fa611 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Lock "59b5b883-4188-471c-8862-444f3ce08cb0" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2084.354593] env[62824]: DEBUG nova.compute.manager [None req-b1f093ef-801c-4886-991f-869391717179 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 8bb90a4d-93c9-4f54-b15e-48fe966e6c97] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2084.354643] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-b1f093ef-801c-4886-991f-869391717179 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 8bb90a4d-93c9-4f54-b15e-48fe966e6c97] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2084.356014] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-419011af-f859-4194-a7ed-f8acdef8befd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2084.366683] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1f093ef-801c-4886-991f-869391717179 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 8bb90a4d-93c9-4f54-b15e-48fe966e6c97] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2084.369022] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-507f878a-9a3a-4df8-b6ec-0601ddf2537c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2084.378145] env[62824]: DEBUG oslo_vmware.api [None req-b1f093ef-801c-4886-991f-869391717179 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Waiting for the task: (returnval){ [ 2084.378145] env[62824]: value = "task-2146268" [ 2084.378145] env[62824]: _type = "Task" [ 2084.378145] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2084.390385] env[62824]: DEBUG oslo_vmware.api [None req-b1f093ef-801c-4886-991f-869391717179 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': task-2146268, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2084.403481] env[62824]: DEBUG oslo_vmware.api [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146264, 'name': Rename_Task, 'duration_secs': 0.190242} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2084.403481] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02f34565-6675-4c79-ac47-b131ceba9df8] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2084.403481] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0b645739-e025-4cac-b89d-8e1a56f91277 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2084.412129] env[62824]: DEBUG oslo_vmware.api [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for the task: (returnval){ [ 2084.412129] env[62824]: value = "task-2146269" [ 2084.412129] env[62824]: _type = "Task" [ 2084.412129] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2084.423700] env[62824]: DEBUG oslo_vmware.api [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146269, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2084.437984] env[62824]: DEBUG nova.compute.manager [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2084.458018] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fa5ec54c-73db-4f7d-8136-a79c953dd1bf tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.017s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2084.463992] env[62824]: DEBUG oslo_concurrency.lockutils [None req-4715e6dc-1d1a-49a2-9d97-4f4b36f634e1 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.004s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2084.463992] env[62824]: DEBUG nova.objects.instance [None req-4715e6dc-1d1a-49a2-9d97-4f4b36f634e1 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Lazy-loading 'resources' on Instance uuid 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2084.678601] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 16b466c3-c749-4f96-a82c-32dad31138ec] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2084.775689] env[62824]: INFO nova.compute.manager [None req-8854c3a5-43a5-4ba3-bce0-709d5c6fa611 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Detaching volume 5c4381e7-4dfa-4683-93fe-8eb8bef1df7e [ 2084.816305] env[62824]: INFO nova.virt.block_device [None req-8854c3a5-43a5-4ba3-bce0-709d5c6fa611 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Attempting to driver detach volume 5c4381e7-4dfa-4683-93fe-8eb8bef1df7e from mountpoint /dev/sdb [ 2084.817045] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-8854c3a5-43a5-4ba3-bce0-709d5c6fa611 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Volume detach. Driver type: vmdk {{(pid=62824) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 2084.817273] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-8854c3a5-43a5-4ba3-bce0-709d5c6fa611 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-438620', 'volume_id': '5c4381e7-4dfa-4683-93fe-8eb8bef1df7e', 'name': 'volume-5c4381e7-4dfa-4683-93fe-8eb8bef1df7e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '59b5b883-4188-471c-8862-444f3ce08cb0', 'attached_at': '', 'detached_at': '', 'volume_id': '5c4381e7-4dfa-4683-93fe-8eb8bef1df7e', 'serial': '5c4381e7-4dfa-4683-93fe-8eb8bef1df7e'} {{(pid=62824) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 2084.818510] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea8cf667-ac29-48d6-a42d-9e3d449d8817 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2084.857501] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-286c1672-79ce-4701-8d2b-3126f7af3fa1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2084.868700] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5794bab0-2e4d-4406-b4d0-e6ac1db73fc8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2084.902297] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0165d3c7-f1d8-40ad-9035-8a2684c37f9f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2084.911278] env[62824]: DEBUG oslo_vmware.api [None req-b1f093ef-801c-4886-991f-869391717179 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': task-2146268, 'name': PowerOffVM_Task, 'duration_secs': 0.262357} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2084.924304] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1f093ef-801c-4886-991f-869391717179 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 8bb90a4d-93c9-4f54-b15e-48fe966e6c97] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2084.924543] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-b1f093ef-801c-4886-991f-869391717179 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 8bb90a4d-93c9-4f54-b15e-48fe966e6c97] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2084.924976] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-8854c3a5-43a5-4ba3-bce0-709d5c6fa611 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] The volume has not been displaced from its original location: [datastore1] volume-5c4381e7-4dfa-4683-93fe-8eb8bef1df7e/volume-5c4381e7-4dfa-4683-93fe-8eb8bef1df7e.vmdk. No consolidation needed. {{(pid=62824) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 2084.931066] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-8854c3a5-43a5-4ba3-bce0-709d5c6fa611 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Reconfiguring VM instance instance-00000022 to detach disk 2001 {{(pid=62824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 2084.931575] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b0ec7d47-00c2-4e91-9096-db6f0bf3c6be {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2084.936765] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-31375a7d-237a-4d26-aafc-e6e3dd144f27 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2084.978026] env[62824]: DEBUG nova.compute.utils [None req-123477fc-8c3e-4f2a-83de-cced4f011020 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2084.978026] env[62824]: DEBUG oslo_vmware.api [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146269, 'name': PowerOnVM_Task} progress is 76%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2084.978745] env[62824]: DEBUG oslo_vmware.api [None req-8854c3a5-43a5-4ba3-bce0-709d5c6fa611 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Waiting for the task: (returnval){ [ 2084.978745] env[62824]: value = "task-2146271" [ 2084.978745] env[62824]: _type = "Task" [ 2084.978745] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2084.989883] env[62824]: DEBUG oslo_vmware.api [None req-8854c3a5-43a5-4ba3-bce0-709d5c6fa611 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Task: {'id': task-2146271, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2085.177590] env[62824]: DEBUG nova.network.neutron [req-3674a087-4dc7-4574-a5f2-5283fb43df26 req-7e4a237e-e515-41aa-a5ab-e84061ae528b service nova] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Updated VIF entry in instance network info cache for port d358ab3c-f4fb-41f3-954b-19de8b22ef15. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2085.181042] env[62824]: DEBUG nova.network.neutron [req-3674a087-4dc7-4574-a5f2-5283fb43df26 req-7e4a237e-e515-41aa-a5ab-e84061ae528b service nova] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Updating instance_info_cache with network_info: [{"id": "d358ab3c-f4fb-41f3-954b-19de8b22ef15", "address": "fa:16:3e:12:72:e1", "network": {"id": "2ed8a86a-9c14-47b4-8cb4-e8496d289a6e", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1207171730-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.147", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "87239f55d6d44606b23ca0c25d9fb4d9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7514a465-f1a4-4a8b-b76b-726b1a9d7e2f", "external-id": "nsx-vlan-transportzone-36", "segmentation_id": 36, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd358ab3c-f4", "ovs_interfaceid": "d358ab3c-f4fb-41f3-954b-19de8b22ef15", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2085.182367] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: ecd6ff7d-802b-4f75-afc4-b99fe5b3ded5] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2085.441805] env[62824]: DEBUG oslo_vmware.api [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146269, 'name': PowerOnVM_Task, 'duration_secs': 0.765722} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2085.442092] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02f34565-6675-4c79-ac47-b131ceba9df8] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2085.442293] env[62824]: INFO nova.compute.manager [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02f34565-6675-4c79-ac47-b131ceba9df8] Took 10.00 seconds to spawn the instance on the hypervisor. [ 2085.442464] env[62824]: DEBUG nova.compute.manager [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02f34565-6675-4c79-ac47-b131ceba9df8] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2085.445535] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5348edb6-a36a-4e96-b7b3-c8e93f3d8e81 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2085.459159] env[62824]: DEBUG nova.compute.manager [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2085.482240] env[62824]: INFO nova.virt.block_device [None req-123477fc-8c3e-4f2a-83de-cced4f011020 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Booting with volume 6d6b6ae6-f851-46e7-9932-6b6c40939347 at /dev/sdb [ 2085.505334] env[62824]: DEBUG oslo_vmware.api [None req-8854c3a5-43a5-4ba3-bce0-709d5c6fa611 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Task: {'id': task-2146271, 'name': ReconfigVM_Task, 'duration_secs': 0.38005} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2085.505647] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-8854c3a5-43a5-4ba3-bce0-709d5c6fa611 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Reconfigured VM instance instance-00000022 to detach disk 2001 {{(pid=62824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 2085.511552] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f16d2398-df0a-4c23-8e90-f2463fd4dbcf {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2085.525548] env[62824]: DEBUG nova.virt.hardware [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2085.525842] env[62824]: DEBUG nova.virt.hardware [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2085.526060] env[62824]: DEBUG nova.virt.hardware [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2085.526308] env[62824]: DEBUG nova.virt.hardware [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2085.526486] env[62824]: DEBUG nova.virt.hardware [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2085.526657] env[62824]: DEBUG nova.virt.hardware [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2085.526932] env[62824]: DEBUG nova.virt.hardware [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2085.527320] env[62824]: DEBUG nova.virt.hardware [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2085.527528] env[62824]: DEBUG nova.virt.hardware [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2085.527663] env[62824]: DEBUG nova.virt.hardware [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2085.528497] env[62824]: DEBUG nova.virt.hardware [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2085.528769] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98719ad8-28ae-4ad0-8759-ec41ac610e68 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2085.538096] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-b1f093ef-801c-4886-991f-869391717179 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 8bb90a4d-93c9-4f54-b15e-48fe966e6c97] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2085.538096] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-b1f093ef-801c-4886-991f-869391717179 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 8bb90a4d-93c9-4f54-b15e-48fe966e6c97] Deleting contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2085.538096] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-b1f093ef-801c-4886-991f-869391717179 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Deleting the datastore file [datastore2] 8bb90a4d-93c9-4f54-b15e-48fe966e6c97 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2085.546685] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b7468153-0131-41b9-b093-f828e2282a64 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2085.556226] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a06e7c31-df5b-4221-827e-7658259d666e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2085.560468] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9689ceaa-9427-4bba-8f63-277240e9e3f4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2085.562102] env[62824]: DEBUG oslo_vmware.api [None req-8854c3a5-43a5-4ba3-bce0-709d5c6fa611 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Waiting for the task: (returnval){ [ 2085.562102] env[62824]: value = "task-2146272" [ 2085.562102] env[62824]: _type = "Task" [ 2085.562102] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2085.582343] env[62824]: DEBUG oslo_vmware.api [None req-b1f093ef-801c-4886-991f-869391717179 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Waiting for the task: (returnval){ [ 2085.582343] env[62824]: value = "task-2146273" [ 2085.582343] env[62824]: _type = "Task" [ 2085.582343] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2085.585250] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e7c028c-e884-4ae2-9e09-c7409d27ca50 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2085.608044] env[62824]: DEBUG oslo_vmware.api [None req-8854c3a5-43a5-4ba3-bce0-709d5c6fa611 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Task: {'id': task-2146272, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2085.616953] env[62824]: DEBUG oslo_vmware.api [None req-b1f093ef-801c-4886-991f-869391717179 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': task-2146273, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2085.629633] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-185bce41-c7f2-4a0f-a963-70145c20c66a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2085.650237] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-91ed6c08-ec61-4499-bbc7-971f6a783583 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2085.655822] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62b8c77f-d0ed-4d76-b51d-c3eb6744c886 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2085.667034] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-518eb694-eca6-4a8e-aaca-581e9c156a62 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2085.707766] env[62824]: DEBUG oslo_concurrency.lockutils [req-3674a087-4dc7-4574-a5f2-5283fb43df26 req-7e4a237e-e515-41aa-a5ab-e84061ae528b service nova] Releasing lock "refresh_cache-2f315f9e-94e4-47ef-9503-ee92b59e5452" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2085.707766] env[62824]: DEBUG nova.compute.manager [req-3674a087-4dc7-4574-a5f2-5283fb43df26 req-7e4a237e-e515-41aa-a5ab-e84061ae528b service nova] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Received event network-vif-deleted-5a010977-7d05-47e0-8688-bbf9a0e70900 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2085.707946] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: e158b5d4-c120-4e6c-89c4-7668e097926d] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2085.713560] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c25dd43c-e2a3-41f7-95c6-7dd1c9525744 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2085.743333] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58c24eeb-d80f-42ae-84cc-b7d664dd0fc3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2085.749726] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c387ce3-8871-488c-9d88-0244e36441eb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2085.765873] env[62824]: DEBUG nova.network.neutron [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8] Successfully created port: bfd88146-4384-4514-8ceb-46c647c80eee {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2085.768699] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0dac23b-0a0a-4f3d-baf4-5624746263a0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2085.771951] env[62824]: DEBUG nova.compute.provider_tree [None req-4715e6dc-1d1a-49a2-9d97-4f4b36f634e1 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2085.779869] env[62824]: DEBUG nova.network.neutron [-] [instance: dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2085.789675] env[62824]: DEBUG nova.network.neutron [-] [instance: c36ece43-3d70-4e67-a740-9057f413c722] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2085.790358] env[62824]: DEBUG nova.virt.block_device [None req-123477fc-8c3e-4f2a-83de-cced4f011020 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Updating existing volume attachment record: 7904770d-90e6-479c-987a-f5083709a54f {{(pid=62824) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 2085.968259] env[62824]: INFO nova.compute.manager [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02f34565-6675-4c79-ac47-b131ceba9df8] Took 39.04 seconds to build instance. [ 2086.073521] env[62824]: DEBUG oslo_vmware.api [None req-8854c3a5-43a5-4ba3-bce0-709d5c6fa611 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Task: {'id': task-2146272, 'name': ReconfigVM_Task, 'duration_secs': 0.173065} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2086.073895] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-8854c3a5-43a5-4ba3-bce0-709d5c6fa611 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-438620', 'volume_id': '5c4381e7-4dfa-4683-93fe-8eb8bef1df7e', 'name': 'volume-5c4381e7-4dfa-4683-93fe-8eb8bef1df7e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '59b5b883-4188-471c-8862-444f3ce08cb0', 'attached_at': '', 'detached_at': '', 'volume_id': '5c4381e7-4dfa-4683-93fe-8eb8bef1df7e', 'serial': '5c4381e7-4dfa-4683-93fe-8eb8bef1df7e'} {{(pid=62824) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 2086.094923] env[62824]: DEBUG oslo_vmware.api [None req-b1f093ef-801c-4886-991f-869391717179 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Task: {'id': task-2146273, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.172598} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2086.095299] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-b1f093ef-801c-4886-991f-869391717179 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2086.095531] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-b1f093ef-801c-4886-991f-869391717179 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 8bb90a4d-93c9-4f54-b15e-48fe966e6c97] Deleted contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2086.095634] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-b1f093ef-801c-4886-991f-869391717179 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 8bb90a4d-93c9-4f54-b15e-48fe966e6c97] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2086.095826] env[62824]: INFO nova.compute.manager [None req-b1f093ef-801c-4886-991f-869391717179 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] [instance: 8bb90a4d-93c9-4f54-b15e-48fe966e6c97] Took 1.74 seconds to destroy the instance on the hypervisor. [ 2086.096088] env[62824]: DEBUG oslo.service.loopingcall [None req-b1f093ef-801c-4886-991f-869391717179 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2086.096279] env[62824]: DEBUG nova.compute.manager [-] [instance: 8bb90a4d-93c9-4f54-b15e-48fe966e6c97] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2086.096380] env[62824]: DEBUG nova.network.neutron [-] [instance: 8bb90a4d-93c9-4f54-b15e-48fe966e6c97] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2086.214037] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 4dfe9f82-7f89-486b-b159-948656d2896c] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2086.278288] env[62824]: DEBUG nova.scheduler.client.report [None req-4715e6dc-1d1a-49a2-9d97-4f4b36f634e1 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2086.286075] env[62824]: INFO nova.compute.manager [-] [instance: dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda] Took 2.73 seconds to deallocate network for instance. [ 2086.296212] env[62824]: INFO nova.compute.manager [-] [instance: c36ece43-3d70-4e67-a740-9057f413c722] Took 2.28 seconds to deallocate network for instance. [ 2086.445220] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Acquiring lock "49818910-61de-4fb8-bbab-d5d61d2f1ada" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2086.446220] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Lock "49818910-61de-4fb8-bbab-d5d61d2f1ada" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2086.470102] env[62824]: DEBUG oslo_concurrency.lockutils [None req-afcb5b14-d051-4e34-bb65-a7ae97f5f453 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Lock "02f34565-6675-4c79-ac47-b131ceba9df8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 50.304s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2086.719138] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 384a96ea-40ce-43e1-a5f9-82f50b710b0b] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2086.784356] env[62824]: DEBUG oslo_concurrency.lockutils [None req-88bc485b-f4ba-418f-94b3-e4e74a1225cb tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquiring lock "cc8b6a13-da40-4c68-afc3-8b2c6ba574d4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2086.784600] env[62824]: DEBUG oslo_concurrency.lockutils [None req-88bc485b-f4ba-418f-94b3-e4e74a1225cb tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "cc8b6a13-da40-4c68-afc3-8b2c6ba574d4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2086.786307] env[62824]: DEBUG oslo_concurrency.lockutils [None req-4715e6dc-1d1a-49a2-9d97-4f4b36f634e1 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.323s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2086.789781] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.391s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2086.791466] env[62824]: INFO nova.compute.claims [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 350cfd12-a8d5-4a5f-b3de-d795fa179dfd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2086.798202] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fb4bd00b-e6f9-4459-a9db-0d3353b82d74 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2086.803329] env[62824]: DEBUG oslo_concurrency.lockutils [None req-56d14257-fc28-4784-a770-ae333154aa8d tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2086.813693] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ffe016d6-cd3f-454a-b9ac-22101a9a9bfd tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Acquiring lock "6a724332-a165-4a2b-9dd8-85f27e7b7637" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2086.813971] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ffe016d6-cd3f-454a-b9ac-22101a9a9bfd tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Lock "6a724332-a165-4a2b-9dd8-85f27e7b7637" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2086.814244] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ffe016d6-cd3f-454a-b9ac-22101a9a9bfd tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Acquiring lock "6a724332-a165-4a2b-9dd8-85f27e7b7637-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2086.814436] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ffe016d6-cd3f-454a-b9ac-22101a9a9bfd tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Lock "6a724332-a165-4a2b-9dd8-85f27e7b7637-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2086.814602] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ffe016d6-cd3f-454a-b9ac-22101a9a9bfd tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Lock "6a724332-a165-4a2b-9dd8-85f27e7b7637-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2086.818978] env[62824]: INFO nova.scheduler.client.report [None req-4715e6dc-1d1a-49a2-9d97-4f4b36f634e1 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Deleted allocations for instance 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2 [ 2086.823709] env[62824]: INFO nova.compute.manager [None req-ffe016d6-cd3f-454a-b9ac-22101a9a9bfd tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 6a724332-a165-4a2b-9dd8-85f27e7b7637] Terminating instance [ 2086.952849] env[62824]: DEBUG nova.compute.manager [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 49818910-61de-4fb8-bbab-d5d61d2f1ada] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2087.156545] env[62824]: DEBUG nova.network.neutron [-] [instance: 8bb90a4d-93c9-4f54-b15e-48fe966e6c97] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2087.224797] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 9062e606-511f-4fe7-9621-90b3c0e51eec] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2087.290646] env[62824]: DEBUG nova.compute.manager [None req-88bc485b-f4ba-418f-94b3-e4e74a1225cb tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: cc8b6a13-da40-4c68-afc3-8b2c6ba574d4] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2087.329314] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ffe016d6-cd3f-454a-b9ac-22101a9a9bfd tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Acquiring lock "refresh_cache-6a724332-a165-4a2b-9dd8-85f27e7b7637" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2087.329314] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ffe016d6-cd3f-454a-b9ac-22101a9a9bfd tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Acquired lock "refresh_cache-6a724332-a165-4a2b-9dd8-85f27e7b7637" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2087.329314] env[62824]: DEBUG nova.network.neutron [None req-ffe016d6-cd3f-454a-b9ac-22101a9a9bfd tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 6a724332-a165-4a2b-9dd8-85f27e7b7637] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2087.334106] env[62824]: DEBUG oslo_concurrency.lockutils [None req-4715e6dc-1d1a-49a2-9d97-4f4b36f634e1 tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Lock "387b7f14-f3c1-43ed-8c4e-6de3ce9822e2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.300s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2087.367308] env[62824]: DEBUG nova.compute.manager [req-4816fdbd-ce11-4953-97d7-f649da4de6a3 req-6570d9a3-ef6e-4029-9c6c-8037c9b2a87a service nova] [instance: dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda] Received event network-vif-deleted-9f1b3f36-9c69-4e99-8030-2d2ec535b324 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2087.367528] env[62824]: DEBUG nova.compute.manager [req-4816fdbd-ce11-4953-97d7-f649da4de6a3 req-6570d9a3-ef6e-4029-9c6c-8037c9b2a87a service nova] [instance: c36ece43-3d70-4e67-a740-9057f413c722] Received event network-vif-deleted-eb387e38-7606-4160-a960-43d51598b964 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2087.484821] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2087.659446] env[62824]: INFO nova.compute.manager [-] [instance: 8bb90a4d-93c9-4f54-b15e-48fe966e6c97] Took 1.56 seconds to deallocate network for instance. [ 2087.690505] env[62824]: DEBUG nova.compute.manager [req-25cf1ed6-5b9d-43c8-b25a-0142a6253ddd req-9f03dff3-3383-406a-a658-cb44c83e9f30 service nova] [instance: 15479062-af75-4925-99b3-77d6a49751ad] Received event network-changed-6b8b3a2d-f55e-4aa8-b995-4880c1adea60 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2087.691151] env[62824]: DEBUG nova.compute.manager [req-25cf1ed6-5b9d-43c8-b25a-0142a6253ddd req-9f03dff3-3383-406a-a658-cb44c83e9f30 service nova] [instance: 15479062-af75-4925-99b3-77d6a49751ad] Refreshing instance network info cache due to event network-changed-6b8b3a2d-f55e-4aa8-b995-4880c1adea60. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2087.691420] env[62824]: DEBUG oslo_concurrency.lockutils [req-25cf1ed6-5b9d-43c8-b25a-0142a6253ddd req-9f03dff3-3383-406a-a658-cb44c83e9f30 service nova] Acquiring lock "refresh_cache-15479062-af75-4925-99b3-77d6a49751ad" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2087.691826] env[62824]: DEBUG oslo_concurrency.lockutils [req-25cf1ed6-5b9d-43c8-b25a-0142a6253ddd req-9f03dff3-3383-406a-a658-cb44c83e9f30 service nova] Acquired lock "refresh_cache-15479062-af75-4925-99b3-77d6a49751ad" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2087.691826] env[62824]: DEBUG nova.network.neutron [req-25cf1ed6-5b9d-43c8-b25a-0142a6253ddd req-9f03dff3-3383-406a-a658-cb44c83e9f30 service nova] [instance: 15479062-af75-4925-99b3-77d6a49751ad] Refreshing network info cache for port 6b8b3a2d-f55e-4aa8-b995-4880c1adea60 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2087.731056] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 517bf2cf-e142-4f36-bf53-79a8d2d8a530] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2087.826822] env[62824]: DEBUG oslo_concurrency.lockutils [None req-88bc485b-f4ba-418f-94b3-e4e74a1225cb tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2087.857727] env[62824]: DEBUG oslo_concurrency.lockutils [None req-925f1353-8f4d-4365-a22f-f1b4ef6bffd3 tempest-ServersAdminTestJSON-510601504 tempest-ServersAdminTestJSON-510601504-project-admin] Acquiring lock "refresh_cache-02f34565-6675-4c79-ac47-b131ceba9df8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2087.857727] env[62824]: DEBUG oslo_concurrency.lockutils [None req-925f1353-8f4d-4365-a22f-f1b4ef6bffd3 tempest-ServersAdminTestJSON-510601504 tempest-ServersAdminTestJSON-510601504-project-admin] Acquired lock "refresh_cache-02f34565-6675-4c79-ac47-b131ceba9df8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2087.857727] env[62824]: DEBUG nova.network.neutron [None req-925f1353-8f4d-4365-a22f-f1b4ef6bffd3 tempest-ServersAdminTestJSON-510601504 tempest-ServersAdminTestJSON-510601504-project-admin] [instance: 02f34565-6675-4c79-ac47-b131ceba9df8] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2087.876399] env[62824]: DEBUG nova.network.neutron [None req-ffe016d6-cd3f-454a-b9ac-22101a9a9bfd tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 6a724332-a165-4a2b-9dd8-85f27e7b7637] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2088.029723] env[62824]: DEBUG nova.network.neutron [None req-ffe016d6-cd3f-454a-b9ac-22101a9a9bfd tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 6a724332-a165-4a2b-9dd8-85f27e7b7637] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2088.168637] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b1f093ef-801c-4886-991f-869391717179 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2088.235337] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 9b1fbb7b-07f1-48bc-851f-d4c35f96cd6c] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2088.411651] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a4b5218-3eac-4eb7-81a8-867dde163f3f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2088.423602] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c24ba376-9dbe-44da-8d88-874b771b7960 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2088.459801] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cef969c9-126f-4d72-8e22-1f62f7249199 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2088.468984] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa87b488-ece5-4a7a-8e84-4c6567aad781 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2088.485059] env[62824]: DEBUG nova.compute.provider_tree [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2088.494483] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-a851343c-36b8-4f6b-b725-ef04164a199a tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 919e8a45-7810-4a8d-a1aa-5046b5ab059c] Volume attach. Driver type: vmdk {{(pid=62824) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 2088.494722] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-a851343c-36b8-4f6b-b725-ef04164a199a tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 919e8a45-7810-4a8d-a1aa-5046b5ab059c] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-438665', 'volume_id': '20c6ff5e-0a8b-4cf4-8bd3-92cb944dd7a3', 'name': 'volume-20c6ff5e-0a8b-4cf4-8bd3-92cb944dd7a3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '919e8a45-7810-4a8d-a1aa-5046b5ab059c', 'attached_at': '', 'detached_at': '', 'volume_id': '20c6ff5e-0a8b-4cf4-8bd3-92cb944dd7a3', 'serial': '20c6ff5e-0a8b-4cf4-8bd3-92cb944dd7a3'} {{(pid=62824) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 2088.495676] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5f06865-e7b5-48ef-b034-35bb03005bc0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2088.514249] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e80eb8e-3a2b-4cf1-9bfb-a8989e82a043 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2088.543731] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-a851343c-36b8-4f6b-b725-ef04164a199a tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 919e8a45-7810-4a8d-a1aa-5046b5ab059c] Reconfiguring VM instance instance-0000003b to attach disk [datastore2] volume-20c6ff5e-0a8b-4cf4-8bd3-92cb944dd7a3/volume-20c6ff5e-0a8b-4cf4-8bd3-92cb944dd7a3.vmdk or device None with type thin {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2088.543934] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ffe016d6-cd3f-454a-b9ac-22101a9a9bfd tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Releasing lock "refresh_cache-6a724332-a165-4a2b-9dd8-85f27e7b7637" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2088.544205] env[62824]: DEBUG nova.compute.manager [None req-ffe016d6-cd3f-454a-b9ac-22101a9a9bfd tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 6a724332-a165-4a2b-9dd8-85f27e7b7637] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2088.544476] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-ffe016d6-cd3f-454a-b9ac-22101a9a9bfd tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 6a724332-a165-4a2b-9dd8-85f27e7b7637] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2088.544653] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e6e886a9-e145-4a0f-b51b-66a250efc675 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2088.559375] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27a4821c-7c5e-4954-9185-38bbb0eb5de6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2088.569558] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-ffe016d6-cd3f-454a-b9ac-22101a9a9bfd tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 6a724332-a165-4a2b-9dd8-85f27e7b7637] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2088.570979] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c29c32a7-e52a-4e08-8a50-ae3f0089d2fa {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2088.572690] env[62824]: DEBUG oslo_vmware.api [None req-a851343c-36b8-4f6b-b725-ef04164a199a tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 2088.572690] env[62824]: value = "task-2146279" [ 2088.572690] env[62824]: _type = "Task" [ 2088.572690] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2088.578166] env[62824]: DEBUG oslo_vmware.api [None req-ffe016d6-cd3f-454a-b9ac-22101a9a9bfd tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Waiting for the task: (returnval){ [ 2088.578166] env[62824]: value = "task-2146280" [ 2088.578166] env[62824]: _type = "Task" [ 2088.578166] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2088.584514] env[62824]: DEBUG oslo_vmware.api [None req-a851343c-36b8-4f6b-b725-ef04164a199a tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146279, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2088.594277] env[62824]: DEBUG oslo_vmware.api [None req-ffe016d6-cd3f-454a-b9ac-22101a9a9bfd tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Task: {'id': task-2146280, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2088.724414] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1254224f-0bf3-41dd-a390-a83d0850a18a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Acquiring lock "bbf1440b-8681-48c8-a178-9a83b925c695" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2088.725726] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1254224f-0bf3-41dd-a390-a83d0850a18a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Lock "bbf1440b-8681-48c8-a178-9a83b925c695" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2088.725949] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1254224f-0bf3-41dd-a390-a83d0850a18a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Acquiring lock "bbf1440b-8681-48c8-a178-9a83b925c695-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2088.726184] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1254224f-0bf3-41dd-a390-a83d0850a18a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Lock "bbf1440b-8681-48c8-a178-9a83b925c695-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2088.726363] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1254224f-0bf3-41dd-a390-a83d0850a18a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Lock "bbf1440b-8681-48c8-a178-9a83b925c695-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2088.732087] env[62824]: INFO nova.compute.manager [None req-1254224f-0bf3-41dd-a390-a83d0850a18a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Terminating instance [ 2088.739233] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 04ac4ade-0bc8-4469-9b1b-f288b2f0a367] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2088.989510] env[62824]: DEBUG nova.scheduler.client.report [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2089.086434] env[62824]: DEBUG oslo_vmware.api [None req-a851343c-36b8-4f6b-b725-ef04164a199a tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146279, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2089.093183] env[62824]: DEBUG nova.objects.instance [None req-8854c3a5-43a5-4ba3-bce0-709d5c6fa611 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Lazy-loading 'flavor' on Instance uuid 59b5b883-4188-471c-8862-444f3ce08cb0 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2089.105833] env[62824]: DEBUG oslo_vmware.api [None req-ffe016d6-cd3f-454a-b9ac-22101a9a9bfd tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Task: {'id': task-2146280, 'name': PowerOffVM_Task, 'duration_secs': 0.193101} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2089.105833] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-ffe016d6-cd3f-454a-b9ac-22101a9a9bfd tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 6a724332-a165-4a2b-9dd8-85f27e7b7637] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2089.105833] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-ffe016d6-cd3f-454a-b9ac-22101a9a9bfd tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 6a724332-a165-4a2b-9dd8-85f27e7b7637] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2089.105833] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-483bb551-8e98-42f2-920a-95f40f7a9e8a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2089.137131] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-ffe016d6-cd3f-454a-b9ac-22101a9a9bfd tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 6a724332-a165-4a2b-9dd8-85f27e7b7637] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2089.140042] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-ffe016d6-cd3f-454a-b9ac-22101a9a9bfd tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 6a724332-a165-4a2b-9dd8-85f27e7b7637] Deleting contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2089.140042] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-ffe016d6-cd3f-454a-b9ac-22101a9a9bfd tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Deleting the datastore file [datastore2] 6a724332-a165-4a2b-9dd8-85f27e7b7637 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2089.140042] env[62824]: DEBUG nova.network.neutron [req-25cf1ed6-5b9d-43c8-b25a-0142a6253ddd req-9f03dff3-3383-406a-a658-cb44c83e9f30 service nova] [instance: 15479062-af75-4925-99b3-77d6a49751ad] Updated VIF entry in instance network info cache for port 6b8b3a2d-f55e-4aa8-b995-4880c1adea60. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2089.140042] env[62824]: DEBUG nova.network.neutron [req-25cf1ed6-5b9d-43c8-b25a-0142a6253ddd req-9f03dff3-3383-406a-a658-cb44c83e9f30 service nova] [instance: 15479062-af75-4925-99b3-77d6a49751ad] Updating instance_info_cache with network_info: [{"id": "6b8b3a2d-f55e-4aa8-b995-4880c1adea60", "address": "fa:16:3e:77:1e:5b", "network": {"id": "6f888a0c-8089-4dd4-aace-14b020e11bee", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-137042394-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.242", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d041a8209e3848eba8e47251e013ca17", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5f60c972-a72d-4c5f-a250-faadfd6eafbe", "external-id": "nsx-vlan-transportzone-932", "segmentation_id": 932, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6b8b3a2d-f5", "ovs_interfaceid": "6b8b3a2d-f55e-4aa8-b995-4880c1adea60", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2089.140042] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3dd11ad7-2407-4f2d-98fd-07c25fc66162 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2089.149427] env[62824]: DEBUG oslo_vmware.api [None req-ffe016d6-cd3f-454a-b9ac-22101a9a9bfd tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Waiting for the task: (returnval){ [ 2089.149427] env[62824]: value = "task-2146282" [ 2089.149427] env[62824]: _type = "Task" [ 2089.149427] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2089.164919] env[62824]: DEBUG oslo_vmware.api [None req-ffe016d6-cd3f-454a-b9ac-22101a9a9bfd tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Task: {'id': task-2146282, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2089.209658] env[62824]: DEBUG nova.network.neutron [None req-925f1353-8f4d-4365-a22f-f1b4ef6bffd3 tempest-ServersAdminTestJSON-510601504 tempest-ServersAdminTestJSON-510601504-project-admin] [instance: 02f34565-6675-4c79-ac47-b131ceba9df8] Updating instance_info_cache with network_info: [{"id": "be36fef6-17a7-4c25-bf33-98cc6a037b53", "address": "fa:16:3e:c1:60:d0", "network": {"id": "fc381ca5-198f-4afc-9ebf-bc6e3de68c35", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-425455451-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "248af2bc3e7b41d2a74b0c6074b359e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4c9a12d2-469f-4199-bfaa-f791d765deac", "external-id": "nsx-vlan-transportzone-96", "segmentation_id": 96, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbe36fef6-17", "ovs_interfaceid": "be36fef6-17a7-4c25-bf33-98cc6a037b53", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2089.239466] env[62824]: DEBUG nova.compute.manager [None req-1254224f-0bf3-41dd-a390-a83d0850a18a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2089.239466] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-1254224f-0bf3-41dd-a390-a83d0850a18a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2089.239466] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07c023b8-d34c-4686-aca6-209bc74e40d5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2089.243172] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: f4157385-43ea-4e8c-887c-0985af37abd1] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2089.249137] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-1254224f-0bf3-41dd-a390-a83d0850a18a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2089.249740] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ebd5f042-ae6d-498a-a115-51a20f05f549 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2089.259664] env[62824]: DEBUG oslo_vmware.api [None req-1254224f-0bf3-41dd-a390-a83d0850a18a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Waiting for the task: (returnval){ [ 2089.259664] env[62824]: value = "task-2146283" [ 2089.259664] env[62824]: _type = "Task" [ 2089.259664] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2089.276734] env[62824]: DEBUG oslo_vmware.api [None req-1254224f-0bf3-41dd-a390-a83d0850a18a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2146283, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2089.336235] env[62824]: DEBUG nova.network.neutron [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8] Successfully updated port: bfd88146-4384-4514-8ceb-46c647c80eee {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2089.496772] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.707s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2089.497391] env[62824]: DEBUG nova.compute.manager [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 350cfd12-a8d5-4a5f-b3de-d795fa179dfd] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2089.506307] env[62824]: DEBUG oslo_concurrency.lockutils [None req-423ea96c-8206-47c1-8b7e-a945775631af tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.995s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2089.506599] env[62824]: DEBUG nova.objects.instance [None req-423ea96c-8206-47c1-8b7e-a945775631af tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Lazy-loading 'resources' on Instance uuid 29b8dd5f-1855-490c-a01b-54840073a753 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2089.585383] env[62824]: DEBUG oslo_vmware.api [None req-a851343c-36b8-4f6b-b725-ef04164a199a tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146279, 'name': ReconfigVM_Task, 'duration_secs': 0.752081} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2089.585383] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-a851343c-36b8-4f6b-b725-ef04164a199a tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 919e8a45-7810-4a8d-a1aa-5046b5ab059c] Reconfigured VM instance instance-0000003b to attach disk [datastore2] volume-20c6ff5e-0a8b-4cf4-8bd3-92cb944dd7a3/volume-20c6ff5e-0a8b-4cf4-8bd3-92cb944dd7a3.vmdk or device None with type thin {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2089.591502] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-39b4860f-00dc-4bef-b2ce-a57b330351fc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2089.613840] env[62824]: DEBUG oslo_vmware.api [None req-a851343c-36b8-4f6b-b725-ef04164a199a tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 2089.613840] env[62824]: value = "task-2146284" [ 2089.613840] env[62824]: _type = "Task" [ 2089.613840] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2089.627323] env[62824]: DEBUG oslo_vmware.api [None req-a851343c-36b8-4f6b-b725-ef04164a199a tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146284, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2089.643469] env[62824]: DEBUG oslo_concurrency.lockutils [req-25cf1ed6-5b9d-43c8-b25a-0142a6253ddd req-9f03dff3-3383-406a-a658-cb44c83e9f30 service nova] Releasing lock "refresh_cache-15479062-af75-4925-99b3-77d6a49751ad" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2089.662524] env[62824]: DEBUG oslo_vmware.api [None req-ffe016d6-cd3f-454a-b9ac-22101a9a9bfd tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Task: {'id': task-2146282, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.114055} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2089.662806] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-ffe016d6-cd3f-454a-b9ac-22101a9a9bfd tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2089.662995] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-ffe016d6-cd3f-454a-b9ac-22101a9a9bfd tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 6a724332-a165-4a2b-9dd8-85f27e7b7637] Deleted contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2089.663192] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-ffe016d6-cd3f-454a-b9ac-22101a9a9bfd tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 6a724332-a165-4a2b-9dd8-85f27e7b7637] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2089.663366] env[62824]: INFO nova.compute.manager [None req-ffe016d6-cd3f-454a-b9ac-22101a9a9bfd tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] [instance: 6a724332-a165-4a2b-9dd8-85f27e7b7637] Took 1.12 seconds to destroy the instance on the hypervisor. [ 2089.663608] env[62824]: DEBUG oslo.service.loopingcall [None req-ffe016d6-cd3f-454a-b9ac-22101a9a9bfd tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2089.663800] env[62824]: DEBUG nova.compute.manager [-] [instance: 6a724332-a165-4a2b-9dd8-85f27e7b7637] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2089.663898] env[62824]: DEBUG nova.network.neutron [-] [instance: 6a724332-a165-4a2b-9dd8-85f27e7b7637] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2089.694090] env[62824]: DEBUG nova.network.neutron [-] [instance: 6a724332-a165-4a2b-9dd8-85f27e7b7637] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2089.716266] env[62824]: DEBUG oslo_concurrency.lockutils [None req-925f1353-8f4d-4365-a22f-f1b4ef6bffd3 tempest-ServersAdminTestJSON-510601504 tempest-ServersAdminTestJSON-510601504-project-admin] Releasing lock "refresh_cache-02f34565-6675-4c79-ac47-b131ceba9df8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2089.716266] env[62824]: DEBUG nova.compute.manager [None req-925f1353-8f4d-4365-a22f-f1b4ef6bffd3 tempest-ServersAdminTestJSON-510601504 tempest-ServersAdminTestJSON-510601504-project-admin] [instance: 02f34565-6675-4c79-ac47-b131ceba9df8] Inject network info {{(pid=62824) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7737}} [ 2089.716707] env[62824]: DEBUG nova.compute.manager [None req-925f1353-8f4d-4365-a22f-f1b4ef6bffd3 tempest-ServersAdminTestJSON-510601504 tempest-ServersAdminTestJSON-510601504-project-admin] [instance: 02f34565-6675-4c79-ac47-b131ceba9df8] network_info to inject: |[{"id": "be36fef6-17a7-4c25-bf33-98cc6a037b53", "address": "fa:16:3e:c1:60:d0", "network": {"id": "fc381ca5-198f-4afc-9ebf-bc6e3de68c35", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-425455451-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "248af2bc3e7b41d2a74b0c6074b359e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4c9a12d2-469f-4199-bfaa-f791d765deac", "external-id": "nsx-vlan-transportzone-96", "segmentation_id": 96, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbe36fef6-17", "ovs_interfaceid": "be36fef6-17a7-4c25-bf33-98cc6a037b53", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7738}} [ 2089.721738] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-925f1353-8f4d-4365-a22f-f1b4ef6bffd3 tempest-ServersAdminTestJSON-510601504 tempest-ServersAdminTestJSON-510601504-project-admin] [instance: 02f34565-6675-4c79-ac47-b131ceba9df8] Reconfiguring VM instance to set the machine id {{(pid=62824) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1796}} [ 2089.722387] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-afbb6502-7b8e-4e42-86ad-5ca049ebbf7a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2089.740833] env[62824]: DEBUG oslo_vmware.api [None req-925f1353-8f4d-4365-a22f-f1b4ef6bffd3 tempest-ServersAdminTestJSON-510601504 tempest-ServersAdminTestJSON-510601504-project-admin] Waiting for the task: (returnval){ [ 2089.740833] env[62824]: value = "task-2146285" [ 2089.740833] env[62824]: _type = "Task" [ 2089.740833] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2090.600318] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2090.601901] env[62824]: DEBUG oslo_concurrency.lockutils [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Acquiring lock "refresh_cache-1870b82a-9783-44ac-8de2-7b9ffc2a1bc8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2090.602045] env[62824]: DEBUG oslo_concurrency.lockutils [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Acquired lock "refresh_cache-1870b82a-9783-44ac-8de2-7b9ffc2a1bc8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2090.602198] env[62824]: DEBUG nova.network.neutron [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2090.606403] env[62824]: DEBUG nova.compute.utils [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2090.607915] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8854c3a5-43a5-4ba3-bce0-709d5c6fa611 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Lock "59b5b883-4188-471c-8862-444f3ce08cb0" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 6.337s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2090.609733] env[62824]: DEBUG nova.network.neutron [-] [instance: 6a724332-a165-4a2b-9dd8-85f27e7b7637] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2090.620241] env[62824]: DEBUG oslo_vmware.api [None req-925f1353-8f4d-4365-a22f-f1b4ef6bffd3 tempest-ServersAdminTestJSON-510601504 tempest-ServersAdminTestJSON-510601504-project-admin] Task: {'id': task-2146285, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2090.620454] env[62824]: WARNING oslo_vmware.common.loopingcall [None req-925f1353-8f4d-4365-a22f-f1b4ef6bffd3 tempest-ServersAdminTestJSON-510601504 tempest-ServersAdminTestJSON-510601504-project-admin] task run outlasted interval by 0.37936499999999995 sec [ 2090.627050] env[62824]: DEBUG nova.compute.manager [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 350cfd12-a8d5-4a5f-b3de-d795fa179dfd] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2090.627050] env[62824]: DEBUG nova.network.neutron [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 350cfd12-a8d5-4a5f-b3de-d795fa179dfd] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2090.639982] env[62824]: DEBUG nova.compute.manager [req-fa2d094e-0159-41d5-9146-13b59197f734 req-a0090131-2fe5-4ea5-ba7b-f7d4a7dac18d service nova] [instance: 8bb90a4d-93c9-4f54-b15e-48fe966e6c97] Received event network-vif-deleted-ec4e83ca-0312-4fc5-8c4f-475c5235bd4e {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2090.642201] env[62824]: DEBUG nova.compute.manager [req-3c896331-a3aa-4dc2-a133-124f14f964d0 req-819be90a-964c-43bb-a59f-b1b46ad5be89 service nova] [instance: 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8] Received event network-vif-plugged-bfd88146-4384-4514-8ceb-46c647c80eee {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2090.642452] env[62824]: DEBUG oslo_concurrency.lockutils [req-3c896331-a3aa-4dc2-a133-124f14f964d0 req-819be90a-964c-43bb-a59f-b1b46ad5be89 service nova] Acquiring lock "1870b82a-9783-44ac-8de2-7b9ffc2a1bc8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2090.642567] env[62824]: DEBUG oslo_concurrency.lockutils [req-3c896331-a3aa-4dc2-a133-124f14f964d0 req-819be90a-964c-43bb-a59f-b1b46ad5be89 service nova] Lock "1870b82a-9783-44ac-8de2-7b9ffc2a1bc8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2090.642724] env[62824]: DEBUG oslo_concurrency.lockutils [req-3c896331-a3aa-4dc2-a133-124f14f964d0 req-819be90a-964c-43bb-a59f-b1b46ad5be89 service nova] Lock "1870b82a-9783-44ac-8de2-7b9ffc2a1bc8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2090.642878] env[62824]: DEBUG nova.compute.manager [req-3c896331-a3aa-4dc2-a133-124f14f964d0 req-819be90a-964c-43bb-a59f-b1b46ad5be89 service nova] [instance: 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8] No waiting events found dispatching network-vif-plugged-bfd88146-4384-4514-8ceb-46c647c80eee {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2090.643053] env[62824]: WARNING nova.compute.manager [req-3c896331-a3aa-4dc2-a133-124f14f964d0 req-819be90a-964c-43bb-a59f-b1b46ad5be89 service nova] [instance: 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8] Received unexpected event network-vif-plugged-bfd88146-4384-4514-8ceb-46c647c80eee for instance with vm_state building and task_state spawning. [ 2090.643206] env[62824]: DEBUG nova.compute.manager [req-3c896331-a3aa-4dc2-a133-124f14f964d0 req-819be90a-964c-43bb-a59f-b1b46ad5be89 service nova] [instance: 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8] Received event network-changed-bfd88146-4384-4514-8ceb-46c647c80eee {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2090.643348] env[62824]: DEBUG nova.compute.manager [req-3c896331-a3aa-4dc2-a133-124f14f964d0 req-819be90a-964c-43bb-a59f-b1b46ad5be89 service nova] [instance: 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8] Refreshing instance network info cache due to event network-changed-bfd88146-4384-4514-8ceb-46c647c80eee. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2090.643613] env[62824]: DEBUG oslo_concurrency.lockutils [req-3c896331-a3aa-4dc2-a133-124f14f964d0 req-819be90a-964c-43bb-a59f-b1b46ad5be89 service nova] Acquiring lock "refresh_cache-1870b82a-9783-44ac-8de2-7b9ffc2a1bc8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2090.644950] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquiring lock "bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2090.645176] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2090.646131] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Acquiring lock "29140542-dc4b-411a-ac10-7d84086eabbb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2090.646420] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Lock "29140542-dc4b-411a-ac10-7d84086eabbb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2090.649842] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Acquiring lock "4c2c2068-664d-404f-a99e-8fc7719f43e5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2090.649842] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Lock "4c2c2068-664d-404f-a99e-8fc7719f43e5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2090.649842] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Acquiring lock "d067fa97-cedc-4e3d-9be4-d860a79a7723" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2090.650118] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Lock "d067fa97-cedc-4e3d-9be4-d860a79a7723" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2090.666365] env[62824]: DEBUG oslo_vmware.api [None req-1254224f-0bf3-41dd-a390-a83d0850a18a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2146283, 'name': PowerOffVM_Task, 'duration_secs': 0.314892} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2090.666565] env[62824]: DEBUG oslo_vmware.api [None req-a851343c-36b8-4f6b-b725-ef04164a199a tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146284, 'name': ReconfigVM_Task, 'duration_secs': 0.19057} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2090.670352] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-1254224f-0bf3-41dd-a390-a83d0850a18a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2090.670532] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-1254224f-0bf3-41dd-a390-a83d0850a18a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2090.671019] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-a851343c-36b8-4f6b-b725-ef04164a199a tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 919e8a45-7810-4a8d-a1aa-5046b5ab059c] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-438665', 'volume_id': '20c6ff5e-0a8b-4cf4-8bd3-92cb944dd7a3', 'name': 'volume-20c6ff5e-0a8b-4cf4-8bd3-92cb944dd7a3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '919e8a45-7810-4a8d-a1aa-5046b5ab059c', 'attached_at': '', 'detached_at': '', 'volume_id': '20c6ff5e-0a8b-4cf4-8bd3-92cb944dd7a3', 'serial': '20c6ff5e-0a8b-4cf4-8bd3-92cb944dd7a3'} {{(pid=62824) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 2090.673257] env[62824]: DEBUG oslo_vmware.api [None req-925f1353-8f4d-4365-a22f-f1b4ef6bffd3 tempest-ServersAdminTestJSON-510601504 tempest-ServersAdminTestJSON-510601504-project-admin] Task: {'id': task-2146285, 'name': ReconfigVM_Task, 'duration_secs': 0.179243} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2090.673380] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dde17e2c-73c8-4ed6-8aaa-6bbabc2a955b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2090.675391] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-925f1353-8f4d-4365-a22f-f1b4ef6bffd3 tempest-ServersAdminTestJSON-510601504 tempest-ServersAdminTestJSON-510601504-project-admin] [instance: 02f34565-6675-4c79-ac47-b131ceba9df8] Reconfigured VM instance to set the machine id {{(pid=62824) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1799}} [ 2090.775081] env[62824]: DEBUG nova.policy [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd64a0b7abab54b15a91ec809b0cf87f3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2e59203d676d48c8bd8402af4eb58f65', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2090.778699] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-1254224f-0bf3-41dd-a390-a83d0850a18a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2090.779073] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-1254224f-0bf3-41dd-a390-a83d0850a18a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Deleting contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2090.779332] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-1254224f-0bf3-41dd-a390-a83d0850a18a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Deleting the datastore file [datastore1] bbf1440b-8681-48c8-a178-9a83b925c695 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2090.779563] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0f8cee10-4bec-4895-b657-239b49b44345 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2090.786977] env[62824]: DEBUG oslo_vmware.api [None req-1254224f-0bf3-41dd-a390-a83d0850a18a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Waiting for the task: (returnval){ [ 2090.786977] env[62824]: value = "task-2146287" [ 2090.786977] env[62824]: _type = "Task" [ 2090.786977] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2090.796651] env[62824]: DEBUG oslo_vmware.api [None req-1254224f-0bf3-41dd-a390-a83d0850a18a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2146287, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2091.123503] env[62824]: DEBUG nova.compute.manager [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 350cfd12-a8d5-4a5f-b3de-d795fa179dfd] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2091.146543] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d43d8aff-0168-42f7-936a-24266643a8a6 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Acquiring lock "59b5b883-4188-471c-8862-444f3ce08cb0" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2091.150658] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d43d8aff-0168-42f7-936a-24266643a8a6 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Lock "59b5b883-4188-471c-8862-444f3ce08cb0" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.004s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2091.150658] env[62824]: DEBUG nova.compute.manager [None req-d43d8aff-0168-42f7-936a-24266643a8a6 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2091.155183] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bb4fa96-ae7c-4cc2-b81a-6351e0e8f9d8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2091.158191] env[62824]: DEBUG nova.compute.manager [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2091.161923] env[62824]: DEBUG nova.compute.manager [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 29140542-dc4b-411a-ac10-7d84086eabbb] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2091.163584] env[62824]: DEBUG nova.compute.manager [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 4c2c2068-664d-404f-a99e-8fc7719f43e5] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2091.170615] env[62824]: DEBUG nova.compute.manager [None req-d43d8aff-0168-42f7-936a-24266643a8a6 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62824) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3404}} [ 2091.171210] env[62824]: DEBUG nova.objects.instance [None req-d43d8aff-0168-42f7-936a-24266643a8a6 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Lazy-loading 'flavor' on Instance uuid 59b5b883-4188-471c-8862-444f3ce08cb0 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2091.175122] env[62824]: DEBUG nova.network.neutron [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2091.186666] env[62824]: INFO nova.compute.manager [-] [instance: 6a724332-a165-4a2b-9dd8-85f27e7b7637] Took 1.52 seconds to deallocate network for instance. [ 2091.224537] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e045cf7e-d29d-4cf8-9d46-4c92f353def0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2091.237115] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5716434e-a687-4e7c-8ce2-ddd04d152d5d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2091.280018] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e294e312-ef01-4475-ad52-b57a52adaafe {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2091.294420] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c760a438-7aa1-46ae-8eec-fed340a908a9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2091.307682] env[62824]: DEBUG oslo_vmware.api [None req-1254224f-0bf3-41dd-a390-a83d0850a18a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Task: {'id': task-2146287, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.14659} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2091.315273] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-1254224f-0bf3-41dd-a390-a83d0850a18a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2091.315606] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-1254224f-0bf3-41dd-a390-a83d0850a18a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Deleted contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2091.315681] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-1254224f-0bf3-41dd-a390-a83d0850a18a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2091.315883] env[62824]: INFO nova.compute.manager [None req-1254224f-0bf3-41dd-a390-a83d0850a18a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Took 2.08 seconds to destroy the instance on the hypervisor. [ 2091.316116] env[62824]: DEBUG oslo.service.loopingcall [None req-1254224f-0bf3-41dd-a390-a83d0850a18a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2091.316938] env[62824]: DEBUG nova.compute.provider_tree [None req-423ea96c-8206-47c1-8b7e-a945775631af tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2091.320526] env[62824]: DEBUG nova.compute.manager [-] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2091.320632] env[62824]: DEBUG nova.network.neutron [-] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2091.435457] env[62824]: DEBUG nova.network.neutron [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8] Updating instance_info_cache with network_info: [{"id": "bfd88146-4384-4514-8ceb-46c647c80eee", "address": "fa:16:3e:91:99:2b", "network": {"id": "aeacc588-ae73-49ec-9aa1-68f7d92055be", "bridge": "br-int", "label": "tempest-ImagesTestJSON-314094063-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10ee5fe93d584725ace90e76605a1b2e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1e7a4976-597e-4636-990e-6062b5faadee", "external-id": "nsx-vlan-transportzone-847", "segmentation_id": 847, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbfd88146-43", "ovs_interfaceid": "bfd88146-4384-4514-8ceb-46c647c80eee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2091.508127] env[62824]: DEBUG nova.network.neutron [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 350cfd12-a8d5-4a5f-b3de-d795fa179dfd] Successfully created port: 75b0f475-5e6e-44de-9480-a71716ecb4e0 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2091.697386] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ffe016d6-cd3f-454a-b9ac-22101a9a9bfd tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2091.702155] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2091.703787] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2091.720127] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2091.730932] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Acquiring lock "dadf590a-7288-4dd2-90de-125fe272f4ba" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2091.731335] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Lock "dadf590a-7288-4dd2-90de-125fe272f4ba" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2091.746929] env[62824]: DEBUG nova.objects.instance [None req-a851343c-36b8-4f6b-b725-ef04164a199a tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lazy-loading 'flavor' on Instance uuid 919e8a45-7810-4a8d-a1aa-5046b5ab059c {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2091.825078] env[62824]: DEBUG nova.scheduler.client.report [None req-423ea96c-8206-47c1-8b7e-a945775631af tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2091.939540] env[62824]: DEBUG oslo_concurrency.lockutils [None req-123477fc-8c3e-4f2a-83de-cced4f011020 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2091.942545] env[62824]: DEBUG oslo_concurrency.lockutils [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Releasing lock "refresh_cache-1870b82a-9783-44ac-8de2-7b9ffc2a1bc8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2091.942911] env[62824]: DEBUG nova.compute.manager [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8] Instance network_info: |[{"id": "bfd88146-4384-4514-8ceb-46c647c80eee", "address": "fa:16:3e:91:99:2b", "network": {"id": "aeacc588-ae73-49ec-9aa1-68f7d92055be", "bridge": "br-int", "label": "tempest-ImagesTestJSON-314094063-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10ee5fe93d584725ace90e76605a1b2e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1e7a4976-597e-4636-990e-6062b5faadee", "external-id": "nsx-vlan-transportzone-847", "segmentation_id": 847, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbfd88146-43", "ovs_interfaceid": "bfd88146-4384-4514-8ceb-46c647c80eee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2091.943229] env[62824]: DEBUG oslo_concurrency.lockutils [req-3c896331-a3aa-4dc2-a133-124f14f964d0 req-819be90a-964c-43bb-a59f-b1b46ad5be89 service nova] Acquired lock "refresh_cache-1870b82a-9783-44ac-8de2-7b9ffc2a1bc8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2091.943405] env[62824]: DEBUG nova.network.neutron [req-3c896331-a3aa-4dc2-a133-124f14f964d0 req-819be90a-964c-43bb-a59f-b1b46ad5be89 service nova] [instance: 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8] Refreshing network info cache for port bfd88146-4384-4514-8ceb-46c647c80eee {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2091.944644] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:91:99:2b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1e7a4976-597e-4636-990e-6062b5faadee', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bfd88146-4384-4514-8ceb-46c647c80eee', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2091.952208] env[62824]: DEBUG oslo.service.loopingcall [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2091.952693] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2091.952976] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-451e602d-2bbb-4861-bdf0-1d1010508639 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2091.976058] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2091.976058] env[62824]: value = "task-2146288" [ 2091.976058] env[62824]: _type = "Task" [ 2091.976058] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2091.982943] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146288, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2092.138817] env[62824]: DEBUG nova.compute.manager [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 350cfd12-a8d5-4a5f-b3de-d795fa179dfd] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2092.166162] env[62824]: DEBUG nova.virt.hardware [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2092.166432] env[62824]: DEBUG nova.virt.hardware [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2092.166590] env[62824]: DEBUG nova.virt.hardware [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2092.166770] env[62824]: DEBUG nova.virt.hardware [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2092.166914] env[62824]: DEBUG nova.virt.hardware [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2092.167165] env[62824]: DEBUG nova.virt.hardware [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2092.167403] env[62824]: DEBUG nova.virt.hardware [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2092.167562] env[62824]: DEBUG nova.virt.hardware [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2092.167725] env[62824]: DEBUG nova.virt.hardware [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2092.167886] env[62824]: DEBUG nova.virt.hardware [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2092.168083] env[62824]: DEBUG nova.virt.hardware [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2092.168960] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c42100d-8f78-46ec-b783-9053347cd20d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2092.177610] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e612c76d-29aa-42e5-960e-a539fa2a8217 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2092.191714] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d43d8aff-0168-42f7-936a-24266643a8a6 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2092.192232] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8346fe8e-7df2-44fb-a400-a86a3fdb595e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2092.200466] env[62824]: DEBUG oslo_vmware.api [None req-d43d8aff-0168-42f7-936a-24266643a8a6 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Waiting for the task: (returnval){ [ 2092.200466] env[62824]: value = "task-2146289" [ 2092.200466] env[62824]: _type = "Task" [ 2092.200466] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2092.209134] env[62824]: DEBUG oslo_vmware.api [None req-d43d8aff-0168-42f7-936a-24266643a8a6 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Task: {'id': task-2146289, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2092.252964] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a851343c-36b8-4f6b-b725-ef04164a199a tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "919e8a45-7810-4a8d-a1aa-5046b5ab059c" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.418s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2092.261108] env[62824]: DEBUG nova.network.neutron [-] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2092.333849] env[62824]: DEBUG oslo_concurrency.lockutils [None req-423ea96c-8206-47c1-8b7e-a945775631af tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.828s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2092.337298] env[62824]: DEBUG oslo_concurrency.lockutils [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.241s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2092.339210] env[62824]: INFO nova.compute.claims [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2092.366018] env[62824]: INFO nova.scheduler.client.report [None req-423ea96c-8206-47c1-8b7e-a945775631af tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Deleted allocations for instance 29b8dd5f-1855-490c-a01b-54840073a753 [ 2092.485749] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146288, 'name': CreateVM_Task, 'duration_secs': 0.430313} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2092.486230] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2092.487198] env[62824]: DEBUG oslo_concurrency.lockutils [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2092.487505] env[62824]: DEBUG oslo_concurrency.lockutils [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2092.487901] env[62824]: DEBUG oslo_concurrency.lockutils [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2092.488254] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a559603b-e861-40c3-bbef-f9357a8d1802 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2092.495131] env[62824]: DEBUG oslo_vmware.api [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the task: (returnval){ [ 2092.495131] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52b1f31e-6eb2-ed4b-e904-f837f8e0fc89" [ 2092.495131] env[62824]: _type = "Task" [ 2092.495131] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2092.505810] env[62824]: DEBUG oslo_vmware.api [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52b1f31e-6eb2-ed4b-e904-f837f8e0fc89, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2092.618261] env[62824]: DEBUG nova.compute.manager [req-da2e1595-65a1-4db0-adc7-3bcf222d5b56 req-24e26783-a976-4a27-930e-1361b37bd4cf service nova] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Received event network-vif-deleted-bd32bec1-cb61-4462-8231-5f4a4ec2ed9d {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2092.713022] env[62824]: DEBUG oslo_vmware.api [None req-d43d8aff-0168-42f7-936a-24266643a8a6 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Task: {'id': task-2146289, 'name': PowerOffVM_Task, 'duration_secs': 0.211015} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2092.714182] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d43d8aff-0168-42f7-936a-24266643a8a6 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2092.714395] env[62824]: DEBUG nova.compute.manager [None req-d43d8aff-0168-42f7-936a-24266643a8a6 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2092.715211] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d40cee3d-994d-408d-9aa1-2b34fac730c7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2092.719563] env[62824]: DEBUG nova.network.neutron [req-3c896331-a3aa-4dc2-a133-124f14f964d0 req-819be90a-964c-43bb-a59f-b1b46ad5be89 service nova] [instance: 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8] Updated VIF entry in instance network info cache for port bfd88146-4384-4514-8ceb-46c647c80eee. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2092.719946] env[62824]: DEBUG nova.network.neutron [req-3c896331-a3aa-4dc2-a133-124f14f964d0 req-819be90a-964c-43bb-a59f-b1b46ad5be89 service nova] [instance: 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8] Updating instance_info_cache with network_info: [{"id": "bfd88146-4384-4514-8ceb-46c647c80eee", "address": "fa:16:3e:91:99:2b", "network": {"id": "aeacc588-ae73-49ec-9aa1-68f7d92055be", "bridge": "br-int", "label": "tempest-ImagesTestJSON-314094063-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10ee5fe93d584725ace90e76605a1b2e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1e7a4976-597e-4636-990e-6062b5faadee", "external-id": "nsx-vlan-transportzone-847", "segmentation_id": 847, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbfd88146-43", "ovs_interfaceid": "bfd88146-4384-4514-8ceb-46c647c80eee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2092.766248] env[62824]: INFO nova.compute.manager [-] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Took 1.45 seconds to deallocate network for instance. [ 2092.874714] env[62824]: DEBUG oslo_concurrency.lockutils [None req-423ea96c-8206-47c1-8b7e-a945775631af tempest-VolumesAssistedSnapshotsTest-586662422 tempest-VolumesAssistedSnapshotsTest-586662422-project-member] Lock "29b8dd5f-1855-490c-a01b-54840073a753" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.528s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2093.007552] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Acquiring lock "7a2a51c4-e558-4f5f-b82c-718bc12c1df5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2093.007552] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Lock "7a2a51c4-e558-4f5f-b82c-718bc12c1df5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2093.015971] env[62824]: DEBUG oslo_vmware.api [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52b1f31e-6eb2-ed4b-e904-f837f8e0fc89, 'name': SearchDatastore_Task, 'duration_secs': 0.010833} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2093.016526] env[62824]: DEBUG oslo_concurrency.lockutils [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2093.017165] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2093.017398] env[62824]: DEBUG oslo_concurrency.lockutils [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2093.017528] env[62824]: DEBUG oslo_concurrency.lockutils [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2093.017803] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2093.018377] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-aa4d43ea-6f28-493c-a941-09f08decd770 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2093.028072] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2093.028292] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2093.029048] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2b2a1501-64b4-4a16-9705-5191d6d590be {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2093.037729] env[62824]: DEBUG oslo_vmware.api [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the task: (returnval){ [ 2093.037729] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52cc788e-69f0-bd2c-7703-d2ca66257d5b" [ 2093.037729] env[62824]: _type = "Task" [ 2093.037729] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2093.047946] env[62824]: DEBUG oslo_vmware.api [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52cc788e-69f0-bd2c-7703-d2ca66257d5b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2093.053620] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Acquiring lock "20f93b46-5e7e-4256-8cc1-e0a1b16740d5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2093.053929] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Lock "20f93b46-5e7e-4256-8cc1-e0a1b16740d5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2093.195405] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b5cb0b06-e659-44b5-a4c3-505f639c37f7 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquiring lock "919e8a45-7810-4a8d-a1aa-5046b5ab059c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2093.195752] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b5cb0b06-e659-44b5-a4c3-505f639c37f7 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "919e8a45-7810-4a8d-a1aa-5046b5ab059c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2093.196138] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b5cb0b06-e659-44b5-a4c3-505f639c37f7 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquiring lock "919e8a45-7810-4a8d-a1aa-5046b5ab059c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2093.196450] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b5cb0b06-e659-44b5-a4c3-505f639c37f7 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "919e8a45-7810-4a8d-a1aa-5046b5ab059c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2093.196689] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b5cb0b06-e659-44b5-a4c3-505f639c37f7 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "919e8a45-7810-4a8d-a1aa-5046b5ab059c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2093.199153] env[62824]: INFO nova.compute.manager [None req-b5cb0b06-e659-44b5-a4c3-505f639c37f7 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 919e8a45-7810-4a8d-a1aa-5046b5ab059c] Terminating instance [ 2093.226723] env[62824]: DEBUG oslo_concurrency.lockutils [req-3c896331-a3aa-4dc2-a133-124f14f964d0 req-819be90a-964c-43bb-a59f-b1b46ad5be89 service nova] Releasing lock "refresh_cache-1870b82a-9783-44ac-8de2-7b9ffc2a1bc8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2093.228617] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d43d8aff-0168-42f7-936a-24266643a8a6 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Lock "59b5b883-4188-471c-8862-444f3ce08cb0" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.078s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2093.273115] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1254224f-0bf3-41dd-a390-a83d0850a18a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2093.469890] env[62824]: DEBUG nova.network.neutron [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 350cfd12-a8d5-4a5f-b3de-d795fa179dfd] Successfully updated port: 75b0f475-5e6e-44de-9480-a71716ecb4e0 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2093.548076] env[62824]: DEBUG oslo_vmware.api [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52cc788e-69f0-bd2c-7703-d2ca66257d5b, 'name': SearchDatastore_Task, 'duration_secs': 0.009446} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2093.548986] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c55ec6e7-da4a-4f5a-a990-c57f5c6fdfeb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2093.558849] env[62824]: DEBUG oslo_vmware.api [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the task: (returnval){ [ 2093.558849] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]523a9ade-baf0-62e0-539e-0cbe41dacc5e" [ 2093.558849] env[62824]: _type = "Task" [ 2093.558849] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2093.569673] env[62824]: DEBUG oslo_vmware.api [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]523a9ade-baf0-62e0-539e-0cbe41dacc5e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2093.571609] env[62824]: DEBUG nova.compute.manager [req-9576c779-4b0c-45ac-a6e8-bdef94bfd8f1 req-e09bfc6e-ed46-4e01-9dca-36829b7855fd service nova] [instance: 350cfd12-a8d5-4a5f-b3de-d795fa179dfd] Received event network-vif-plugged-75b0f475-5e6e-44de-9480-a71716ecb4e0 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2093.571811] env[62824]: DEBUG oslo_concurrency.lockutils [req-9576c779-4b0c-45ac-a6e8-bdef94bfd8f1 req-e09bfc6e-ed46-4e01-9dca-36829b7855fd service nova] Acquiring lock "350cfd12-a8d5-4a5f-b3de-d795fa179dfd-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2093.572734] env[62824]: DEBUG oslo_concurrency.lockutils [req-9576c779-4b0c-45ac-a6e8-bdef94bfd8f1 req-e09bfc6e-ed46-4e01-9dca-36829b7855fd service nova] Lock "350cfd12-a8d5-4a5f-b3de-d795fa179dfd-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2093.572734] env[62824]: DEBUG oslo_concurrency.lockutils [req-9576c779-4b0c-45ac-a6e8-bdef94bfd8f1 req-e09bfc6e-ed46-4e01-9dca-36829b7855fd service nova] Lock "350cfd12-a8d5-4a5f-b3de-d795fa179dfd-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2093.572734] env[62824]: DEBUG nova.compute.manager [req-9576c779-4b0c-45ac-a6e8-bdef94bfd8f1 req-e09bfc6e-ed46-4e01-9dca-36829b7855fd service nova] [instance: 350cfd12-a8d5-4a5f-b3de-d795fa179dfd] No waiting events found dispatching network-vif-plugged-75b0f475-5e6e-44de-9480-a71716ecb4e0 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2093.572734] env[62824]: WARNING nova.compute.manager [req-9576c779-4b0c-45ac-a6e8-bdef94bfd8f1 req-e09bfc6e-ed46-4e01-9dca-36829b7855fd service nova] [instance: 350cfd12-a8d5-4a5f-b3de-d795fa179dfd] Received unexpected event network-vif-plugged-75b0f475-5e6e-44de-9480-a71716ecb4e0 for instance with vm_state building and task_state spawning. [ 2093.706885] env[62824]: DEBUG nova.compute.manager [None req-b5cb0b06-e659-44b5-a4c3-505f639c37f7 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 919e8a45-7810-4a8d-a1aa-5046b5ab059c] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2093.707682] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5cb0b06-e659-44b5-a4c3-505f639c37f7 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 919e8a45-7810-4a8d-a1aa-5046b5ab059c] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2093.713680] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d50ebbe3-0b46-440a-8ff8-04d1324a1e0b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2093.721082] env[62824]: DEBUG oslo_vmware.api [None req-b5cb0b06-e659-44b5-a4c3-505f639c37f7 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 2093.721082] env[62824]: value = "task-2146290" [ 2093.721082] env[62824]: _type = "Task" [ 2093.721082] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2093.729785] env[62824]: DEBUG oslo_vmware.api [None req-b5cb0b06-e659-44b5-a4c3-505f639c37f7 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146290, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2093.862950] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74186fb4-cae6-47f0-b067-a51b1f981799 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2093.871591] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e957b121-a28d-4c6b-9f4f-e6374601ea68 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2093.916035] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b386f5f-f48c-40ff-b5df-8c279013a349 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2093.924259] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4a80687-875f-4e59-886a-84491e04626a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2093.938667] env[62824]: DEBUG nova.compute.provider_tree [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2093.973646] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Acquiring lock "refresh_cache-350cfd12-a8d5-4a5f-b3de-d795fa179dfd" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2093.974060] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Acquired lock "refresh_cache-350cfd12-a8d5-4a5f-b3de-d795fa179dfd" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2093.974105] env[62824]: DEBUG nova.network.neutron [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 350cfd12-a8d5-4a5f-b3de-d795fa179dfd] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2094.069652] env[62824]: DEBUG oslo_vmware.api [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]523a9ade-baf0-62e0-539e-0cbe41dacc5e, 'name': SearchDatastore_Task, 'duration_secs': 0.010427} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2094.069652] env[62824]: DEBUG oslo_concurrency.lockutils [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2094.069774] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8/1870b82a-9783-44ac-8de2-7b9ffc2a1bc8.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2094.070019] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-47819a30-984d-4e6f-858f-0330cc59d06b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2094.076944] env[62824]: DEBUG oslo_vmware.api [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the task: (returnval){ [ 2094.076944] env[62824]: value = "task-2146291" [ 2094.076944] env[62824]: _type = "Task" [ 2094.076944] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2094.085065] env[62824]: DEBUG oslo_vmware.api [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146291, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2094.237078] env[62824]: DEBUG oslo_vmware.api [None req-b5cb0b06-e659-44b5-a4c3-505f639c37f7 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146290, 'name': PowerOffVM_Task, 'duration_secs': 0.214775} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2094.237623] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5cb0b06-e659-44b5-a4c3-505f639c37f7 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 919e8a45-7810-4a8d-a1aa-5046b5ab059c] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2094.240019] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-b5cb0b06-e659-44b5-a4c3-505f639c37f7 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 919e8a45-7810-4a8d-a1aa-5046b5ab059c] Volume detach. Driver type: vmdk {{(pid=62824) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 2094.240019] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-b5cb0b06-e659-44b5-a4c3-505f639c37f7 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 919e8a45-7810-4a8d-a1aa-5046b5ab059c] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-438665', 'volume_id': '20c6ff5e-0a8b-4cf4-8bd3-92cb944dd7a3', 'name': 'volume-20c6ff5e-0a8b-4cf4-8bd3-92cb944dd7a3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '919e8a45-7810-4a8d-a1aa-5046b5ab059c', 'attached_at': '', 'detached_at': '', 'volume_id': '20c6ff5e-0a8b-4cf4-8bd3-92cb944dd7a3', 'serial': '20c6ff5e-0a8b-4cf4-8bd3-92cb944dd7a3'} {{(pid=62824) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 2094.240019] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c39f6b3f-1259-4669-9a44-5d10fbb2e8cc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2094.277475] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a11561d-5399-4baa-a55f-4c19ddde0079 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2094.289865] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-caffbd48-1d8b-4595-b175-e1e86a8fc130 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2094.323609] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae693a1f-50df-42df-867f-97348e7fe6a4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2094.340744] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-b5cb0b06-e659-44b5-a4c3-505f639c37f7 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] The volume has not been displaced from its original location: [datastore2] volume-20c6ff5e-0a8b-4cf4-8bd3-92cb944dd7a3/volume-20c6ff5e-0a8b-4cf4-8bd3-92cb944dd7a3.vmdk. No consolidation needed. {{(pid=62824) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 2094.346901] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-b5cb0b06-e659-44b5-a4c3-505f639c37f7 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 919e8a45-7810-4a8d-a1aa-5046b5ab059c] Reconfiguring VM instance instance-0000003b to detach disk 2001 {{(pid=62824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 2094.347360] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c6173948-ae2d-4ee0-b016-441df978eb18 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2094.368279] env[62824]: DEBUG oslo_vmware.api [None req-b5cb0b06-e659-44b5-a4c3-505f639c37f7 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 2094.368279] env[62824]: value = "task-2146292" [ 2094.368279] env[62824]: _type = "Task" [ 2094.368279] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2094.379186] env[62824]: DEBUG oslo_vmware.api [None req-b5cb0b06-e659-44b5-a4c3-505f639c37f7 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146292, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2094.443039] env[62824]: DEBUG nova.scheduler.client.report [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2094.504591] env[62824]: DEBUG nova.objects.instance [None req-e2025339-4a38-43ec-b070-966aa57df690 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Lazy-loading 'flavor' on Instance uuid 59b5b883-4188-471c-8862-444f3ce08cb0 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2094.552201] env[62824]: DEBUG nova.network.neutron [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 350cfd12-a8d5-4a5f-b3de-d795fa179dfd] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2094.590254] env[62824]: DEBUG oslo_vmware.api [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146291, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.450311} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2094.590559] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8/1870b82a-9783-44ac-8de2-7b9ffc2a1bc8.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2094.590819] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2094.591070] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-63ef4bb8-b552-4fdc-a8e4-ae8b5275693d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2094.604025] env[62824]: DEBUG oslo_vmware.api [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the task: (returnval){ [ 2094.604025] env[62824]: value = "task-2146293" [ 2094.604025] env[62824]: _type = "Task" [ 2094.604025] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2094.607457] env[62824]: DEBUG oslo_vmware.api [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146293, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2094.775765] env[62824]: DEBUG nova.network.neutron [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 350cfd12-a8d5-4a5f-b3de-d795fa179dfd] Updating instance_info_cache with network_info: [{"id": "75b0f475-5e6e-44de-9480-a71716ecb4e0", "address": "fa:16:3e:fb:93:96", "network": {"id": "60037885-38d9-4586-9fd7-4907321e9c7e", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-370573912-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2e59203d676d48c8bd8402af4eb58f65", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f856fca-9fb5-41ea-a057-ac4193bd323d", "external-id": "nsx-vlan-transportzone-148", "segmentation_id": 148, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap75b0f475-5e", "ovs_interfaceid": "75b0f475-5e6e-44de-9480-a71716ecb4e0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2094.880267] env[62824]: DEBUG oslo_vmware.api [None req-b5cb0b06-e659-44b5-a4c3-505f639c37f7 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146292, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2094.950267] env[62824]: DEBUG oslo_concurrency.lockutils [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.613s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2094.951482] env[62824]: DEBUG nova.compute.manager [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2094.954724] env[62824]: DEBUG oslo_concurrency.lockutils [None req-84ba9f19-2925-48b1-85d5-00fe075f71bc tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.044s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2094.954989] env[62824]: DEBUG nova.objects.instance [None req-84ba9f19-2925-48b1-85d5-00fe075f71bc tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Lazy-loading 'resources' on Instance uuid ed7d3e95-f26e-40fd-a370-e17922bbff8e {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2095.014872] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e2025339-4a38-43ec-b070-966aa57df690 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Acquiring lock "refresh_cache-59b5b883-4188-471c-8862-444f3ce08cb0" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2095.014872] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e2025339-4a38-43ec-b070-966aa57df690 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Acquired lock "refresh_cache-59b5b883-4188-471c-8862-444f3ce08cb0" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2095.014872] env[62824]: DEBUG nova.network.neutron [None req-e2025339-4a38-43ec-b070-966aa57df690 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2095.015135] env[62824]: DEBUG nova.objects.instance [None req-e2025339-4a38-43ec-b070-966aa57df690 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Lazy-loading 'info_cache' on Instance uuid 59b5b883-4188-471c-8862-444f3ce08cb0 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2095.112024] env[62824]: DEBUG oslo_vmware.api [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146293, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071074} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2095.112024] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2095.112024] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc2e2c26-3909-4bf3-920c-fea9676509ba {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2095.137338] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8] Reconfiguring VM instance instance-0000003f to attach disk [datastore2] 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8/1870b82a-9783-44ac-8de2-7b9ffc2a1bc8.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2095.138190] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-58f6e897-c3f8-4bc2-9f77-60654e0926ed {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2095.162682] env[62824]: DEBUG oslo_vmware.api [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the task: (returnval){ [ 2095.162682] env[62824]: value = "task-2146294" [ 2095.162682] env[62824]: _type = "Task" [ 2095.162682] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2095.171072] env[62824]: DEBUG oslo_vmware.api [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146294, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2095.279808] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Releasing lock "refresh_cache-350cfd12-a8d5-4a5f-b3de-d795fa179dfd" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2095.279808] env[62824]: DEBUG nova.compute.manager [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 350cfd12-a8d5-4a5f-b3de-d795fa179dfd] Instance network_info: |[{"id": "75b0f475-5e6e-44de-9480-a71716ecb4e0", "address": "fa:16:3e:fb:93:96", "network": {"id": "60037885-38d9-4586-9fd7-4907321e9c7e", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-370573912-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2e59203d676d48c8bd8402af4eb58f65", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f856fca-9fb5-41ea-a057-ac4193bd323d", "external-id": "nsx-vlan-transportzone-148", "segmentation_id": 148, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap75b0f475-5e", "ovs_interfaceid": "75b0f475-5e6e-44de-9480-a71716ecb4e0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2095.279808] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 350cfd12-a8d5-4a5f-b3de-d795fa179dfd] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fb:93:96', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9f856fca-9fb5-41ea-a057-ac4193bd323d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '75b0f475-5e6e-44de-9480-a71716ecb4e0', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2095.288359] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Creating folder: Project (2e59203d676d48c8bd8402af4eb58f65). Parent ref: group-v438503. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 2095.289144] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-09c08de4-9684-4dd8-a2d2-73a3390e91a0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2095.301847] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Created folder: Project (2e59203d676d48c8bd8402af4eb58f65) in parent group-v438503. [ 2095.303903] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Creating folder: Instances. Parent ref: group-v438669. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 2095.303903] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0cd538dc-ccee-4d9b-b9cb-ae83b1b7c705 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2095.313237] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Created folder: Instances in parent group-v438669. [ 2095.313491] env[62824]: DEBUG oslo.service.loopingcall [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2095.313683] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 350cfd12-a8d5-4a5f-b3de-d795fa179dfd] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2095.313890] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cdf9db6a-3ee4-4938-9d12-849041e3cb3f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2095.334027] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2095.334027] env[62824]: value = "task-2146297" [ 2095.334027] env[62824]: _type = "Task" [ 2095.334027] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2095.341839] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146297, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2095.377741] env[62824]: DEBUG oslo_vmware.api [None req-b5cb0b06-e659-44b5-a4c3-505f639c37f7 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146292, 'name': ReconfigVM_Task, 'duration_secs': 0.571674} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2095.379055] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-b5cb0b06-e659-44b5-a4c3-505f639c37f7 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 919e8a45-7810-4a8d-a1aa-5046b5ab059c] Reconfigured VM instance instance-0000003b to detach disk 2001 {{(pid=62824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 2095.383635] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d32ee292-9c0a-4467-8180-8815deaf9669 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2095.399076] env[62824]: DEBUG oslo_vmware.api [None req-b5cb0b06-e659-44b5-a4c3-505f639c37f7 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 2095.399076] env[62824]: value = "task-2146298" [ 2095.399076] env[62824]: _type = "Task" [ 2095.399076] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2095.409170] env[62824]: DEBUG oslo_vmware.api [None req-b5cb0b06-e659-44b5-a4c3-505f639c37f7 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146298, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2095.458718] env[62824]: DEBUG nova.compute.utils [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2095.463510] env[62824]: DEBUG nova.compute.manager [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2095.463510] env[62824]: DEBUG nova.network.neutron [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: 363177c0-dc40-429a-a74b-e690da133edb] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2095.520334] env[62824]: DEBUG nova.objects.base [None req-e2025339-4a38-43ec-b070-966aa57df690 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Object Instance<59b5b883-4188-471c-8862-444f3ce08cb0> lazy-loaded attributes: flavor,info_cache {{(pid=62824) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 2095.567205] env[62824]: DEBUG nova.policy [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1ba0f76cb10d41939831abe409efb30b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0503b434636c4ef5bae8db1b0c74d2ce', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2095.679228] env[62824]: DEBUG oslo_vmware.api [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146294, 'name': ReconfigVM_Task, 'duration_secs': 0.291403} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2095.679546] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8] Reconfigured VM instance instance-0000003f to attach disk [datastore2] 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8/1870b82a-9783-44ac-8de2-7b9ffc2a1bc8.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2095.680370] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a48c5e84-3b08-4839-aaa9-c2c605030d78 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2095.694614] env[62824]: DEBUG oslo_vmware.api [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the task: (returnval){ [ 2095.694614] env[62824]: value = "task-2146299" [ 2095.694614] env[62824]: _type = "Task" [ 2095.694614] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2095.704475] env[62824]: DEBUG oslo_vmware.api [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146299, 'name': Rename_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2095.770359] env[62824]: DEBUG nova.compute.manager [req-de65f808-a1d5-4dd0-95de-3b0e323913e3 req-733ed3c1-dd12-4e6b-b835-367f421d054d service nova] [instance: 350cfd12-a8d5-4a5f-b3de-d795fa179dfd] Received event network-changed-75b0f475-5e6e-44de-9480-a71716ecb4e0 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2095.770553] env[62824]: DEBUG nova.compute.manager [req-de65f808-a1d5-4dd0-95de-3b0e323913e3 req-733ed3c1-dd12-4e6b-b835-367f421d054d service nova] [instance: 350cfd12-a8d5-4a5f-b3de-d795fa179dfd] Refreshing instance network info cache due to event network-changed-75b0f475-5e6e-44de-9480-a71716ecb4e0. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2095.770768] env[62824]: DEBUG oslo_concurrency.lockutils [req-de65f808-a1d5-4dd0-95de-3b0e323913e3 req-733ed3c1-dd12-4e6b-b835-367f421d054d service nova] Acquiring lock "refresh_cache-350cfd12-a8d5-4a5f-b3de-d795fa179dfd" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2095.770914] env[62824]: DEBUG oslo_concurrency.lockutils [req-de65f808-a1d5-4dd0-95de-3b0e323913e3 req-733ed3c1-dd12-4e6b-b835-367f421d054d service nova] Acquired lock "refresh_cache-350cfd12-a8d5-4a5f-b3de-d795fa179dfd" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2095.771221] env[62824]: DEBUG nova.network.neutron [req-de65f808-a1d5-4dd0-95de-3b0e323913e3 req-733ed3c1-dd12-4e6b-b835-367f421d054d service nova] [instance: 350cfd12-a8d5-4a5f-b3de-d795fa179dfd] Refreshing network info cache for port 75b0f475-5e6e-44de-9480-a71716ecb4e0 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2095.778844] env[62824]: INFO nova.compute.manager [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Rebuilding instance [ 2095.829696] env[62824]: DEBUG nova.compute.manager [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2095.830586] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e60403fb-8c42-485e-9a7b-b3a4ed4699df {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2095.851539] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146297, 'name': CreateVM_Task, 'duration_secs': 0.499317} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2095.852948] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 350cfd12-a8d5-4a5f-b3de-d795fa179dfd] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2095.853854] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2095.853896] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2095.854298] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2095.854525] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fe952f3a-c04f-46a4-99d7-dd5887875c5f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2095.864515] env[62824]: DEBUG oslo_vmware.api [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Waiting for the task: (returnval){ [ 2095.864515] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5247f871-37ed-03d4-f6ba-a5d99a9d4d5e" [ 2095.864515] env[62824]: _type = "Task" [ 2095.864515] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2095.870786] env[62824]: DEBUG oslo_vmware.api [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5247f871-37ed-03d4-f6ba-a5d99a9d4d5e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2095.910483] env[62824]: DEBUG oslo_vmware.api [None req-b5cb0b06-e659-44b5-a4c3-505f639c37f7 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146298, 'name': ReconfigVM_Task, 'duration_secs': 0.147123} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2095.911099] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-b5cb0b06-e659-44b5-a4c3-505f639c37f7 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 919e8a45-7810-4a8d-a1aa-5046b5ab059c] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-438665', 'volume_id': '20c6ff5e-0a8b-4cf4-8bd3-92cb944dd7a3', 'name': 'volume-20c6ff5e-0a8b-4cf4-8bd3-92cb944dd7a3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '919e8a45-7810-4a8d-a1aa-5046b5ab059c', 'attached_at': '', 'detached_at': '', 'volume_id': '20c6ff5e-0a8b-4cf4-8bd3-92cb944dd7a3', 'serial': '20c6ff5e-0a8b-4cf4-8bd3-92cb944dd7a3'} {{(pid=62824) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 2095.911527] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-b5cb0b06-e659-44b5-a4c3-505f639c37f7 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 919e8a45-7810-4a8d-a1aa-5046b5ab059c] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2095.912199] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7b3f2d4-992e-4925-af2f-fe78b18c5399 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2095.921761] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-b5cb0b06-e659-44b5-a4c3-505f639c37f7 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 919e8a45-7810-4a8d-a1aa-5046b5ab059c] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2095.922224] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dcd89431-9643-41f3-9fd7-1d564cba09ea {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2095.966725] env[62824]: DEBUG nova.compute.manager [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2096.058160] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0356b47d-6d51-4dfd-870e-3401ae56c3ce {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2096.065971] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec71e1d3-eedf-4c84-9fe1-1a5469eb2a9f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2096.102673] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34843cce-4366-4065-81a4-3546b5b6fd3f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2096.111250] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-649e56e4-624b-4461-962c-d2e27f8466a5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2096.129849] env[62824]: DEBUG nova.compute.provider_tree [None req-84ba9f19-2925-48b1-85d5-00fe075f71bc tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2096.132128] env[62824]: DEBUG nova.network.neutron [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Successfully created port: 8e8b16ef-e3f8-48c1-aab6-51d94d356c77 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2096.140754] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-b5cb0b06-e659-44b5-a4c3-505f639c37f7 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 919e8a45-7810-4a8d-a1aa-5046b5ab059c] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2096.140991] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-b5cb0b06-e659-44b5-a4c3-505f639c37f7 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 919e8a45-7810-4a8d-a1aa-5046b5ab059c] Deleting contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2096.141187] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-b5cb0b06-e659-44b5-a4c3-505f639c37f7 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Deleting the datastore file [datastore2] 919e8a45-7810-4a8d-a1aa-5046b5ab059c {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2096.142160] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-631d4c42-e834-49dc-a173-4ea31e87eefe {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2096.151309] env[62824]: DEBUG oslo_vmware.api [None req-b5cb0b06-e659-44b5-a4c3-505f639c37f7 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 2096.151309] env[62824]: value = "task-2146301" [ 2096.151309] env[62824]: _type = "Task" [ 2096.151309] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2096.159841] env[62824]: DEBUG oslo_vmware.api [None req-b5cb0b06-e659-44b5-a4c3-505f639c37f7 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146301, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2096.204536] env[62824]: DEBUG oslo_vmware.api [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146299, 'name': Rename_Task, 'duration_secs': 0.190055} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2096.205441] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2096.205724] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f26d1d7c-f324-4bdb-9942-dbc3e110cead {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2096.212354] env[62824]: DEBUG oslo_vmware.api [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the task: (returnval){ [ 2096.212354] env[62824]: value = "task-2146302" [ 2096.212354] env[62824]: _type = "Task" [ 2096.212354] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2096.220519] env[62824]: DEBUG oslo_vmware.api [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146302, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2096.373154] env[62824]: DEBUG oslo_vmware.api [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5247f871-37ed-03d4-f6ba-a5d99a9d4d5e, 'name': SearchDatastore_Task, 'duration_secs': 0.009293} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2096.373509] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2096.373760] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 350cfd12-a8d5-4a5f-b3de-d795fa179dfd] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2096.374013] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2096.374318] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2096.374414] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2096.374617] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-31bb1cd3-5d3c-4d2d-bf51-7f94ef8c31bc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2096.382696] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2096.382863] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2096.383832] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7777ca47-625e-413c-a5a0-50d1aec1643a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2096.388787] env[62824]: DEBUG oslo_vmware.api [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Waiting for the task: (returnval){ [ 2096.388787] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]529039d7-27e2-b4c5-5a18-8f4e59a073ff" [ 2096.388787] env[62824]: _type = "Task" [ 2096.388787] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2096.396481] env[62824]: DEBUG oslo_vmware.api [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]529039d7-27e2-b4c5-5a18-8f4e59a073ff, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2096.518785] env[62824]: DEBUG nova.network.neutron [None req-e2025339-4a38-43ec-b070-966aa57df690 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Updating instance_info_cache with network_info: [{"id": "1e74959f-64c2-4c37-85f8-fa8f81a30d60", "address": "fa:16:3e:6d:26:78", "network": {"id": "ae1dfabd-223a-4d4c-a8c4-fe65a83dd9c0", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-797102932-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.148", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "102d2f886fd44e1da6780e48e36f1b62", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "517421c3-bea0-419c-ab0b-987815e5d160", "external-id": "nsx-vlan-transportzone-68", "segmentation_id": 68, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1e74959f-64", "ovs_interfaceid": "1e74959f-64c2-4c37-85f8-fa8f81a30d60", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2096.638026] env[62824]: DEBUG nova.scheduler.client.report [None req-84ba9f19-2925-48b1-85d5-00fe075f71bc tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2096.662732] env[62824]: DEBUG oslo_vmware.api [None req-b5cb0b06-e659-44b5-a4c3-505f639c37f7 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146301, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.151821} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2096.663459] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-b5cb0b06-e659-44b5-a4c3-505f639c37f7 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2096.663772] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-b5cb0b06-e659-44b5-a4c3-505f639c37f7 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 919e8a45-7810-4a8d-a1aa-5046b5ab059c] Deleted contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2096.664098] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-b5cb0b06-e659-44b5-a4c3-505f639c37f7 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 919e8a45-7810-4a8d-a1aa-5046b5ab059c] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2096.664431] env[62824]: INFO nova.compute.manager [None req-b5cb0b06-e659-44b5-a4c3-505f639c37f7 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 919e8a45-7810-4a8d-a1aa-5046b5ab059c] Took 2.96 seconds to destroy the instance on the hypervisor. [ 2096.664790] env[62824]: DEBUG oslo.service.loopingcall [None req-b5cb0b06-e659-44b5-a4c3-505f639c37f7 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2096.667204] env[62824]: DEBUG nova.compute.manager [-] [instance: 919e8a45-7810-4a8d-a1aa-5046b5ab059c] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2096.667204] env[62824]: DEBUG nova.network.neutron [-] [instance: 919e8a45-7810-4a8d-a1aa-5046b5ab059c] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2096.723740] env[62824]: DEBUG oslo_vmware.api [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146302, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2096.852994] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2096.852994] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-29cac50b-128d-4a76-9222-e762a9eee7c9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2096.857925] env[62824]: DEBUG oslo_vmware.api [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for the task: (returnval){ [ 2096.857925] env[62824]: value = "task-2146303" [ 2096.857925] env[62824]: _type = "Task" [ 2096.857925] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2096.867995] env[62824]: DEBUG oslo_vmware.api [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146303, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2096.903868] env[62824]: DEBUG oslo_vmware.api [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]529039d7-27e2-b4c5-5a18-8f4e59a073ff, 'name': SearchDatastore_Task, 'duration_secs': 0.011535} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2096.905502] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-81a4926e-ff1c-4699-bec1-86bab9a7b657 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2096.911197] env[62824]: DEBUG oslo_vmware.api [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Waiting for the task: (returnval){ [ 2096.911197] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52b9f90b-2858-dfd8-b66d-159b90b81ae0" [ 2096.911197] env[62824]: _type = "Task" [ 2096.911197] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2096.921792] env[62824]: DEBUG oslo_vmware.api [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52b9f90b-2858-dfd8-b66d-159b90b81ae0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2096.941497] env[62824]: DEBUG nova.network.neutron [req-de65f808-a1d5-4dd0-95de-3b0e323913e3 req-733ed3c1-dd12-4e6b-b835-367f421d054d service nova] [instance: 350cfd12-a8d5-4a5f-b3de-d795fa179dfd] Updated VIF entry in instance network info cache for port 75b0f475-5e6e-44de-9480-a71716ecb4e0. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2096.941854] env[62824]: DEBUG nova.network.neutron [req-de65f808-a1d5-4dd0-95de-3b0e323913e3 req-733ed3c1-dd12-4e6b-b835-367f421d054d service nova] [instance: 350cfd12-a8d5-4a5f-b3de-d795fa179dfd] Updating instance_info_cache with network_info: [{"id": "75b0f475-5e6e-44de-9480-a71716ecb4e0", "address": "fa:16:3e:fb:93:96", "network": {"id": "60037885-38d9-4586-9fd7-4907321e9c7e", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-370573912-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2e59203d676d48c8bd8402af4eb58f65", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f856fca-9fb5-41ea-a057-ac4193bd323d", "external-id": "nsx-vlan-transportzone-148", "segmentation_id": 148, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap75b0f475-5e", "ovs_interfaceid": "75b0f475-5e6e-44de-9480-a71716ecb4e0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2096.977439] env[62824]: DEBUG nova.compute.manager [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2097.010509] env[62824]: DEBUG nova.virt.hardware [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2097.010754] env[62824]: DEBUG nova.virt.hardware [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2097.010909] env[62824]: DEBUG nova.virt.hardware [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2097.011291] env[62824]: DEBUG nova.virt.hardware [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2097.011492] env[62824]: DEBUG nova.virt.hardware [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2097.011641] env[62824]: DEBUG nova.virt.hardware [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2097.011851] env[62824]: DEBUG nova.virt.hardware [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2097.012017] env[62824]: DEBUG nova.virt.hardware [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2097.012191] env[62824]: DEBUG nova.virt.hardware [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2097.012354] env[62824]: DEBUG nova.virt.hardware [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2097.012527] env[62824]: DEBUG nova.virt.hardware [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2097.013422] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67c5fd61-46cf-4324-aa65-72e600abfd75 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2097.022274] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e2025339-4a38-43ec-b070-966aa57df690 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Releasing lock "refresh_cache-59b5b883-4188-471c-8862-444f3ce08cb0" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2097.025703] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d92fad3-75e5-42cf-9eb1-8e620d5e0346 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2097.145101] env[62824]: DEBUG oslo_concurrency.lockutils [None req-84ba9f19-2925-48b1-85d5-00fe075f71bc tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.188s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2097.145101] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3aa78573-24a8-42b5-8896-df119ccd8889 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.016s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2097.145950] env[62824]: DEBUG nova.objects.instance [None req-3aa78573-24a8-42b5-8896-df119ccd8889 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lazy-loading 'resources' on Instance uuid f4d63a93-23af-470c-b36a-662af81dc386 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2097.175531] env[62824]: INFO nova.scheduler.client.report [None req-84ba9f19-2925-48b1-85d5-00fe075f71bc tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Deleted allocations for instance ed7d3e95-f26e-40fd-a370-e17922bbff8e [ 2097.224970] env[62824]: DEBUG oslo_vmware.api [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146302, 'name': PowerOnVM_Task, 'duration_secs': 0.599822} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2097.224970] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2097.225068] env[62824]: INFO nova.compute.manager [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8] Took 11.77 seconds to spawn the instance on the hypervisor. [ 2097.225236] env[62824]: DEBUG nova.compute.manager [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2097.226309] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cc843ed-ef7e-46dc-af7d-6f2498fc9860 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2097.372928] env[62824]: DEBUG oslo_vmware.api [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146303, 'name': PowerOffVM_Task, 'duration_secs': 0.2177} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2097.373213] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2097.373580] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2097.374797] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4448be09-5683-4efb-a728-0182bac6a94d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2097.386771] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2097.387164] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c0e2b930-feed-4a0d-83ac-0ee418c492a8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2097.438838] env[62824]: DEBUG oslo_vmware.api [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52b9f90b-2858-dfd8-b66d-159b90b81ae0, 'name': SearchDatastore_Task, 'duration_secs': 0.010722} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2097.439290] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2097.440198] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 350cfd12-a8d5-4a5f-b3de-d795fa179dfd/350cfd12-a8d5-4a5f-b3de-d795fa179dfd.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2097.440198] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-198a7b25-14b1-4a62-a5ba-3cde6c1338ab {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2097.446455] env[62824]: DEBUG oslo_concurrency.lockutils [req-de65f808-a1d5-4dd0-95de-3b0e323913e3 req-733ed3c1-dd12-4e6b-b835-367f421d054d service nova] Releasing lock "refresh_cache-350cfd12-a8d5-4a5f-b3de-d795fa179dfd" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2097.448232] env[62824]: DEBUG nova.compute.manager [req-520cc285-ec2d-45d7-aa0b-7503a34ef644 req-60b00087-a2b9-4d1a-8ea3-70b7f504e705 service nova] [instance: 919e8a45-7810-4a8d-a1aa-5046b5ab059c] Received event network-vif-deleted-ec8bab9e-6a51-4cc3-9fe8-4764dfd97d45 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2097.448532] env[62824]: INFO nova.compute.manager [req-520cc285-ec2d-45d7-aa0b-7503a34ef644 req-60b00087-a2b9-4d1a-8ea3-70b7f504e705 service nova] [instance: 919e8a45-7810-4a8d-a1aa-5046b5ab059c] Neutron deleted interface ec8bab9e-6a51-4cc3-9fe8-4764dfd97d45; detaching it from the instance and deleting it from the info cache [ 2097.448819] env[62824]: DEBUG nova.network.neutron [req-520cc285-ec2d-45d7-aa0b-7503a34ef644 req-60b00087-a2b9-4d1a-8ea3-70b7f504e705 service nova] [instance: 919e8a45-7810-4a8d-a1aa-5046b5ab059c] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2097.463028] env[62824]: DEBUG oslo_vmware.api [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Waiting for the task: (returnval){ [ 2097.463028] env[62824]: value = "task-2146305" [ 2097.463028] env[62824]: _type = "Task" [ 2097.463028] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2097.479470] env[62824]: DEBUG oslo_vmware.api [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Task: {'id': task-2146305, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2097.604059] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2097.604059] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Deleting contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2097.604059] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Deleting the datastore file [datastore1] 07ce931f-45ef-409b-b714-9f1cd47a3a88 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2097.604455] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cb99e9a6-d05c-4ae5-a886-fed788062897 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2097.611903] env[62824]: DEBUG oslo_vmware.api [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for the task: (returnval){ [ 2097.611903] env[62824]: value = "task-2146306" [ 2097.611903] env[62824]: _type = "Task" [ 2097.611903] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2097.624908] env[62824]: DEBUG oslo_vmware.api [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146306, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2097.628464] env[62824]: DEBUG nova.network.neutron [-] [instance: 919e8a45-7810-4a8d-a1aa-5046b5ab059c] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2097.687845] env[62824]: DEBUG oslo_concurrency.lockutils [None req-84ba9f19-2925-48b1-85d5-00fe075f71bc tempest-ServerShowV254Test-518260744 tempest-ServerShowV254Test-518260744-project-member] Lock "ed7d3e95-f26e-40fd-a370-e17922bbff8e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.200s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2097.747498] env[62824]: INFO nova.compute.manager [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8] Took 25.87 seconds to build instance. [ 2097.955821] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-adaa8e06-2493-4e86-9151-9640c3082a8c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2097.974496] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f74c74b-859f-48f0-bcdf-fa6c751dcedd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2097.994984] env[62824]: DEBUG oslo_vmware.api [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Task: {'id': task-2146305, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2098.027700] env[62824]: DEBUG nova.compute.manager [req-520cc285-ec2d-45d7-aa0b-7503a34ef644 req-60b00087-a2b9-4d1a-8ea3-70b7f504e705 service nova] [instance: 919e8a45-7810-4a8d-a1aa-5046b5ab059c] Detach interface failed, port_id=ec8bab9e-6a51-4cc3-9fe8-4764dfd97d45, reason: Instance 919e8a45-7810-4a8d-a1aa-5046b5ab059c could not be found. {{(pid=62824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11482}} [ 2098.031311] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2025339-4a38-43ec-b070-966aa57df690 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2098.031827] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c8dcc057-c2d6-43f6-8e94-1e629806669b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2098.040214] env[62824]: DEBUG oslo_vmware.api [None req-e2025339-4a38-43ec-b070-966aa57df690 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Waiting for the task: (returnval){ [ 2098.040214] env[62824]: value = "task-2146307" [ 2098.040214] env[62824]: _type = "Task" [ 2098.040214] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2098.052598] env[62824]: DEBUG oslo_vmware.api [None req-e2025339-4a38-43ec-b070-966aa57df690 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Task: {'id': task-2146307, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2098.124758] env[62824]: DEBUG oslo_vmware.api [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146306, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.261488} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2098.128253] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2098.128497] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Deleted contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2098.128738] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2098.133624] env[62824]: INFO nova.compute.manager [-] [instance: 919e8a45-7810-4a8d-a1aa-5046b5ab059c] Took 1.47 seconds to deallocate network for instance. [ 2098.256892] env[62824]: DEBUG oslo_concurrency.lockutils [None req-57438db8-2184-4c3f-a6d4-ec4ad5799dc4 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Lock "1870b82a-9783-44ac-8de2-7b9ffc2a1bc8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.401s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2098.293628] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbdd2e72-2957-4192-8163-e8d0b85c3ec6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2098.302237] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff4ed316-c1c5-4a79-a643-f55b1fbc223f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2098.335913] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73782d5c-378f-4be9-a708-d557638a0bce {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2098.345193] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2cf1b67-7715-48b9-a3cb-85e914f638a9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2098.360263] env[62824]: DEBUG nova.compute.provider_tree [None req-3aa78573-24a8-42b5-8896-df119ccd8889 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2098.472998] env[62824]: DEBUG oslo_vmware.api [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Task: {'id': task-2146305, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.658112} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2098.473344] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 350cfd12-a8d5-4a5f-b3de-d795fa179dfd/350cfd12-a8d5-4a5f-b3de-d795fa179dfd.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2098.473652] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 350cfd12-a8d5-4a5f-b3de-d795fa179dfd] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2098.473889] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-801d421d-3ce4-42a2-824a-91b4a7aae487 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2098.481061] env[62824]: DEBUG oslo_vmware.api [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Waiting for the task: (returnval){ [ 2098.481061] env[62824]: value = "task-2146308" [ 2098.481061] env[62824]: _type = "Task" [ 2098.481061] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2098.490162] env[62824]: DEBUG oslo_vmware.api [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Task: {'id': task-2146308, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2098.552705] env[62824]: DEBUG oslo_vmware.api [None req-e2025339-4a38-43ec-b070-966aa57df690 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Task: {'id': task-2146307, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2098.703485] env[62824]: INFO nova.compute.manager [None req-b5cb0b06-e659-44b5-a4c3-505f639c37f7 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 919e8a45-7810-4a8d-a1aa-5046b5ab059c] Took 0.57 seconds to detach 1 volumes for instance. [ 2098.760026] env[62824]: DEBUG nova.compute.manager [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: d067fa97-cedc-4e3d-9be4-d860a79a7723] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2098.855994] env[62824]: DEBUG nova.network.neutron [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Successfully updated port: 8e8b16ef-e3f8-48c1-aab6-51d94d356c77 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2098.865147] env[62824]: DEBUG nova.scheduler.client.report [None req-3aa78573-24a8-42b5-8896-df119ccd8889 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2098.993349] env[62824]: DEBUG oslo_vmware.api [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Task: {'id': task-2146308, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.198716} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2098.993349] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 350cfd12-a8d5-4a5f-b3de-d795fa179dfd] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2098.994420] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b4799db-9665-458a-b214-88a0a6b0c784 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2099.018220] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 350cfd12-a8d5-4a5f-b3de-d795fa179dfd] Reconfiguring VM instance instance-00000040 to attach disk [datastore2] 350cfd12-a8d5-4a5f-b3de-d795fa179dfd/350cfd12-a8d5-4a5f-b3de-d795fa179dfd.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2099.018907] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f73f12cf-6745-4062-8a29-cf3daf1821ff {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2099.048999] env[62824]: DEBUG oslo_vmware.api [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Waiting for the task: (returnval){ [ 2099.048999] env[62824]: value = "task-2146309" [ 2099.048999] env[62824]: _type = "Task" [ 2099.048999] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2099.054054] env[62824]: DEBUG oslo_vmware.api [None req-e2025339-4a38-43ec-b070-966aa57df690 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Task: {'id': task-2146307, 'name': PowerOnVM_Task, 'duration_secs': 0.757388} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2099.057310] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2025339-4a38-43ec-b070-966aa57df690 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2099.057500] env[62824]: DEBUG nova.compute.manager [None req-e2025339-4a38-43ec-b070-966aa57df690 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2099.057776] env[62824]: DEBUG oslo_vmware.api [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Task: {'id': task-2146309, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2099.058486] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd10a222-84d2-4f27-8411-b21df3cc2085 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2099.152850] env[62824]: DEBUG nova.compute.manager [req-6966d2ea-ea57-4b89-9ccf-3b82e8f5e4b3 req-6f62c07c-2515-41bf-b1a9-b6d831860d37 service nova] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Received event network-vif-plugged-8e8b16ef-e3f8-48c1-aab6-51d94d356c77 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2099.153096] env[62824]: DEBUG oslo_concurrency.lockutils [req-6966d2ea-ea57-4b89-9ccf-3b82e8f5e4b3 req-6f62c07c-2515-41bf-b1a9-b6d831860d37 service nova] Acquiring lock "363177c0-dc40-429a-a74b-e690da133edb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2099.154017] env[62824]: DEBUG oslo_concurrency.lockutils [req-6966d2ea-ea57-4b89-9ccf-3b82e8f5e4b3 req-6f62c07c-2515-41bf-b1a9-b6d831860d37 service nova] Lock "363177c0-dc40-429a-a74b-e690da133edb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2099.154017] env[62824]: DEBUG oslo_concurrency.lockutils [req-6966d2ea-ea57-4b89-9ccf-3b82e8f5e4b3 req-6f62c07c-2515-41bf-b1a9-b6d831860d37 service nova] Lock "363177c0-dc40-429a-a74b-e690da133edb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2099.154017] env[62824]: DEBUG nova.compute.manager [req-6966d2ea-ea57-4b89-9ccf-3b82e8f5e4b3 req-6f62c07c-2515-41bf-b1a9-b6d831860d37 service nova] [instance: 363177c0-dc40-429a-a74b-e690da133edb] No waiting events found dispatching network-vif-plugged-8e8b16ef-e3f8-48c1-aab6-51d94d356c77 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2099.154128] env[62824]: WARNING nova.compute.manager [req-6966d2ea-ea57-4b89-9ccf-3b82e8f5e4b3 req-6f62c07c-2515-41bf-b1a9-b6d831860d37 service nova] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Received unexpected event network-vif-plugged-8e8b16ef-e3f8-48c1-aab6-51d94d356c77 for instance with vm_state building and task_state spawning. [ 2099.179908] env[62824]: DEBUG nova.virt.hardware [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2099.180239] env[62824]: DEBUG nova.virt.hardware [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2099.180309] env[62824]: DEBUG nova.virt.hardware [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2099.180491] env[62824]: DEBUG nova.virt.hardware [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2099.180642] env[62824]: DEBUG nova.virt.hardware [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2099.180801] env[62824]: DEBUG nova.virt.hardware [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2099.181032] env[62824]: DEBUG nova.virt.hardware [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2099.181205] env[62824]: DEBUG nova.virt.hardware [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2099.181386] env[62824]: DEBUG nova.virt.hardware [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2099.181566] env[62824]: DEBUG nova.virt.hardware [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2099.181743] env[62824]: DEBUG nova.virt.hardware [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2099.182927] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-974c40f0-c831-49ba-9e92-dbac63203ff9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2099.191275] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75bec857-d624-4029-9b5b-075dff52e881 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2099.205386] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:07:e9:a0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4c9a12d2-469f-4199-bfaa-f791d765deac', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '309b4346-1935-41f1-a374-d322c4a8c016', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2099.212999] env[62824]: DEBUG oslo.service.loopingcall [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2099.213988] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b5cb0b06-e659-44b5-a4c3-505f639c37f7 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2099.214249] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2099.214478] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-33efc5a2-baa8-4bfd-b85f-77bcfc859446 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2099.241069] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2099.241069] env[62824]: value = "task-2146310" [ 2099.241069] env[62824]: _type = "Task" [ 2099.241069] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2099.251525] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146310, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2099.286080] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2099.359473] env[62824]: DEBUG oslo_concurrency.lockutils [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Acquiring lock "refresh_cache-363177c0-dc40-429a-a74b-e690da133edb" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2099.359657] env[62824]: DEBUG oslo_concurrency.lockutils [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Acquired lock "refresh_cache-363177c0-dc40-429a-a74b-e690da133edb" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2099.359801] env[62824]: DEBUG nova.network.neutron [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2099.376701] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3aa78573-24a8-42b5-8896-df119ccd8889 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.229s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2099.376701] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.898s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2099.378688] env[62824]: INFO nova.compute.claims [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2099.404146] env[62824]: INFO nova.scheduler.client.report [None req-3aa78573-24a8-42b5-8896-df119ccd8889 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Deleted allocations for instance f4d63a93-23af-470c-b36a-662af81dc386 [ 2099.561017] env[62824]: DEBUG oslo_vmware.api [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Task: {'id': task-2146309, 'name': ReconfigVM_Task, 'duration_secs': 0.486708} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2099.561451] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 350cfd12-a8d5-4a5f-b3de-d795fa179dfd] Reconfigured VM instance instance-00000040 to attach disk [datastore2] 350cfd12-a8d5-4a5f-b3de-d795fa179dfd/350cfd12-a8d5-4a5f-b3de-d795fa179dfd.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2099.562689] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5a8665c1-d73a-41b4-a4bd-9b8c79c7a0ea {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2099.572975] env[62824]: DEBUG oslo_vmware.api [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Waiting for the task: (returnval){ [ 2099.572975] env[62824]: value = "task-2146311" [ 2099.572975] env[62824]: _type = "Task" [ 2099.572975] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2099.588549] env[62824]: DEBUG oslo_vmware.api [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Task: {'id': task-2146311, 'name': Rename_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2099.761190] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146310, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2099.911814] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3aa78573-24a8-42b5-8896-df119ccd8889 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lock "f4d63a93-23af-470c-b36a-662af81dc386" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.545s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2099.927173] env[62824]: DEBUG nova.network.neutron [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2100.090232] env[62824]: DEBUG oslo_vmware.api [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Task: {'id': task-2146311, 'name': Rename_Task, 'duration_secs': 0.335868} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2100.090232] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 350cfd12-a8d5-4a5f-b3de-d795fa179dfd] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2100.090232] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5bdfa346-0784-4658-917a-5e41a80f1eef {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2100.095872] env[62824]: DEBUG oslo_vmware.api [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Waiting for the task: (returnval){ [ 2100.095872] env[62824]: value = "task-2146312" [ 2100.095872] env[62824]: _type = "Task" [ 2100.095872] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2100.104700] env[62824]: DEBUG oslo_vmware.api [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Task: {'id': task-2146312, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2100.253411] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146310, 'name': CreateVM_Task, 'duration_secs': 0.698936} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2100.254056] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2100.255080] env[62824]: DEBUG oslo_concurrency.lockutils [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2100.255571] env[62824]: DEBUG oslo_concurrency.lockutils [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2100.255918] env[62824]: DEBUG oslo_concurrency.lockutils [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2100.256313] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7257f692-6b5a-4aa7-86f1-8eff5b7bfd79 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2100.263600] env[62824]: DEBUG oslo_vmware.api [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for the task: (returnval){ [ 2100.263600] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52ee1a9e-932f-5696-b550-e21c81c9a08e" [ 2100.263600] env[62824]: _type = "Task" [ 2100.263600] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2100.275102] env[62824]: DEBUG oslo_vmware.api [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52ee1a9e-932f-5696-b550-e21c81c9a08e, 'name': SearchDatastore_Task, 'duration_secs': 0.009118} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2100.276534] env[62824]: DEBUG nova.network.neutron [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Updating instance_info_cache with network_info: [{"id": "8e8b16ef-e3f8-48c1-aab6-51d94d356c77", "address": "fa:16:3e:48:83:8c", "network": {"id": "bcf30d77-cac1-4deb-971d-efb7c1fb8354", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1030043097-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "0503b434636c4ef5bae8db1b0c74d2ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3cc0a33d-17c0-4b87-b48f-413a87a4cc6a", "external-id": "nsx-vlan-transportzone-865", "segmentation_id": 865, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8e8b16ef-e3", "ovs_interfaceid": "8e8b16ef-e3f8-48c1-aab6-51d94d356c77", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2100.278620] env[62824]: DEBUG oslo_concurrency.lockutils [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2100.278938] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2100.279550] env[62824]: DEBUG oslo_concurrency.lockutils [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2100.279812] env[62824]: DEBUG oslo_concurrency.lockutils [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2100.280084] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2100.280629] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e5c6b23e-2f71-4565-9217-6b5c1d5d057d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2100.289584] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2100.290061] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2100.290894] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3d1e8fdc-2525-42fc-bc09-0f79ff278657 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2100.298998] env[62824]: DEBUG oslo_vmware.api [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for the task: (returnval){ [ 2100.298998] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]529f49bb-945e-2054-f4ca-0a3fc2b32ef3" [ 2100.298998] env[62824]: _type = "Task" [ 2100.298998] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2100.310274] env[62824]: DEBUG oslo_vmware.api [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]529f49bb-945e-2054-f4ca-0a3fc2b32ef3, 'name': SearchDatastore_Task, 'duration_secs': 0.008394} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2100.312591] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d2ea0658-a8e3-4383-81c4-6f7fdc3a3bf0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2100.318262] env[62824]: DEBUG oslo_vmware.api [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for the task: (returnval){ [ 2100.318262] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]524bdcb3-79c0-5830-9662-ed79359e69d7" [ 2100.318262] env[62824]: _type = "Task" [ 2100.318262] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2100.327656] env[62824]: DEBUG oslo_vmware.api [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]524bdcb3-79c0-5830-9662-ed79359e69d7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2100.613094] env[62824]: DEBUG oslo_vmware.api [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Task: {'id': task-2146312, 'name': PowerOnVM_Task, 'duration_secs': 0.4915} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2100.613534] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 350cfd12-a8d5-4a5f-b3de-d795fa179dfd] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2100.613981] env[62824]: INFO nova.compute.manager [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 350cfd12-a8d5-4a5f-b3de-d795fa179dfd] Took 8.48 seconds to spawn the instance on the hypervisor. [ 2100.614299] env[62824]: DEBUG nova.compute.manager [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 350cfd12-a8d5-4a5f-b3de-d795fa179dfd] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2100.615449] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4926a0a-0db6-48cb-bf35-4628f3d1c892 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2100.713062] env[62824]: DEBUG nova.compute.manager [None req-4a76d89b-cbff-4ca0-a9aa-3ebc58bfda76 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2100.714125] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5718c864-dc58-45aa-90f6-7f62715e5ea3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2100.785125] env[62824]: DEBUG oslo_concurrency.lockutils [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Releasing lock "refresh_cache-363177c0-dc40-429a-a74b-e690da133edb" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2100.785125] env[62824]: DEBUG nova.compute.manager [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Instance network_info: |[{"id": "8e8b16ef-e3f8-48c1-aab6-51d94d356c77", "address": "fa:16:3e:48:83:8c", "network": {"id": "bcf30d77-cac1-4deb-971d-efb7c1fb8354", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1030043097-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "0503b434636c4ef5bae8db1b0c74d2ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3cc0a33d-17c0-4b87-b48f-413a87a4cc6a", "external-id": "nsx-vlan-transportzone-865", "segmentation_id": 865, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8e8b16ef-e3", "ovs_interfaceid": "8e8b16ef-e3f8-48c1-aab6-51d94d356c77", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2100.785125] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:48:83:8c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3cc0a33d-17c0-4b87-b48f-413a87a4cc6a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8e8b16ef-e3f8-48c1-aab6-51d94d356c77', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2100.790543] env[62824]: DEBUG oslo.service.loopingcall [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2100.793349] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2100.793718] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-80819d75-00d9-432d-86d9-62c7e0dd4120 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2100.820444] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2100.820444] env[62824]: value = "task-2146313" [ 2100.820444] env[62824]: _type = "Task" [ 2100.820444] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2100.833859] env[62824]: DEBUG oslo_vmware.api [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]524bdcb3-79c0-5830-9662-ed79359e69d7, 'name': SearchDatastore_Task, 'duration_secs': 0.022179} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2100.834051] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146313, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2100.836997] env[62824]: DEBUG oslo_concurrency.lockutils [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2100.837428] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 07ce931f-45ef-409b-b714-9f1cd47a3a88/07ce931f-45ef-409b-b714-9f1cd47a3a88.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2100.837979] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-40a254b1-e4da-460a-81bc-4af0db1a98c9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2100.846677] env[62824]: DEBUG oslo_vmware.api [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for the task: (returnval){ [ 2100.846677] env[62824]: value = "task-2146314" [ 2100.846677] env[62824]: _type = "Task" [ 2100.846677] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2100.858432] env[62824]: DEBUG oslo_vmware.api [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146314, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2101.068122] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1437298b-7669-4861-b77a-b1d2c99e80a7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2101.078554] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46d9f3db-08e4-405d-b32d-b30627862760 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2101.117713] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f2c1cd4-2cc4-49e6-9d19-d2b693cda19d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2101.130997] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ff8ada3-2fc1-44ca-b845-b7709a9b3a86 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2101.150556] env[62824]: DEBUG nova.compute.provider_tree [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2101.158551] env[62824]: INFO nova.compute.manager [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 350cfd12-a8d5-4a5f-b3de-d795fa179dfd] Took 23.79 seconds to build instance. [ 2101.230419] env[62824]: INFO nova.compute.manager [None req-4a76d89b-cbff-4ca0-a9aa-3ebc58bfda76 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8] instance snapshotting [ 2101.236721] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67cb7b07-2f84-40cd-a350-6ba8cfc9b771 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2101.262177] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6dbf853-c46a-4017-bf37-9375da3ac12f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2101.338554] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146313, 'name': CreateVM_Task, 'duration_secs': 0.491201} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2101.338554] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2101.338554] env[62824]: DEBUG oslo_concurrency.lockutils [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2101.338554] env[62824]: DEBUG oslo_concurrency.lockutils [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2101.338554] env[62824]: DEBUG oslo_concurrency.lockutils [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2101.338554] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-24a0a65a-030c-40fa-b0b8-01be1745b534 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2101.343883] env[62824]: DEBUG oslo_vmware.api [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Waiting for the task: (returnval){ [ 2101.343883] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]529941ad-f4e8-05a9-a649-177d7c4fb502" [ 2101.343883] env[62824]: _type = "Task" [ 2101.343883] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2101.356265] env[62824]: DEBUG oslo_vmware.api [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]529941ad-f4e8-05a9-a649-177d7c4fb502, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2101.360211] env[62824]: DEBUG oslo_vmware.api [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146314, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2101.443499] env[62824]: DEBUG nova.compute.manager [req-d684e001-e872-4375-aca0-fbad54924094 req-ac87ca48-2f76-42d8-b8eb-ce066dad79fd service nova] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Received event network-changed-8e8b16ef-e3f8-48c1-aab6-51d94d356c77 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2101.443697] env[62824]: DEBUG nova.compute.manager [req-d684e001-e872-4375-aca0-fbad54924094 req-ac87ca48-2f76-42d8-b8eb-ce066dad79fd service nova] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Refreshing instance network info cache due to event network-changed-8e8b16ef-e3f8-48c1-aab6-51d94d356c77. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2101.443907] env[62824]: DEBUG oslo_concurrency.lockutils [req-d684e001-e872-4375-aca0-fbad54924094 req-ac87ca48-2f76-42d8-b8eb-ce066dad79fd service nova] Acquiring lock "refresh_cache-363177c0-dc40-429a-a74b-e690da133edb" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2101.444055] env[62824]: DEBUG oslo_concurrency.lockutils [req-d684e001-e872-4375-aca0-fbad54924094 req-ac87ca48-2f76-42d8-b8eb-ce066dad79fd service nova] Acquired lock "refresh_cache-363177c0-dc40-429a-a74b-e690da133edb" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2101.444223] env[62824]: DEBUG nova.network.neutron [req-d684e001-e872-4375-aca0-fbad54924094 req-ac87ca48-2f76-42d8-b8eb-ce066dad79fd service nova] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Refreshing network info cache for port 8e8b16ef-e3f8-48c1-aab6-51d94d356c77 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2101.658021] env[62824]: DEBUG nova.scheduler.client.report [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2101.662913] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5b2f43c1-2f5d-40f0-8a93-eb160f342723 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Lock "350cfd12-a8d5-4a5f-b3de-d795fa179dfd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.303s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2101.777371] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-4a76d89b-cbff-4ca0-a9aa-3ebc58bfda76 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8] Creating Snapshot of the VM instance {{(pid=62824) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 2101.777564] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-12b35eee-3f0f-4d4d-bd30-a1c1ca8862ec {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2101.787788] env[62824]: DEBUG oslo_vmware.api [None req-4a76d89b-cbff-4ca0-a9aa-3ebc58bfda76 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the task: (returnval){ [ 2101.787788] env[62824]: value = "task-2146315" [ 2101.787788] env[62824]: _type = "Task" [ 2101.787788] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2101.796151] env[62824]: DEBUG oslo_vmware.api [None req-4a76d89b-cbff-4ca0-a9aa-3ebc58bfda76 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146315, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2101.857324] env[62824]: DEBUG oslo_vmware.api [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146314, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.541313} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2101.861008] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 07ce931f-45ef-409b-b714-9f1cd47a3a88/07ce931f-45ef-409b-b714-9f1cd47a3a88.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2101.861458] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2101.861458] env[62824]: DEBUG oslo_vmware.api [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]529941ad-f4e8-05a9-a649-177d7c4fb502, 'name': SearchDatastore_Task, 'duration_secs': 0.016005} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2101.861658] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5e5e0955-f339-474e-a1fe-3ed839a3ebfb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2101.864267] env[62824]: DEBUG oslo_concurrency.lockutils [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2101.864497] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2101.864720] env[62824]: DEBUG oslo_concurrency.lockutils [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2101.864992] env[62824]: DEBUG oslo_concurrency.lockutils [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2101.865079] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2101.865864] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-97117224-55f4-4ff7-a3e4-c2464e2de2da {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2101.872474] env[62824]: DEBUG oslo_vmware.api [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for the task: (returnval){ [ 2101.872474] env[62824]: value = "task-2146316" [ 2101.872474] env[62824]: _type = "Task" [ 2101.872474] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2101.876832] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2101.877055] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2101.879312] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-65b9c1db-2a49-4520-bd73-bedb722756ea {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2101.883818] env[62824]: DEBUG oslo_vmware.api [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146316, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2101.886464] env[62824]: DEBUG oslo_vmware.api [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Waiting for the task: (returnval){ [ 2101.886464] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52eafde3-7596-5443-5dae-2e917570fdff" [ 2101.886464] env[62824]: _type = "Task" [ 2101.886464] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2101.894267] env[62824]: DEBUG oslo_vmware.api [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52eafde3-7596-5443-5dae-2e917570fdff, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2102.167072] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.788s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2102.167072] env[62824]: DEBUG nova.compute.manager [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2102.176042] env[62824]: DEBUG oslo_concurrency.lockutils [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.746s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2102.176042] env[62824]: INFO nova.compute.claims [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2102.176042] env[62824]: DEBUG nova.compute.manager [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] [instance: dadf590a-7288-4dd2-90de-125fe272f4ba] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2102.298659] env[62824]: DEBUG oslo_vmware.api [None req-4a76d89b-cbff-4ca0-a9aa-3ebc58bfda76 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146315, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2102.388131] env[62824]: DEBUG oslo_vmware.api [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146316, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.081276} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2102.392208] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2102.393342] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26fe7843-9b6a-4d63-bea7-a03e5b105cf9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2102.404810] env[62824]: DEBUG oslo_vmware.api [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52eafde3-7596-5443-5dae-2e917570fdff, 'name': SearchDatastore_Task, 'duration_secs': 0.019295} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2102.431373] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Reconfiguring VM instance instance-00000018 to attach disk [datastore2] 07ce931f-45ef-409b-b714-9f1cd47a3a88/07ce931f-45ef-409b-b714-9f1cd47a3a88.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2102.431964] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-13ad4a56-4bfa-4398-a86f-7e0295f0a6c0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2102.435682] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-68be3272-f45e-4ffd-b2f2-02167b9f7560 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2102.455216] env[62824]: DEBUG oslo_vmware.api [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Waiting for the task: (returnval){ [ 2102.455216] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]526ef776-1227-185c-5a80-3254f84e886f" [ 2102.455216] env[62824]: _type = "Task" [ 2102.455216] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2102.459914] env[62824]: DEBUG oslo_vmware.api [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for the task: (returnval){ [ 2102.459914] env[62824]: value = "task-2146317" [ 2102.459914] env[62824]: _type = "Task" [ 2102.459914] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2102.468259] env[62824]: DEBUG oslo_vmware.api [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]526ef776-1227-185c-5a80-3254f84e886f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2102.473915] env[62824]: DEBUG oslo_vmware.api [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146317, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2102.474884] env[62824]: DEBUG nova.network.neutron [req-d684e001-e872-4375-aca0-fbad54924094 req-ac87ca48-2f76-42d8-b8eb-ce066dad79fd service nova] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Updated VIF entry in instance network info cache for port 8e8b16ef-e3f8-48c1-aab6-51d94d356c77. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2102.475268] env[62824]: DEBUG nova.network.neutron [req-d684e001-e872-4375-aca0-fbad54924094 req-ac87ca48-2f76-42d8-b8eb-ce066dad79fd service nova] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Updating instance_info_cache with network_info: [{"id": "8e8b16ef-e3f8-48c1-aab6-51d94d356c77", "address": "fa:16:3e:48:83:8c", "network": {"id": "bcf30d77-cac1-4deb-971d-efb7c1fb8354", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1030043097-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "0503b434636c4ef5bae8db1b0c74d2ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3cc0a33d-17c0-4b87-b48f-413a87a4cc6a", "external-id": "nsx-vlan-transportzone-865", "segmentation_id": 865, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8e8b16ef-e3", "ovs_interfaceid": "8e8b16ef-e3f8-48c1-aab6-51d94d356c77", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2102.672079] env[62824]: DEBUG nova.compute.utils [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2102.678023] env[62824]: DEBUG nova.compute.manager [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2102.678023] env[62824]: DEBUG nova.network.neutron [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2102.705322] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2102.775550] env[62824]: DEBUG nova.policy [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9cc01135021d4a90a718dc605ce98af2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '815ecf9340934733a83d4dcb26612fd4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2102.799670] env[62824]: DEBUG oslo_vmware.api [None req-4a76d89b-cbff-4ca0-a9aa-3ebc58bfda76 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146315, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2102.968723] env[62824]: DEBUG oslo_vmware.api [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]526ef776-1227-185c-5a80-3254f84e886f, 'name': SearchDatastore_Task, 'duration_secs': 0.027761} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2102.969428] env[62824]: DEBUG oslo_concurrency.lockutils [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2102.969680] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 363177c0-dc40-429a-a74b-e690da133edb/363177c0-dc40-429a-a74b-e690da133edb.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2102.969933] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f0d2e050-1dc8-418f-952a-59b72914e961 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2102.975201] env[62824]: DEBUG oslo_vmware.api [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146317, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2102.979376] env[62824]: DEBUG oslo_concurrency.lockutils [req-d684e001-e872-4375-aca0-fbad54924094 req-ac87ca48-2f76-42d8-b8eb-ce066dad79fd service nova] Releasing lock "refresh_cache-363177c0-dc40-429a-a74b-e690da133edb" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2102.983038] env[62824]: DEBUG oslo_vmware.api [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Waiting for the task: (returnval){ [ 2102.983038] env[62824]: value = "task-2146318" [ 2102.983038] env[62824]: _type = "Task" [ 2102.983038] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2102.994663] env[62824]: DEBUG oslo_vmware.api [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146318, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2103.180972] env[62824]: DEBUG nova.compute.manager [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2103.305332] env[62824]: DEBUG oslo_vmware.api [None req-4a76d89b-cbff-4ca0-a9aa-3ebc58bfda76 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146315, 'name': CreateSnapshot_Task, 'duration_secs': 1.338979} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2103.305722] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-4a76d89b-cbff-4ca0-a9aa-3ebc58bfda76 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8] Created Snapshot of the VM instance {{(pid=62824) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 2103.306893] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-119bd128-c032-4b48-98e2-6793f18ff85a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2103.425640] env[62824]: DEBUG nova.network.neutron [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] Successfully created port: 3eb965bf-7ceb-4a51-bc6f-00803bc6b99e {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2103.477516] env[62824]: DEBUG oslo_vmware.api [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146317, 'name': ReconfigVM_Task, 'duration_secs': 0.62352} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2103.478040] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Reconfigured VM instance instance-00000018 to attach disk [datastore2] 07ce931f-45ef-409b-b714-9f1cd47a3a88/07ce931f-45ef-409b-b714-9f1cd47a3a88.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2103.478817] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0fb81768-031a-4135-b769-9f9a2d450e3e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2103.493200] env[62824]: DEBUG oslo_vmware.api [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for the task: (returnval){ [ 2103.493200] env[62824]: value = "task-2146319" [ 2103.493200] env[62824]: _type = "Task" [ 2103.493200] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2103.501953] env[62824]: DEBUG oslo_vmware.api [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146318, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.509817} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2103.502661] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 363177c0-dc40-429a-a74b-e690da133edb/363177c0-dc40-429a-a74b-e690da133edb.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2103.502845] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2103.503128] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ecf34863-29f1-449c-a7ef-a10f275cc972 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2103.508968] env[62824]: DEBUG oslo_vmware.api [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146319, 'name': Rename_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2103.516235] env[62824]: DEBUG oslo_vmware.api [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Waiting for the task: (returnval){ [ 2103.516235] env[62824]: value = "task-2146320" [ 2103.516235] env[62824]: _type = "Task" [ 2103.516235] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2103.527867] env[62824]: DEBUG oslo_vmware.api [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146320, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2103.722579] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquiring lock "fc77c44a-180c-46ff-9690-9072c6213c91" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2103.723392] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lock "fc77c44a-180c-46ff-9690-9072c6213c91" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2103.776919] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6e1ad5f-aede-4ac6-b3e0-3ccfd7be8fa3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2103.787798] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e361821-fe6a-4990-8a48-f80268b92910 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2103.822193] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05487645-a45e-48ce-864a-0b29b86dc6d7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2103.833453] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-4a76d89b-cbff-4ca0-a9aa-3ebc58bfda76 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8] Creating linked-clone VM from snapshot {{(pid=62824) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 2103.836044] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-7a6293e9-6497-43c5-9e59-a684824b9e2c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2103.840340] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-167784c1-e279-404e-be36-0787c0103a1d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2103.856272] env[62824]: DEBUG nova.compute.provider_tree [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2103.857832] env[62824]: DEBUG oslo_vmware.api [None req-4a76d89b-cbff-4ca0-a9aa-3ebc58bfda76 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the task: (returnval){ [ 2103.857832] env[62824]: value = "task-2146321" [ 2103.857832] env[62824]: _type = "Task" [ 2103.857832] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2103.867564] env[62824]: DEBUG oslo_vmware.api [None req-4a76d89b-cbff-4ca0-a9aa-3ebc58bfda76 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146321, 'name': CloneVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2104.013622] env[62824]: DEBUG oslo_vmware.api [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146319, 'name': Rename_Task, 'duration_secs': 0.162257} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2104.013952] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2104.014277] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-25ed3393-da76-48a6-b4fe-0ed5c7cdc3e1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2104.022679] env[62824]: DEBUG oslo_vmware.api [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for the task: (returnval){ [ 2104.022679] env[62824]: value = "task-2146322" [ 2104.022679] env[62824]: _type = "Task" [ 2104.022679] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2104.035622] env[62824]: DEBUG oslo_vmware.api [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146320, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065604} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2104.036377] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2104.037747] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86468f11-246f-4dbd-8dcd-366e76c23258 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2104.043650] env[62824]: DEBUG oslo_vmware.api [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146322, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2104.063143] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Reconfiguring VM instance instance-00000041 to attach disk [datastore2] 363177c0-dc40-429a-a74b-e690da133edb/363177c0-dc40-429a-a74b-e690da133edb.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2104.063495] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-541c7b5d-662a-4650-a27f-420261969369 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2104.089849] env[62824]: DEBUG oslo_vmware.api [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Waiting for the task: (returnval){ [ 2104.089849] env[62824]: value = "task-2146323" [ 2104.089849] env[62824]: _type = "Task" [ 2104.089849] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2104.099346] env[62824]: DEBUG oslo_vmware.api [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146323, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2104.111239] env[62824]: DEBUG nova.compute.manager [req-78e34d2d-d748-42fb-bb05-7fbe2ae7cd42 req-f7fbb8d5-8758-4f38-9c8f-feb2b4752ffa service nova] [instance: 350cfd12-a8d5-4a5f-b3de-d795fa179dfd] Received event network-changed-75b0f475-5e6e-44de-9480-a71716ecb4e0 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2104.111510] env[62824]: DEBUG nova.compute.manager [req-78e34d2d-d748-42fb-bb05-7fbe2ae7cd42 req-f7fbb8d5-8758-4f38-9c8f-feb2b4752ffa service nova] [instance: 350cfd12-a8d5-4a5f-b3de-d795fa179dfd] Refreshing instance network info cache due to event network-changed-75b0f475-5e6e-44de-9480-a71716ecb4e0. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2104.111747] env[62824]: DEBUG oslo_concurrency.lockutils [req-78e34d2d-d748-42fb-bb05-7fbe2ae7cd42 req-f7fbb8d5-8758-4f38-9c8f-feb2b4752ffa service nova] Acquiring lock "refresh_cache-350cfd12-a8d5-4a5f-b3de-d795fa179dfd" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2104.112101] env[62824]: DEBUG oslo_concurrency.lockutils [req-78e34d2d-d748-42fb-bb05-7fbe2ae7cd42 req-f7fbb8d5-8758-4f38-9c8f-feb2b4752ffa service nova] Acquired lock "refresh_cache-350cfd12-a8d5-4a5f-b3de-d795fa179dfd" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2104.112360] env[62824]: DEBUG nova.network.neutron [req-78e34d2d-d748-42fb-bb05-7fbe2ae7cd42 req-f7fbb8d5-8758-4f38-9c8f-feb2b4752ffa service nova] [instance: 350cfd12-a8d5-4a5f-b3de-d795fa179dfd] Refreshing network info cache for port 75b0f475-5e6e-44de-9480-a71716ecb4e0 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2104.193422] env[62824]: DEBUG nova.compute.manager [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2104.232780] env[62824]: DEBUG nova.virt.hardware [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2104.233045] env[62824]: DEBUG nova.virt.hardware [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2104.233212] env[62824]: DEBUG nova.virt.hardware [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2104.233393] env[62824]: DEBUG nova.virt.hardware [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2104.233705] env[62824]: DEBUG nova.virt.hardware [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2104.233924] env[62824]: DEBUG nova.virt.hardware [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2104.234211] env[62824]: DEBUG nova.virt.hardware [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2104.234404] env[62824]: DEBUG nova.virt.hardware [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2104.234574] env[62824]: DEBUG nova.virt.hardware [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2104.234762] env[62824]: DEBUG nova.virt.hardware [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2104.234940] env[62824]: DEBUG nova.virt.hardware [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2104.235836] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb1b3a12-4351-46c6-9e67-2b1c1e8f0746 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2104.245332] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c6805f9-6338-46f3-8cec-7025d118291d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2104.359700] env[62824]: DEBUG nova.scheduler.client.report [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2104.376719] env[62824]: DEBUG oslo_vmware.api [None req-4a76d89b-cbff-4ca0-a9aa-3ebc58bfda76 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146321, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2104.534852] env[62824]: DEBUG oslo_vmware.api [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146322, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2104.600864] env[62824]: DEBUG oslo_vmware.api [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146323, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2104.865270] env[62824]: DEBUG nova.network.neutron [req-78e34d2d-d748-42fb-bb05-7fbe2ae7cd42 req-f7fbb8d5-8758-4f38-9c8f-feb2b4752ffa service nova] [instance: 350cfd12-a8d5-4a5f-b3de-d795fa179dfd] Updated VIF entry in instance network info cache for port 75b0f475-5e6e-44de-9480-a71716ecb4e0. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2104.865753] env[62824]: DEBUG nova.network.neutron [req-78e34d2d-d748-42fb-bb05-7fbe2ae7cd42 req-f7fbb8d5-8758-4f38-9c8f-feb2b4752ffa service nova] [instance: 350cfd12-a8d5-4a5f-b3de-d795fa179dfd] Updating instance_info_cache with network_info: [{"id": "75b0f475-5e6e-44de-9480-a71716ecb4e0", "address": "fa:16:3e:fb:93:96", "network": {"id": "60037885-38d9-4586-9fd7-4907321e9c7e", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-370573912-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2e59203d676d48c8bd8402af4eb58f65", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f856fca-9fb5-41ea-a057-ac4193bd323d", "external-id": "nsx-vlan-transportzone-148", "segmentation_id": 148, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap75b0f475-5e", "ovs_interfaceid": "75b0f475-5e6e-44de-9480-a71716ecb4e0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2104.873400] env[62824]: DEBUG oslo_concurrency.lockutils [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.703s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2104.873939] env[62824]: DEBUG nova.compute.manager [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2104.878963] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fb4bd00b-e6f9-4459-a9db-0d3353b82d74 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.080s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2104.878963] env[62824]: DEBUG nova.objects.instance [None req-fb4bd00b-e6f9-4459-a9db-0d3353b82d74 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Lazy-loading 'resources' on Instance uuid dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2104.886049] env[62824]: DEBUG oslo_vmware.api [None req-4a76d89b-cbff-4ca0-a9aa-3ebc58bfda76 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146321, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2105.041035] env[62824]: DEBUG oslo_vmware.api [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146322, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2105.102276] env[62824]: DEBUG oslo_vmware.api [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146323, 'name': ReconfigVM_Task, 'duration_secs': 0.95342} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2105.102992] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Reconfigured VM instance instance-00000041 to attach disk [datastore2] 363177c0-dc40-429a-a74b-e690da133edb/363177c0-dc40-429a-a74b-e690da133edb.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2105.103645] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-57f84997-8e1c-458d-943e-2e27fa2347c8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2105.111185] env[62824]: DEBUG oslo_vmware.api [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Waiting for the task: (returnval){ [ 2105.111185] env[62824]: value = "task-2146324" [ 2105.111185] env[62824]: _type = "Task" [ 2105.111185] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2105.123568] env[62824]: DEBUG oslo_vmware.api [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146324, 'name': Rename_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2105.374135] env[62824]: DEBUG oslo_concurrency.lockutils [req-78e34d2d-d748-42fb-bb05-7fbe2ae7cd42 req-f7fbb8d5-8758-4f38-9c8f-feb2b4752ffa service nova] Releasing lock "refresh_cache-350cfd12-a8d5-4a5f-b3de-d795fa179dfd" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2105.374427] env[62824]: DEBUG nova.compute.manager [req-78e34d2d-d748-42fb-bb05-7fbe2ae7cd42 req-f7fbb8d5-8758-4f38-9c8f-feb2b4752ffa service nova] [instance: 350cfd12-a8d5-4a5f-b3de-d795fa179dfd] Received event network-changed-75b0f475-5e6e-44de-9480-a71716ecb4e0 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2105.374606] env[62824]: DEBUG nova.compute.manager [req-78e34d2d-d748-42fb-bb05-7fbe2ae7cd42 req-f7fbb8d5-8758-4f38-9c8f-feb2b4752ffa service nova] [instance: 350cfd12-a8d5-4a5f-b3de-d795fa179dfd] Refreshing instance network info cache due to event network-changed-75b0f475-5e6e-44de-9480-a71716ecb4e0. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2105.374813] env[62824]: DEBUG oslo_concurrency.lockutils [req-78e34d2d-d748-42fb-bb05-7fbe2ae7cd42 req-f7fbb8d5-8758-4f38-9c8f-feb2b4752ffa service nova] Acquiring lock "refresh_cache-350cfd12-a8d5-4a5f-b3de-d795fa179dfd" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2105.374950] env[62824]: DEBUG oslo_concurrency.lockutils [req-78e34d2d-d748-42fb-bb05-7fbe2ae7cd42 req-f7fbb8d5-8758-4f38-9c8f-feb2b4752ffa service nova] Acquired lock "refresh_cache-350cfd12-a8d5-4a5f-b3de-d795fa179dfd" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2105.375124] env[62824]: DEBUG nova.network.neutron [req-78e34d2d-d748-42fb-bb05-7fbe2ae7cd42 req-f7fbb8d5-8758-4f38-9c8f-feb2b4752ffa service nova] [instance: 350cfd12-a8d5-4a5f-b3de-d795fa179dfd] Refreshing network info cache for port 75b0f475-5e6e-44de-9480-a71716ecb4e0 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2105.380935] env[62824]: DEBUG oslo_vmware.api [None req-4a76d89b-cbff-4ca0-a9aa-3ebc58bfda76 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146321, 'name': CloneVM_Task} progress is 95%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2105.385744] env[62824]: DEBUG nova.compute.utils [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2105.387427] env[62824]: DEBUG nova.compute.manager [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2105.387755] env[62824]: DEBUG nova.network.neutron [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2105.465537] env[62824]: DEBUG nova.policy [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5a5891a75fee4caf92d4bba152254ad5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7fcdbfc29df64a1ba8d982bdcc667b64', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2105.553506] env[62824]: DEBUG oslo_vmware.api [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146322, 'name': PowerOnVM_Task, 'duration_secs': 1.074825} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2105.553506] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2105.553506] env[62824]: DEBUG nova.compute.manager [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2105.554519] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71aa73de-4639-47b1-96b7-880af80a340c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2105.625247] env[62824]: DEBUG oslo_vmware.api [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146324, 'name': Rename_Task, 'duration_secs': 0.193605} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2105.625532] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2105.625778] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-748819de-9c64-4a0e-af42-ac6705614010 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2105.634287] env[62824]: DEBUG oslo_vmware.api [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Waiting for the task: (returnval){ [ 2105.634287] env[62824]: value = "task-2146325" [ 2105.634287] env[62824]: _type = "Task" [ 2105.634287] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2105.645894] env[62824]: DEBUG oslo_vmware.api [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146325, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2105.732031] env[62824]: DEBUG nova.network.neutron [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] Successfully updated port: 3eb965bf-7ceb-4a51-bc6f-00803bc6b99e {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2105.879721] env[62824]: DEBUG oslo_vmware.api [None req-4a76d89b-cbff-4ca0-a9aa-3ebc58bfda76 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146321, 'name': CloneVM_Task, 'duration_secs': 1.652045} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2105.880154] env[62824]: INFO nova.virt.vmwareapi.vmops [None req-4a76d89b-cbff-4ca0-a9aa-3ebc58bfda76 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8] Created linked-clone VM from snapshot [ 2105.881492] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96099f7f-6fd2-4335-9b47-b986006dff5c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2105.897530] env[62824]: DEBUG nova.compute.manager [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2105.900707] env[62824]: DEBUG nova.virt.vmwareapi.images [None req-4a76d89b-cbff-4ca0-a9aa-3ebc58bfda76 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8] Uploading image 758d9d98-91b8-46a1-9c95-5705c074be1a {{(pid=62824) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 2105.903322] env[62824]: DEBUG nova.network.neutron [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Successfully created port: 7be95c5e-6e98-4fec-8550-1c22a787b95b {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2105.952028] env[62824]: DEBUG oslo_vmware.rw_handles [None req-4a76d89b-cbff-4ca0-a9aa-3ebc58bfda76 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 2105.952028] env[62824]: value = "vm-438675" [ 2105.952028] env[62824]: _type = "VirtualMachine" [ 2105.952028] env[62824]: }. {{(pid=62824) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 2105.957024] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-1d5d4209-4ab1-4919-8dca-10788a6b7712 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2105.972446] env[62824]: DEBUG oslo_vmware.rw_handles [None req-4a76d89b-cbff-4ca0-a9aa-3ebc58bfda76 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Lease: (returnval){ [ 2105.972446] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5272ca6f-5d50-63ac-afbd-35d66ad14d6d" [ 2105.972446] env[62824]: _type = "HttpNfcLease" [ 2105.972446] env[62824]: } obtained for exporting VM: (result){ [ 2105.972446] env[62824]: value = "vm-438675" [ 2105.972446] env[62824]: _type = "VirtualMachine" [ 2105.972446] env[62824]: }. {{(pid=62824) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 2105.973411] env[62824]: DEBUG oslo_vmware.api [None req-4a76d89b-cbff-4ca0-a9aa-3ebc58bfda76 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the lease: (returnval){ [ 2105.973411] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5272ca6f-5d50-63ac-afbd-35d66ad14d6d" [ 2105.973411] env[62824]: _type = "HttpNfcLease" [ 2105.973411] env[62824]: } to be ready. {{(pid=62824) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 2105.991116] env[62824]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 2105.991116] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5272ca6f-5d50-63ac-afbd-35d66ad14d6d" [ 2105.991116] env[62824]: _type = "HttpNfcLease" [ 2105.991116] env[62824]: } is initializing. {{(pid=62824) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 2105.991341] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Acquiring lock "bd7f2ed5-fa08-4a48-9755-60b61791ea0b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2105.991448] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Lock "bd7f2ed5-fa08-4a48-9755-60b61791ea0b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2106.027759] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d482fa4-db41-4fa2-9600-44ac35e65bde {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2106.039655] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30b4c11f-b58e-49e1-8925-3eb8b1c94847 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2106.080070] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d6882d3-1bf9-4603-bb57-348b652d622b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2106.091575] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57548265-6869-4ea7-8214-180cc6bfd243 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2106.099853] env[62824]: DEBUG oslo_concurrency.lockutils [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2106.110704] env[62824]: DEBUG nova.compute.provider_tree [None req-fb4bd00b-e6f9-4459-a9db-0d3353b82d74 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2106.145519] env[62824]: DEBUG oslo_vmware.api [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146325, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2106.236153] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquiring lock "refresh_cache-ca42fef0-1e90-4ab6-9d60-8ef7e4997884" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2106.236408] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquired lock "refresh_cache-ca42fef0-1e90-4ab6-9d60-8ef7e4997884" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2106.236493] env[62824]: DEBUG nova.network.neutron [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2106.366571] env[62824]: DEBUG nova.network.neutron [req-78e34d2d-d748-42fb-bb05-7fbe2ae7cd42 req-f7fbb8d5-8758-4f38-9c8f-feb2b4752ffa service nova] [instance: 350cfd12-a8d5-4a5f-b3de-d795fa179dfd] Updated VIF entry in instance network info cache for port 75b0f475-5e6e-44de-9480-a71716ecb4e0. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2106.366728] env[62824]: DEBUG nova.network.neutron [req-78e34d2d-d748-42fb-bb05-7fbe2ae7cd42 req-f7fbb8d5-8758-4f38-9c8f-feb2b4752ffa service nova] [instance: 350cfd12-a8d5-4a5f-b3de-d795fa179dfd] Updating instance_info_cache with network_info: [{"id": "75b0f475-5e6e-44de-9480-a71716ecb4e0", "address": "fa:16:3e:fb:93:96", "network": {"id": "60037885-38d9-4586-9fd7-4907321e9c7e", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-370573912-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2e59203d676d48c8bd8402af4eb58f65", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f856fca-9fb5-41ea-a057-ac4193bd323d", "external-id": "nsx-vlan-transportzone-148", "segmentation_id": 148, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap75b0f475-5e", "ovs_interfaceid": "75b0f475-5e6e-44de-9480-a71716ecb4e0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2106.372416] env[62824]: DEBUG nova.compute.manager [req-f0892d10-9b25-45da-8f73-a0499a271abf req-21a7dd6d-7d22-4cac-be6d-d437eeef2e46 service nova] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] Received event network-vif-plugged-3eb965bf-7ceb-4a51-bc6f-00803bc6b99e {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2106.372647] env[62824]: DEBUG oslo_concurrency.lockutils [req-f0892d10-9b25-45da-8f73-a0499a271abf req-21a7dd6d-7d22-4cac-be6d-d437eeef2e46 service nova] Acquiring lock "ca42fef0-1e90-4ab6-9d60-8ef7e4997884-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2106.372891] env[62824]: DEBUG oslo_concurrency.lockutils [req-f0892d10-9b25-45da-8f73-a0499a271abf req-21a7dd6d-7d22-4cac-be6d-d437eeef2e46 service nova] Lock "ca42fef0-1e90-4ab6-9d60-8ef7e4997884-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2106.373024] env[62824]: DEBUG oslo_concurrency.lockutils [req-f0892d10-9b25-45da-8f73-a0499a271abf req-21a7dd6d-7d22-4cac-be6d-d437eeef2e46 service nova] Lock "ca42fef0-1e90-4ab6-9d60-8ef7e4997884-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2106.373188] env[62824]: DEBUG nova.compute.manager [req-f0892d10-9b25-45da-8f73-a0499a271abf req-21a7dd6d-7d22-4cac-be6d-d437eeef2e46 service nova] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] No waiting events found dispatching network-vif-plugged-3eb965bf-7ceb-4a51-bc6f-00803bc6b99e {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2106.373352] env[62824]: WARNING nova.compute.manager [req-f0892d10-9b25-45da-8f73-a0499a271abf req-21a7dd6d-7d22-4cac-be6d-d437eeef2e46 service nova] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] Received unexpected event network-vif-plugged-3eb965bf-7ceb-4a51-bc6f-00803bc6b99e for instance with vm_state building and task_state spawning. [ 2106.373534] env[62824]: DEBUG nova.compute.manager [req-f0892d10-9b25-45da-8f73-a0499a271abf req-21a7dd6d-7d22-4cac-be6d-d437eeef2e46 service nova] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] Received event network-changed-3eb965bf-7ceb-4a51-bc6f-00803bc6b99e {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2106.373696] env[62824]: DEBUG nova.compute.manager [req-f0892d10-9b25-45da-8f73-a0499a271abf req-21a7dd6d-7d22-4cac-be6d-d437eeef2e46 service nova] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] Refreshing instance network info cache due to event network-changed-3eb965bf-7ceb-4a51-bc6f-00803bc6b99e. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2106.373814] env[62824]: DEBUG oslo_concurrency.lockutils [req-f0892d10-9b25-45da-8f73-a0499a271abf req-21a7dd6d-7d22-4cac-be6d-d437eeef2e46 service nova] Acquiring lock "refresh_cache-ca42fef0-1e90-4ab6-9d60-8ef7e4997884" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2106.481763] env[62824]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 2106.481763] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5272ca6f-5d50-63ac-afbd-35d66ad14d6d" [ 2106.481763] env[62824]: _type = "HttpNfcLease" [ 2106.481763] env[62824]: } is ready. {{(pid=62824) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 2106.482193] env[62824]: DEBUG oslo_vmware.rw_handles [None req-4a76d89b-cbff-4ca0-a9aa-3ebc58bfda76 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 2106.482193] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5272ca6f-5d50-63ac-afbd-35d66ad14d6d" [ 2106.482193] env[62824]: _type = "HttpNfcLease" [ 2106.482193] env[62824]: }. {{(pid=62824) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 2106.482836] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94bc2b6e-81ae-44db-9d35-ff78eb46e341 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2106.491142] env[62824]: DEBUG oslo_vmware.rw_handles [None req-4a76d89b-cbff-4ca0-a9aa-3ebc58bfda76 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52f9e045-1c7a-e2c9-c853-b2b6a239bfe8/disk-0.vmdk from lease info. {{(pid=62824) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 2106.491333] env[62824]: DEBUG oslo_vmware.rw_handles [None req-4a76d89b-cbff-4ca0-a9aa-3ebc58bfda76 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52f9e045-1c7a-e2c9-c853-b2b6a239bfe8/disk-0.vmdk for reading. {{(pid=62824) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 2106.597226] env[62824]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-4443e352-a437-4a13-9ed3-579dc336b6df {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2106.616465] env[62824]: DEBUG nova.scheduler.client.report [None req-fb4bd00b-e6f9-4459-a9db-0d3353b82d74 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2106.657245] env[62824]: DEBUG oslo_vmware.api [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146325, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2106.786099] env[62824]: DEBUG nova.network.neutron [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2106.869420] env[62824]: DEBUG oslo_concurrency.lockutils [req-78e34d2d-d748-42fb-bb05-7fbe2ae7cd42 req-f7fbb8d5-8758-4f38-9c8f-feb2b4752ffa service nova] Releasing lock "refresh_cache-350cfd12-a8d5-4a5f-b3de-d795fa179dfd" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2106.921385] env[62824]: DEBUG nova.compute.manager [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2106.950922] env[62824]: DEBUG nova.virt.hardware [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2106.951227] env[62824]: DEBUG nova.virt.hardware [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2106.951451] env[62824]: DEBUG nova.virt.hardware [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2106.951633] env[62824]: DEBUG nova.virt.hardware [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2106.951862] env[62824]: DEBUG nova.virt.hardware [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2106.953017] env[62824]: DEBUG nova.virt.hardware [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2106.954090] env[62824]: DEBUG nova.virt.hardware [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2106.955992] env[62824]: DEBUG nova.virt.hardware [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2106.956333] env[62824]: DEBUG nova.virt.hardware [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2106.956574] env[62824]: DEBUG nova.virt.hardware [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2106.957189] env[62824]: DEBUG nova.virt.hardware [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2106.957946] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74267e02-448a-4f4e-9770-8ed51bddf82d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2106.972040] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53f37c35-4739-454b-b39a-5fb6b29d632f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2107.096211] env[62824]: DEBUG nova.network.neutron [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] Updating instance_info_cache with network_info: [{"id": "3eb965bf-7ceb-4a51-bc6f-00803bc6b99e", "address": "fa:16:3e:3f:e9:09", "network": {"id": "b4ed285c-26a7-461b-8a71-d18e07582ab8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-997150974-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "815ecf9340934733a83d4dcb26612fd4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69f65356-c85e-4b7f-ad28-7c7b5e8cf50c", "external-id": "nsx-vlan-transportzone-281", "segmentation_id": 281, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3eb965bf-7c", "ovs_interfaceid": "3eb965bf-7ceb-4a51-bc6f-00803bc6b99e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2107.123632] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fb4bd00b-e6f9-4459-a9db-0d3353b82d74 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.246s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2107.127977] env[62824]: DEBUG oslo_concurrency.lockutils [None req-56d14257-fc28-4784-a770-ae333154aa8d tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.325s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2107.128653] env[62824]: DEBUG nova.objects.instance [None req-56d14257-fc28-4784-a770-ae333154aa8d tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Lazy-loading 'resources' on Instance uuid c36ece43-3d70-4e67-a740-9057f413c722 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2107.152966] env[62824]: DEBUG oslo_vmware.api [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146325, 'name': PowerOnVM_Task, 'duration_secs': 1.46927} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2107.153266] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2107.153544] env[62824]: INFO nova.compute.manager [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Took 10.18 seconds to spawn the instance on the hypervisor. [ 2107.154141] env[62824]: DEBUG nova.compute.manager [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2107.154655] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c8d52cf-f1a3-4978-b98f-0c9f67192ba5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2107.165699] env[62824]: INFO nova.scheduler.client.report [None req-fb4bd00b-e6f9-4459-a9db-0d3353b82d74 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Deleted allocations for instance dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda [ 2107.548699] env[62824]: INFO nova.compute.manager [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Rebuilding instance [ 2107.602028] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Releasing lock "refresh_cache-ca42fef0-1e90-4ab6-9d60-8ef7e4997884" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2107.602028] env[62824]: DEBUG nova.compute.manager [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] Instance network_info: |[{"id": "3eb965bf-7ceb-4a51-bc6f-00803bc6b99e", "address": "fa:16:3e:3f:e9:09", "network": {"id": "b4ed285c-26a7-461b-8a71-d18e07582ab8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-997150974-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "815ecf9340934733a83d4dcb26612fd4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69f65356-c85e-4b7f-ad28-7c7b5e8cf50c", "external-id": "nsx-vlan-transportzone-281", "segmentation_id": 281, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3eb965bf-7c", "ovs_interfaceid": "3eb965bf-7ceb-4a51-bc6f-00803bc6b99e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2107.602028] env[62824]: DEBUG oslo_concurrency.lockutils [req-f0892d10-9b25-45da-8f73-a0499a271abf req-21a7dd6d-7d22-4cac-be6d-d437eeef2e46 service nova] Acquired lock "refresh_cache-ca42fef0-1e90-4ab6-9d60-8ef7e4997884" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2107.602028] env[62824]: DEBUG nova.network.neutron [req-f0892d10-9b25-45da-8f73-a0499a271abf req-21a7dd6d-7d22-4cac-be6d-d437eeef2e46 service nova] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] Refreshing network info cache for port 3eb965bf-7ceb-4a51-bc6f-00803bc6b99e {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2107.602028] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3f:e9:09', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '69f65356-c85e-4b7f-ad28-7c7b5e8cf50c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3eb965bf-7ceb-4a51-bc6f-00803bc6b99e', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2107.611472] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Creating folder: Project (815ecf9340934733a83d4dcb26612fd4). Parent ref: group-v438503. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 2107.616556] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d52f531d-bb1b-430f-9768-4ff50d50cfdd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2107.621226] env[62824]: DEBUG nova.compute.manager [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2107.622589] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c29e9d2-ed14-409f-82d8-7f577f56c333 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2107.638884] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Created folder: Project (815ecf9340934733a83d4dcb26612fd4) in parent group-v438503. [ 2107.639188] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Creating folder: Instances. Parent ref: group-v438676. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 2107.640823] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-88cde296-e4d9-434c-9e20-57f844a9566c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2107.651944] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Created folder: Instances in parent group-v438676. [ 2107.652052] env[62824]: DEBUG oslo.service.loopingcall [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2107.652351] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2107.652960] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4c213863-55b2-490f-bab0-49d9545689ad {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2107.685427] env[62824]: DEBUG nova.compute.manager [req-eaeac17c-e697-4821-93bc-3d78f945efec req-1e5addec-e296-484b-bd3a-f70216d88f83 service nova] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Received event network-vif-plugged-7be95c5e-6e98-4fec-8550-1c22a787b95b {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2107.685710] env[62824]: DEBUG oslo_concurrency.lockutils [req-eaeac17c-e697-4821-93bc-3d78f945efec req-1e5addec-e296-484b-bd3a-f70216d88f83 service nova] Acquiring lock "5b3df31f-15fe-473f-992c-ddb272661c53-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2107.685925] env[62824]: DEBUG oslo_concurrency.lockutils [req-eaeac17c-e697-4821-93bc-3d78f945efec req-1e5addec-e296-484b-bd3a-f70216d88f83 service nova] Lock "5b3df31f-15fe-473f-992c-ddb272661c53-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2107.686110] env[62824]: DEBUG oslo_concurrency.lockutils [req-eaeac17c-e697-4821-93bc-3d78f945efec req-1e5addec-e296-484b-bd3a-f70216d88f83 service nova] Lock "5b3df31f-15fe-473f-992c-ddb272661c53-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2107.686664] env[62824]: DEBUG nova.compute.manager [req-eaeac17c-e697-4821-93bc-3d78f945efec req-1e5addec-e296-484b-bd3a-f70216d88f83 service nova] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] No waiting events found dispatching network-vif-plugged-7be95c5e-6e98-4fec-8550-1c22a787b95b {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2107.686664] env[62824]: WARNING nova.compute.manager [req-eaeac17c-e697-4821-93bc-3d78f945efec req-1e5addec-e296-484b-bd3a-f70216d88f83 service nova] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Received unexpected event network-vif-plugged-7be95c5e-6e98-4fec-8550-1c22a787b95b for instance with vm_state building and task_state spawning. [ 2107.689919] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fb4bd00b-e6f9-4459-a9db-0d3353b82d74 tempest-ServersTestManualDisk-1896826768 tempest-ServersTestManualDisk-1896826768-project-member] Lock "dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.872s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2107.691345] env[62824]: INFO nova.compute.manager [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Took 27.62 seconds to build instance. [ 2107.700259] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2107.700259] env[62824]: value = "task-2146329" [ 2107.700259] env[62824]: _type = "Task" [ 2107.700259] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2107.721148] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146329, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2107.757838] env[62824]: DEBUG nova.network.neutron [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Successfully updated port: 7be95c5e-6e98-4fec-8550-1c22a787b95b {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2108.193728] env[62824]: DEBUG oslo_concurrency.lockutils [None req-436cc472-788a-4df8-ad03-1a32e2023795 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Lock "363177c0-dc40-429a-a74b-e690da133edb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.129s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2108.214309] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146329, 'name': CreateVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2108.254803] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac272185-9078-43ca-9212-ebe6b01498cc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2108.266791] env[62824]: DEBUG oslo_concurrency.lockutils [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquiring lock "refresh_cache-5b3df31f-15fe-473f-992c-ddb272661c53" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2108.267737] env[62824]: DEBUG oslo_concurrency.lockutils [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquired lock "refresh_cache-5b3df31f-15fe-473f-992c-ddb272661c53" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2108.267737] env[62824]: DEBUG nova.network.neutron [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2108.271634] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4939ecb-3b67-4801-bd64-17ac11deec04 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2108.311950] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f114f66e-fed0-4433-a402-cd2d86330d47 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2108.321305] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3676ab03-1a34-497d-9010-fc2e49be6e2d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2108.342805] env[62824]: DEBUG nova.compute.provider_tree [None req-56d14257-fc28-4784-a770-ae333154aa8d tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2108.581313] env[62824]: DEBUG nova.network.neutron [req-f0892d10-9b25-45da-8f73-a0499a271abf req-21a7dd6d-7d22-4cac-be6d-d437eeef2e46 service nova] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] Updated VIF entry in instance network info cache for port 3eb965bf-7ceb-4a51-bc6f-00803bc6b99e. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2108.581711] env[62824]: DEBUG nova.network.neutron [req-f0892d10-9b25-45da-8f73-a0499a271abf req-21a7dd6d-7d22-4cac-be6d-d437eeef2e46 service nova] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] Updating instance_info_cache with network_info: [{"id": "3eb965bf-7ceb-4a51-bc6f-00803bc6b99e", "address": "fa:16:3e:3f:e9:09", "network": {"id": "b4ed285c-26a7-461b-8a71-d18e07582ab8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-997150974-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "815ecf9340934733a83d4dcb26612fd4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69f65356-c85e-4b7f-ad28-7c7b5e8cf50c", "external-id": "nsx-vlan-transportzone-281", "segmentation_id": 281, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3eb965bf-7c", "ovs_interfaceid": "3eb965bf-7ceb-4a51-bc6f-00803bc6b99e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2108.648020] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2108.648020] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-740105ba-eddb-4ffd-8529-526ed09ac516 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2108.655300] env[62824]: DEBUG oslo_vmware.api [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for the task: (returnval){ [ 2108.655300] env[62824]: value = "task-2146330" [ 2108.655300] env[62824]: _type = "Task" [ 2108.655300] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2108.665786] env[62824]: DEBUG oslo_vmware.api [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146330, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2108.705516] env[62824]: DEBUG nova.compute.manager [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7a2a51c4-e558-4f5f-b82c-718bc12c1df5] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2108.707312] env[62824]: INFO nova.compute.manager [None req-e9274965-3b8a-49a7-9276-3a0e6a892115 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Rescuing [ 2108.707744] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e9274965-3b8a-49a7-9276-3a0e6a892115 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Acquiring lock "refresh_cache-363177c0-dc40-429a-a74b-e690da133edb" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2108.708043] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e9274965-3b8a-49a7-9276-3a0e6a892115 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Acquired lock "refresh_cache-363177c0-dc40-429a-a74b-e690da133edb" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2108.708470] env[62824]: DEBUG nova.network.neutron [None req-e9274965-3b8a-49a7-9276-3a0e6a892115 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2108.722133] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146329, 'name': CreateVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2108.849024] env[62824]: DEBUG nova.scheduler.client.report [None req-56d14257-fc28-4784-a770-ae333154aa8d tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2109.042077] env[62824]: DEBUG nova.network.neutron [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2109.085147] env[62824]: DEBUG oslo_concurrency.lockutils [req-f0892d10-9b25-45da-8f73-a0499a271abf req-21a7dd6d-7d22-4cac-be6d-d437eeef2e46 service nova] Releasing lock "refresh_cache-ca42fef0-1e90-4ab6-9d60-8ef7e4997884" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2109.166334] env[62824]: DEBUG oslo_vmware.api [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146330, 'name': PowerOffVM_Task, 'duration_secs': 0.247441} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2109.166747] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2109.166926] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2109.167693] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b666a4fc-b333-465c-91b5-989d4252ac4e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2109.176271] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2109.176552] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-313d6233-c0f9-4d42-95ac-17251fa04757 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2109.227338] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146329, 'name': CreateVM_Task, 'duration_secs': 1.035574} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2109.227555] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2109.228506] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2109.228901] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2109.229391] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2109.229831] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-964b56bd-f221-4049-99b0-0f81bb2464f1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2109.235179] env[62824]: DEBUG oslo_vmware.api [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2109.235179] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52cabc83-8de0-54b7-5b7c-1419e84be892" [ 2109.235179] env[62824]: _type = "Task" [ 2109.235179] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2109.241026] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2109.245017] env[62824]: DEBUG oslo_vmware.api [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52cabc83-8de0-54b7-5b7c-1419e84be892, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2109.273374] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2109.273586] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Deleting contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2109.273771] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Deleting the datastore file [datastore2] 07ce931f-45ef-409b-b714-9f1cd47a3a88 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2109.274060] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b6cf8bc7-48fb-4e75-819d-99c2f52e080d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2109.281526] env[62824]: DEBUG oslo_vmware.api [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for the task: (returnval){ [ 2109.281526] env[62824]: value = "task-2146332" [ 2109.281526] env[62824]: _type = "Task" [ 2109.281526] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2109.288230] env[62824]: DEBUG nova.network.neutron [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Updating instance_info_cache with network_info: [{"id": "7be95c5e-6e98-4fec-8550-1c22a787b95b", "address": "fa:16:3e:e8:69:3a", "network": {"id": "bba9b98c-c7eb-4890-8de5-eafacfdebd7d", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1928394728-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7fcdbfc29df64a1ba8d982bdcc667b64", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b49c5024-2ced-42ca-90cc-6066766d43e6", "external-id": "nsx-vlan-transportzone-239", "segmentation_id": 239, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7be95c5e-6e", "ovs_interfaceid": "7be95c5e-6e98-4fec-8550-1c22a787b95b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2109.294323] env[62824]: DEBUG oslo_vmware.api [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146332, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2109.352978] env[62824]: DEBUG oslo_concurrency.lockutils [None req-56d14257-fc28-4784-a770-ae333154aa8d tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.225s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2109.356707] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.872s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2109.360071] env[62824]: INFO nova.compute.claims [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 49818910-61de-4fb8-bbab-d5d61d2f1ada] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2109.375406] env[62824]: INFO nova.scheduler.client.report [None req-56d14257-fc28-4784-a770-ae333154aa8d tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Deleted allocations for instance c36ece43-3d70-4e67-a740-9057f413c722 [ 2109.748372] env[62824]: DEBUG oslo_vmware.api [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52cabc83-8de0-54b7-5b7c-1419e84be892, 'name': SearchDatastore_Task, 'duration_secs': 0.010192} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2109.748962] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2109.749237] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2109.749491] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2109.749645] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2109.749829] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2109.750143] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e30911da-6581-450f-b5dd-ce5554022ff1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2109.761563] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2109.761563] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2109.761958] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-75561c0b-5494-447a-800c-78e82edd276f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2109.772562] env[62824]: DEBUG oslo_vmware.api [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2109.772562] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5228034d-b04c-7600-96f5-f3959f1661f8" [ 2109.772562] env[62824]: _type = "Task" [ 2109.772562] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2109.782501] env[62824]: DEBUG nova.network.neutron [None req-e9274965-3b8a-49a7-9276-3a0e6a892115 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Updating instance_info_cache with network_info: [{"id": "8e8b16ef-e3f8-48c1-aab6-51d94d356c77", "address": "fa:16:3e:48:83:8c", "network": {"id": "bcf30d77-cac1-4deb-971d-efb7c1fb8354", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1030043097-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "0503b434636c4ef5bae8db1b0c74d2ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3cc0a33d-17c0-4b87-b48f-413a87a4cc6a", "external-id": "nsx-vlan-transportzone-865", "segmentation_id": 865, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8e8b16ef-e3", "ovs_interfaceid": "8e8b16ef-e3f8-48c1-aab6-51d94d356c77", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2109.794748] env[62824]: DEBUG oslo_vmware.api [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5228034d-b04c-7600-96f5-f3959f1661f8, 'name': SearchDatastore_Task, 'duration_secs': 0.009754} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2109.794748] env[62824]: DEBUG oslo_concurrency.lockutils [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Releasing lock "refresh_cache-5b3df31f-15fe-473f-992c-ddb272661c53" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2109.794748] env[62824]: DEBUG nova.compute.manager [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Instance network_info: |[{"id": "7be95c5e-6e98-4fec-8550-1c22a787b95b", "address": "fa:16:3e:e8:69:3a", "network": {"id": "bba9b98c-c7eb-4890-8de5-eafacfdebd7d", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1928394728-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7fcdbfc29df64a1ba8d982bdcc667b64", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b49c5024-2ced-42ca-90cc-6066766d43e6", "external-id": "nsx-vlan-transportzone-239", "segmentation_id": 239, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7be95c5e-6e", "ovs_interfaceid": "7be95c5e-6e98-4fec-8550-1c22a787b95b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2109.796785] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e8:69:3a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b49c5024-2ced-42ca-90cc-6066766d43e6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7be95c5e-6e98-4fec-8550-1c22a787b95b', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2109.804745] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Creating folder: Project (7fcdbfc29df64a1ba8d982bdcc667b64). Parent ref: group-v438503. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 2109.812250] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-49790287-2b5b-4b37-8e15-29a907a0f574 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2109.812250] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f31833f1-d6e9-4f6c-a0ee-699331247020 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2109.814458] env[62824]: DEBUG oslo_vmware.api [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146332, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.17012} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2109.815012] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2109.815212] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Deleted contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2109.815386] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2109.825697] env[62824]: DEBUG oslo_vmware.api [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2109.825697] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]526c7887-8e0d-eefb-dc5c-e4954f5963d1" [ 2109.825697] env[62824]: _type = "Task" [ 2109.825697] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2109.830062] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Created folder: Project (7fcdbfc29df64a1ba8d982bdcc667b64) in parent group-v438503. [ 2109.830217] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Creating folder: Instances. Parent ref: group-v438679. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 2109.830971] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b55e9480-9301-4957-b971-9fa965a21696 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2109.839169] env[62824]: DEBUG oslo_vmware.api [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]526c7887-8e0d-eefb-dc5c-e4954f5963d1, 'name': SearchDatastore_Task, 'duration_secs': 0.009956} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2109.839169] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2109.839169] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] ca42fef0-1e90-4ab6-9d60-8ef7e4997884/ca42fef0-1e90-4ab6-9d60-8ef7e4997884.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2109.839169] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c4f5cb43-d575-4a8e-afa6-7ca87ff74243 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2109.846165] env[62824]: DEBUG oslo_vmware.api [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2109.846165] env[62824]: value = "task-2146335" [ 2109.846165] env[62824]: _type = "Task" [ 2109.846165] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2109.848210] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Created folder: Instances in parent group-v438679. [ 2109.849150] env[62824]: DEBUG oslo.service.loopingcall [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2109.851813] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2109.853121] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-19cdd6f0-8d8f-4f4a-90b4-c518301d64df {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2109.877568] env[62824]: DEBUG oslo_vmware.api [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146335, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2109.884317] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2109.884317] env[62824]: value = "task-2146336" [ 2109.884317] env[62824]: _type = "Task" [ 2109.884317] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2109.889184] env[62824]: DEBUG oslo_concurrency.lockutils [None req-56d14257-fc28-4784-a770-ae333154aa8d tempest-ServersAdminNegativeTestJSON-246685314 tempest-ServersAdminNegativeTestJSON-246685314-project-member] Lock "c36ece43-3d70-4e67-a740-9057f413c722" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.622s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2109.897994] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146336, 'name': CreateVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2110.293942] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e9274965-3b8a-49a7-9276-3a0e6a892115 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Releasing lock "refresh_cache-363177c0-dc40-429a-a74b-e690da133edb" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2110.349152] env[62824]: DEBUG nova.compute.manager [req-ebd60a84-bbd0-4da7-b659-0e00a424bc87 req-19df813e-66c9-4a1e-80a0-a4fddf4fe9ad service nova] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Received event network-changed-7be95c5e-6e98-4fec-8550-1c22a787b95b {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2110.349447] env[62824]: DEBUG nova.compute.manager [req-ebd60a84-bbd0-4da7-b659-0e00a424bc87 req-19df813e-66c9-4a1e-80a0-a4fddf4fe9ad service nova] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Refreshing instance network info cache due to event network-changed-7be95c5e-6e98-4fec-8550-1c22a787b95b. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2110.349689] env[62824]: DEBUG oslo_concurrency.lockutils [req-ebd60a84-bbd0-4da7-b659-0e00a424bc87 req-19df813e-66c9-4a1e-80a0-a4fddf4fe9ad service nova] Acquiring lock "refresh_cache-5b3df31f-15fe-473f-992c-ddb272661c53" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2110.349944] env[62824]: DEBUG oslo_concurrency.lockutils [req-ebd60a84-bbd0-4da7-b659-0e00a424bc87 req-19df813e-66c9-4a1e-80a0-a4fddf4fe9ad service nova] Acquired lock "refresh_cache-5b3df31f-15fe-473f-992c-ddb272661c53" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2110.350219] env[62824]: DEBUG nova.network.neutron [req-ebd60a84-bbd0-4da7-b659-0e00a424bc87 req-19df813e-66c9-4a1e-80a0-a4fddf4fe9ad service nova] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Refreshing network info cache for port 7be95c5e-6e98-4fec-8550-1c22a787b95b {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2110.366556] env[62824]: DEBUG oslo_vmware.api [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146335, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.459458} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2110.366887] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] ca42fef0-1e90-4ab6-9d60-8ef7e4997884/ca42fef0-1e90-4ab6-9d60-8ef7e4997884.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2110.367110] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2110.367629] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-87e89e0f-e3c6-4d09-967d-819665bfc9ef {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2110.378919] env[62824]: DEBUG oslo_vmware.api [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2110.378919] env[62824]: value = "task-2146337" [ 2110.378919] env[62824]: _type = "Task" [ 2110.378919] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2110.396309] env[62824]: DEBUG oslo_vmware.api [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146337, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2110.403484] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146336, 'name': CreateVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2110.477488] env[62824]: DEBUG oslo_concurrency.lockutils [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Acquiring lock "2afd1f18-234c-40b5-9ede-7413ad30dafe" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2110.477722] env[62824]: DEBUG oslo_concurrency.lockutils [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Lock "2afd1f18-234c-40b5-9ede-7413ad30dafe" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2110.872606] env[62824]: DEBUG nova.virt.hardware [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2110.872907] env[62824]: DEBUG nova.virt.hardware [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2110.872907] env[62824]: DEBUG nova.virt.hardware [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2110.873054] env[62824]: DEBUG nova.virt.hardware [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2110.873290] env[62824]: DEBUG nova.virt.hardware [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2110.873555] env[62824]: DEBUG nova.virt.hardware [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2110.873797] env[62824]: DEBUG nova.virt.hardware [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2110.873963] env[62824]: DEBUG nova.virt.hardware [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2110.874149] env[62824]: DEBUG nova.virt.hardware [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2110.874327] env[62824]: DEBUG nova.virt.hardware [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2110.874492] env[62824]: DEBUG nova.virt.hardware [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2110.875679] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60493d49-6167-42cd-934e-3eeae8baa2a7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2110.886358] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e04d0562-ef86-4748-afce-ed11d9df4268 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2110.896488] env[62824]: DEBUG oslo_vmware.api [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146337, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076453} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2110.899588] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2110.900356] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ee8b29b-af91-4827-9bab-84a57b7160c6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2110.912995] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:07:e9:a0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4c9a12d2-469f-4199-bfaa-f791d765deac', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '309b4346-1935-41f1-a374-d322c4a8c016', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2110.920881] env[62824]: DEBUG oslo.service.loopingcall [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2110.927323] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2110.927323] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146336, 'name': CreateVM_Task, 'duration_secs': 0.598457} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2110.928978] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8de79574-a65b-40bb-9a13-a466ab5fcbbe {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2110.942954] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2110.954588] env[62824]: DEBUG oslo_concurrency.lockutils [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2110.954588] env[62824]: DEBUG oslo_concurrency.lockutils [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2110.954769] env[62824]: DEBUG oslo_concurrency.lockutils [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2110.963943] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] Reconfiguring VM instance instance-00000042 to attach disk [datastore1] ca42fef0-1e90-4ab6-9d60-8ef7e4997884/ca42fef0-1e90-4ab6-9d60-8ef7e4997884.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2110.969503] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-85e58d07-4f2f-4af0-b99c-dcf9df6a47dd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2110.971516] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-59afc917-f0bf-4868-b6b5-f0a3f1aefd28 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2110.987982] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2110.987982] env[62824]: value = "task-2146338" [ 2110.987982] env[62824]: _type = "Task" [ 2110.987982] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2110.994800] env[62824]: DEBUG oslo_vmware.api [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2110.994800] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52372f32-c658-69cc-ed3b-5c9dbb8d2426" [ 2110.994800] env[62824]: _type = "Task" [ 2110.994800] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2110.996249] env[62824]: DEBUG oslo_vmware.api [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2110.996249] env[62824]: value = "task-2146339" [ 2110.996249] env[62824]: _type = "Task" [ 2110.996249] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2111.007242] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146338, 'name': CreateVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2111.010616] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f878c520-cb4e-47c4-bd06-93b02fad5ec4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2111.022423] env[62824]: DEBUG oslo_vmware.api [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52372f32-c658-69cc-ed3b-5c9dbb8d2426, 'name': SearchDatastore_Task, 'duration_secs': 0.010916} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2111.023027] env[62824]: DEBUG oslo_concurrency.lockutils [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2111.023565] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2111.023628] env[62824]: DEBUG oslo_concurrency.lockutils [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2111.023778] env[62824]: DEBUG oslo_concurrency.lockutils [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2111.023994] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2111.029265] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b3118240-9214-4d6d-b1d3-e5cfce393bfa {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2111.030871] env[62824]: DEBUG oslo_vmware.api [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146339, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2111.034215] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1bd1fb1-afe8-4ab4-8436-cb13da2e01a4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2111.039791] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2111.040111] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2111.066903] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-08f7f0a1-ce9b-4133-8e4e-984b0f397432 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2111.070540] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78b04559-d202-46fa-b329-cdffdf12aaed {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2111.078956] env[62824]: DEBUG oslo_vmware.api [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2111.078956] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]529d9a24-a521-065b-884e-4a4c6be63c2a" [ 2111.078956] env[62824]: _type = "Task" [ 2111.078956] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2111.080376] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-854ca1f7-7c0a-421c-be47-b993ddf5a52a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2111.094297] env[62824]: DEBUG oslo_vmware.api [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]529d9a24-a521-065b-884e-4a4c6be63c2a, 'name': SearchDatastore_Task, 'duration_secs': 0.010238} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2111.103678] env[62824]: DEBUG nova.compute.provider_tree [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2111.107434] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fe1b0cfc-b86f-4969-ad94-e137e8b0dd44 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2111.117425] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Acquiring lock "ab37b1a2-0012-40fb-9341-b613525e89cf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2111.117670] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Lock "ab37b1a2-0012-40fb-9341-b613525e89cf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2111.117966] env[62824]: DEBUG oslo_vmware.api [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2111.117966] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52d53cc4-b12d-c8cf-d91b-5f97dc96f54a" [ 2111.117966] env[62824]: _type = "Task" [ 2111.117966] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2111.128455] env[62824]: DEBUG oslo_vmware.api [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52d53cc4-b12d-c8cf-d91b-5f97dc96f54a, 'name': SearchDatastore_Task, 'duration_secs': 0.010003} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2111.128721] env[62824]: DEBUG oslo_concurrency.lockutils [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2111.128980] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 5b3df31f-15fe-473f-992c-ddb272661c53/5b3df31f-15fe-473f-992c-ddb272661c53.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2111.129261] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-dbd69cd0-23b0-48f9-965f-a0a2ee7f6e10 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2111.136446] env[62824]: DEBUG oslo_vmware.api [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2111.136446] env[62824]: value = "task-2146340" [ 2111.136446] env[62824]: _type = "Task" [ 2111.136446] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2111.147739] env[62824]: DEBUG oslo_vmware.api [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2146340, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2111.505117] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146338, 'name': CreateVM_Task, 'duration_secs': 0.392763} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2111.512923] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2111.514076] env[62824]: DEBUG oslo_concurrency.lockutils [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2111.514446] env[62824]: DEBUG oslo_concurrency.lockutils [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2111.514878] env[62824]: DEBUG oslo_concurrency.lockutils [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2111.515947] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ddf4a310-62e4-445f-b13f-343d0ef236c3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2111.524065] env[62824]: DEBUG oslo_vmware.api [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146339, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2111.530296] env[62824]: DEBUG oslo_vmware.api [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for the task: (returnval){ [ 2111.530296] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]520d85c1-4fa1-6b11-ea87-e4b2f87f0261" [ 2111.530296] env[62824]: _type = "Task" [ 2111.530296] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2111.539236] env[62824]: DEBUG oslo_vmware.api [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]520d85c1-4fa1-6b11-ea87-e4b2f87f0261, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2111.551589] env[62824]: DEBUG nova.network.neutron [req-ebd60a84-bbd0-4da7-b659-0e00a424bc87 req-19df813e-66c9-4a1e-80a0-a4fddf4fe9ad service nova] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Updated VIF entry in instance network info cache for port 7be95c5e-6e98-4fec-8550-1c22a787b95b. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2111.551950] env[62824]: DEBUG nova.network.neutron [req-ebd60a84-bbd0-4da7-b659-0e00a424bc87 req-19df813e-66c9-4a1e-80a0-a4fddf4fe9ad service nova] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Updating instance_info_cache with network_info: [{"id": "7be95c5e-6e98-4fec-8550-1c22a787b95b", "address": "fa:16:3e:e8:69:3a", "network": {"id": "bba9b98c-c7eb-4890-8de5-eafacfdebd7d", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1928394728-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7fcdbfc29df64a1ba8d982bdcc667b64", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b49c5024-2ced-42ca-90cc-6066766d43e6", "external-id": "nsx-vlan-transportzone-239", "segmentation_id": 239, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7be95c5e-6e", "ovs_interfaceid": "7be95c5e-6e98-4fec-8550-1c22a787b95b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2111.613366] env[62824]: DEBUG nova.scheduler.client.report [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2111.652753] env[62824]: DEBUG oslo_vmware.api [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2146340, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2111.850038] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9274965-3b8a-49a7-9276-3a0e6a892115 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2111.850038] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1e658c58-4bb1-4c52-a3eb-40831d0a5015 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2111.859026] env[62824]: DEBUG oslo_vmware.api [None req-e9274965-3b8a-49a7-9276-3a0e6a892115 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Waiting for the task: (returnval){ [ 2111.859026] env[62824]: value = "task-2146341" [ 2111.859026] env[62824]: _type = "Task" [ 2111.859026] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2111.871271] env[62824]: DEBUG oslo_vmware.api [None req-e9274965-3b8a-49a7-9276-3a0e6a892115 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146341, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2112.017317] env[62824]: DEBUG oslo_vmware.api [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146339, 'name': ReconfigVM_Task, 'duration_secs': 0.647308} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2112.017679] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] Reconfigured VM instance instance-00000042 to attach disk [datastore1] ca42fef0-1e90-4ab6-9d60-8ef7e4997884/ca42fef0-1e90-4ab6-9d60-8ef7e4997884.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2112.018496] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-19b1cd2a-c4e4-4dc6-aba7-090b4cbb4acf {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2112.027301] env[62824]: DEBUG oslo_vmware.api [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2112.027301] env[62824]: value = "task-2146342" [ 2112.027301] env[62824]: _type = "Task" [ 2112.027301] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2112.039160] env[62824]: DEBUG oslo_vmware.api [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]520d85c1-4fa1-6b11-ea87-e4b2f87f0261, 'name': SearchDatastore_Task, 'duration_secs': 0.050771} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2112.045902] env[62824]: DEBUG oslo_concurrency.lockutils [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2112.045902] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2112.045902] env[62824]: DEBUG oslo_concurrency.lockutils [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2112.045902] env[62824]: DEBUG oslo_concurrency.lockutils [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2112.046058] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2112.046302] env[62824]: DEBUG oslo_vmware.api [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146342, 'name': Rename_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2112.046599] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-62af1d3b-0ef4-4ae3-9f5f-a687dbaaca31 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2112.055303] env[62824]: DEBUG oslo_concurrency.lockutils [req-ebd60a84-bbd0-4da7-b659-0e00a424bc87 req-19df813e-66c9-4a1e-80a0-a4fddf4fe9ad service nova] Releasing lock "refresh_cache-5b3df31f-15fe-473f-992c-ddb272661c53" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2112.056593] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2112.056793] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2112.057793] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1d5e3cf2-0a23-4c54-9066-05e75a72ed84 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2112.064226] env[62824]: DEBUG oslo_vmware.api [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for the task: (returnval){ [ 2112.064226] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52eb704d-f2b6-e53e-3926-203507061438" [ 2112.064226] env[62824]: _type = "Task" [ 2112.064226] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2112.073945] env[62824]: DEBUG oslo_vmware.api [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52eb704d-f2b6-e53e-3926-203507061438, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2112.120205] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.764s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2112.120759] env[62824]: DEBUG nova.compute.manager [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 49818910-61de-4fb8-bbab-d5d61d2f1ada] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2112.123932] env[62824]: DEBUG oslo_concurrency.lockutils [None req-88bc485b-f4ba-418f-94b3-e4e74a1225cb tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.297s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2112.129305] env[62824]: INFO nova.compute.claims [None req-88bc485b-f4ba-418f-94b3-e4e74a1225cb tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: cc8b6a13-da40-4c68-afc3-8b2c6ba574d4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2112.154599] env[62824]: DEBUG oslo_vmware.api [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2146340, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.522733} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2112.154922] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 5b3df31f-15fe-473f-992c-ddb272661c53/5b3df31f-15fe-473f-992c-ddb272661c53.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2112.155536] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2112.155667] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-71458086-0d68-477c-b436-42efbb4934cc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2112.165872] env[62824]: DEBUG oslo_vmware.api [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2112.165872] env[62824]: value = "task-2146343" [ 2112.165872] env[62824]: _type = "Task" [ 2112.165872] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2112.178090] env[62824]: DEBUG oslo_vmware.api [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2146343, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2112.370120] env[62824]: DEBUG oslo_vmware.api [None req-e9274965-3b8a-49a7-9276-3a0e6a892115 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146341, 'name': PowerOffVM_Task, 'duration_secs': 0.237784} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2112.370515] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9274965-3b8a-49a7-9276-3a0e6a892115 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2112.371575] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8737277c-0cf0-4e92-a649-36b51524a68b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2112.390938] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c958a7ef-f64f-4ce5-b89c-7d21f3919f8f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2112.422934] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9274965-3b8a-49a7-9276-3a0e6a892115 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2112.423269] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7d8ef723-a953-4af6-b6a3-72ab8aa956ca {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2112.430638] env[62824]: DEBUG oslo_vmware.api [None req-e9274965-3b8a-49a7-9276-3a0e6a892115 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Waiting for the task: (returnval){ [ 2112.430638] env[62824]: value = "task-2146344" [ 2112.430638] env[62824]: _type = "Task" [ 2112.430638] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2112.439037] env[62824]: DEBUG oslo_vmware.api [None req-e9274965-3b8a-49a7-9276-3a0e6a892115 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146344, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2112.535600] env[62824]: DEBUG oslo_vmware.api [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146342, 'name': Rename_Task, 'duration_secs': 0.18287} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2112.535898] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2112.536262] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-47887b8b-21ff-4cc1-9eaa-c9645e70c72f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2112.543214] env[62824]: DEBUG oslo_vmware.api [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2112.543214] env[62824]: value = "task-2146345" [ 2112.543214] env[62824]: _type = "Task" [ 2112.543214] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2112.553376] env[62824]: DEBUG oslo_vmware.api [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146345, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2112.574639] env[62824]: DEBUG oslo_vmware.api [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52eb704d-f2b6-e53e-3926-203507061438, 'name': SearchDatastore_Task, 'duration_secs': 0.009596} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2112.575509] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-89a022fe-ee08-4508-877d-d9648a9ab74c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2112.581635] env[62824]: DEBUG oslo_vmware.api [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for the task: (returnval){ [ 2112.581635] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5280e5b4-fbde-c882-d9b9-bec38adea2b9" [ 2112.581635] env[62824]: _type = "Task" [ 2112.581635] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2112.590256] env[62824]: DEBUG oslo_vmware.api [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5280e5b4-fbde-c882-d9b9-bec38adea2b9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2112.631111] env[62824]: DEBUG nova.compute.utils [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2112.632856] env[62824]: DEBUG oslo_concurrency.lockutils [None req-88bc485b-f4ba-418f-94b3-e4e74a1225cb tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 0.509s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2112.633551] env[62824]: DEBUG nova.compute.utils [None req-88bc485b-f4ba-418f-94b3-e4e74a1225cb tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: cc8b6a13-da40-4c68-afc3-8b2c6ba574d4] Instance cc8b6a13-da40-4c68-afc3-8b2c6ba574d4 could not be found. {{(pid=62824) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 2112.634862] env[62824]: DEBUG nova.compute.manager [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 49818910-61de-4fb8-bbab-d5d61d2f1ada] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2112.635052] env[62824]: DEBUG nova.network.neutron [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 49818910-61de-4fb8-bbab-d5d61d2f1ada] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2112.636863] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b1f093ef-801c-4886-991f-869391717179 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.468s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2112.637149] env[62824]: DEBUG nova.objects.instance [None req-b1f093ef-801c-4886-991f-869391717179 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Lazy-loading 'resources' on Instance uuid 8bb90a4d-93c9-4f54-b15e-48fe966e6c97 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2112.638429] env[62824]: DEBUG nova.compute.manager [None req-88bc485b-f4ba-418f-94b3-e4e74a1225cb tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: cc8b6a13-da40-4c68-afc3-8b2c6ba574d4] Instance disappeared during build. {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2520}} [ 2112.638597] env[62824]: DEBUG nova.compute.manager [None req-88bc485b-f4ba-418f-94b3-e4e74a1225cb tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: cc8b6a13-da40-4c68-afc3-8b2c6ba574d4] Unplugging VIFs for instance {{(pid=62824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3013}} [ 2112.638820] env[62824]: DEBUG oslo_concurrency.lockutils [None req-88bc485b-f4ba-418f-94b3-e4e74a1225cb tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquiring lock "refresh_cache-cc8b6a13-da40-4c68-afc3-8b2c6ba574d4" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2112.638965] env[62824]: DEBUG oslo_concurrency.lockutils [None req-88bc485b-f4ba-418f-94b3-e4e74a1225cb tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquired lock "refresh_cache-cc8b6a13-da40-4c68-afc3-8b2c6ba574d4" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2112.639215] env[62824]: DEBUG nova.network.neutron [None req-88bc485b-f4ba-418f-94b3-e4e74a1225cb tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: cc8b6a13-da40-4c68-afc3-8b2c6ba574d4] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2112.677763] env[62824]: DEBUG oslo_vmware.api [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2146343, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071489} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2112.679051] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2112.680249] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-522531ba-523f-4d60-9896-18e3dc7f9622 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2112.711125] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Reconfiguring VM instance instance-00000043 to attach disk [datastore1] 5b3df31f-15fe-473f-992c-ddb272661c53/5b3df31f-15fe-473f-992c-ddb272661c53.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2112.711721] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d6409c31-c0da-4747-8b04-681faa262c6f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2112.738745] env[62824]: DEBUG oslo_vmware.api [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2112.738745] env[62824]: value = "task-2146346" [ 2112.738745] env[62824]: _type = "Task" [ 2112.738745] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2112.738745] env[62824]: DEBUG nova.policy [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '28495352107b45cd9cbd746c6affc4fe', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '64042a790d6c459186c68d73da32c019', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2112.750610] env[62824]: DEBUG oslo_vmware.api [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2146346, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2112.946173] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9274965-3b8a-49a7-9276-3a0e6a892115 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: 363177c0-dc40-429a-a74b-e690da133edb] VM already powered off {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 2112.946173] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e9274965-3b8a-49a7-9276-3a0e6a892115 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2112.946558] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e9274965-3b8a-49a7-9276-3a0e6a892115 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2112.946682] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e9274965-3b8a-49a7-9276-3a0e6a892115 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2112.946875] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-e9274965-3b8a-49a7-9276-3a0e6a892115 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2112.948017] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dfc93950-d8f7-4a27-9f3d-97a4819ec478 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2112.961098] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-e9274965-3b8a-49a7-9276-3a0e6a892115 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2112.961323] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e9274965-3b8a-49a7-9276-3a0e6a892115 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2112.962447] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0e54365a-c16d-43c8-908d-a19d26ba3c6a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2112.968370] env[62824]: DEBUG oslo_vmware.api [None req-e9274965-3b8a-49a7-9276-3a0e6a892115 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Waiting for the task: (returnval){ [ 2112.968370] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52ee0fd8-fb86-3ba1-a138-b2c7a8adb012" [ 2112.968370] env[62824]: _type = "Task" [ 2112.968370] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2112.977074] env[62824]: DEBUG oslo_vmware.api [None req-e9274965-3b8a-49a7-9276-3a0e6a892115 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52ee0fd8-fb86-3ba1-a138-b2c7a8adb012, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2113.054854] env[62824]: DEBUG oslo_vmware.api [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146345, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2113.096496] env[62824]: DEBUG oslo_vmware.api [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5280e5b4-fbde-c882-d9b9-bec38adea2b9, 'name': SearchDatastore_Task, 'duration_secs': 0.010068} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2113.096730] env[62824]: DEBUG oslo_concurrency.lockutils [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2113.096827] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 07ce931f-45ef-409b-b714-9f1cd47a3a88/07ce931f-45ef-409b-b714-9f1cd47a3a88.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2113.097071] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6aa6c8b5-e48e-477b-97f7-92cb386d2770 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2113.104769] env[62824]: DEBUG oslo_vmware.api [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for the task: (returnval){ [ 2113.104769] env[62824]: value = "task-2146347" [ 2113.104769] env[62824]: _type = "Task" [ 2113.104769] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2113.114960] env[62824]: DEBUG oslo_vmware.api [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146347, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2113.135854] env[62824]: DEBUG nova.compute.manager [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 49818910-61de-4fb8-bbab-d5d61d2f1ada] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2113.145760] env[62824]: DEBUG nova.compute.utils [None req-88bc485b-f4ba-418f-94b3-e4e74a1225cb tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: cc8b6a13-da40-4c68-afc3-8b2c6ba574d4] Can not refresh info_cache because instance was not found {{(pid=62824) refresh_info_cache_for_instance /opt/stack/nova/nova/compute/utils.py:1055}} [ 2113.185494] env[62824]: DEBUG nova.network.neutron [None req-88bc485b-f4ba-418f-94b3-e4e74a1225cb tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: cc8b6a13-da40-4c68-afc3-8b2c6ba574d4] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2113.256487] env[62824]: DEBUG nova.network.neutron [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 49818910-61de-4fb8-bbab-d5d61d2f1ada] Successfully created port: f26c1cce-02ce-4d96-b8b7-b206825081b3 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2113.260257] env[62824]: DEBUG oslo_vmware.api [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2146346, 'name': ReconfigVM_Task, 'duration_secs': 0.397872} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2113.263168] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Reconfigured VM instance instance-00000043 to attach disk [datastore1] 5b3df31f-15fe-473f-992c-ddb272661c53/5b3df31f-15fe-473f-992c-ddb272661c53.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2113.264659] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dd3799bf-099a-4be3-ac84-1edb0bfb0f6e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2113.272760] env[62824]: DEBUG oslo_vmware.api [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2113.272760] env[62824]: value = "task-2146348" [ 2113.272760] env[62824]: _type = "Task" [ 2113.272760] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2113.287953] env[62824]: DEBUG oslo_vmware.api [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2146348, 'name': Rename_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2113.308212] env[62824]: DEBUG nova.network.neutron [None req-88bc485b-f4ba-418f-94b3-e4e74a1225cb tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: cc8b6a13-da40-4c68-afc3-8b2c6ba574d4] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2113.485969] env[62824]: DEBUG oslo_vmware.api [None req-e9274965-3b8a-49a7-9276-3a0e6a892115 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52ee0fd8-fb86-3ba1-a138-b2c7a8adb012, 'name': SearchDatastore_Task, 'duration_secs': 0.009211} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2113.487674] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-631e46b0-f406-430f-bfb5-707564154981 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2113.494932] env[62824]: DEBUG oslo_vmware.api [None req-e9274965-3b8a-49a7-9276-3a0e6a892115 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Waiting for the task: (returnval){ [ 2113.494932] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]528fadfb-31fe-f876-f2da-70747bb81fac" [ 2113.494932] env[62824]: _type = "Task" [ 2113.494932] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2113.506675] env[62824]: DEBUG oslo_vmware.api [None req-e9274965-3b8a-49a7-9276-3a0e6a892115 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]528fadfb-31fe-f876-f2da-70747bb81fac, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2113.557046] env[62824]: DEBUG oslo_vmware.api [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146345, 'name': PowerOnVM_Task, 'duration_secs': 0.550015} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2113.557340] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2113.557584] env[62824]: INFO nova.compute.manager [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] Took 9.37 seconds to spawn the instance on the hypervisor. [ 2113.557774] env[62824]: DEBUG nova.compute.manager [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2113.558618] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67d59b26-26a0-4dda-acda-54cabedcb83e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2113.617127] env[62824]: DEBUG oslo_vmware.api [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146347, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.49312} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2113.617418] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 07ce931f-45ef-409b-b714-9f1cd47a3a88/07ce931f-45ef-409b-b714-9f1cd47a3a88.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2113.617636] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2113.617889] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-513e8f0f-b141-4fb9-927a-5561a7211059 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2113.626965] env[62824]: DEBUG oslo_vmware.api [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for the task: (returnval){ [ 2113.626965] env[62824]: value = "task-2146349" [ 2113.626965] env[62824]: _type = "Task" [ 2113.626965] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2113.635123] env[62824]: DEBUG oslo_vmware.api [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146349, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2113.779450] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a15d5bc0-c0d4-4806-b592-c68006606bf6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2113.784949] env[62824]: DEBUG oslo_vmware.api [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2146348, 'name': Rename_Task, 'duration_secs': 0.346416} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2113.785656] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2113.785922] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2361d287-f914-489d-add5-1bcff5aba12f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2113.792020] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89662e69-f85f-4745-993e-dbec23bea8f2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2113.795357] env[62824]: DEBUG oslo_vmware.api [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2113.795357] env[62824]: value = "task-2146350" [ 2113.795357] env[62824]: _type = "Task" [ 2113.795357] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2113.826387] env[62824]: DEBUG oslo_concurrency.lockutils [None req-88bc485b-f4ba-418f-94b3-e4e74a1225cb tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Releasing lock "refresh_cache-cc8b6a13-da40-4c68-afc3-8b2c6ba574d4" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2113.826663] env[62824]: DEBUG nova.compute.manager [None req-88bc485b-f4ba-418f-94b3-e4e74a1225cb tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3036}} [ 2113.826857] env[62824]: DEBUG nova.compute.manager [None req-88bc485b-f4ba-418f-94b3-e4e74a1225cb tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: cc8b6a13-da40-4c68-afc3-8b2c6ba574d4] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2113.827038] env[62824]: DEBUG nova.network.neutron [None req-88bc485b-f4ba-418f-94b3-e4e74a1225cb tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: cc8b6a13-da40-4c68-afc3-8b2c6ba574d4] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2113.830291] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c2db18b-412f-4a95-af1e-376fbe9cc7cf {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2113.836258] env[62824]: DEBUG oslo_vmware.api [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2146350, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2113.841337] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fee5b0f7-a9a3-47cc-811a-63f0621be6e4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2113.856182] env[62824]: DEBUG nova.compute.provider_tree [None req-b1f093ef-801c-4886-991f-869391717179 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2113.858248] env[62824]: DEBUG nova.network.neutron [None req-88bc485b-f4ba-418f-94b3-e4e74a1225cb tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: cc8b6a13-da40-4c68-afc3-8b2c6ba574d4] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2114.007762] env[62824]: DEBUG oslo_vmware.api [None req-e9274965-3b8a-49a7-9276-3a0e6a892115 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]528fadfb-31fe-f876-f2da-70747bb81fac, 'name': SearchDatastore_Task, 'duration_secs': 0.026872} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2114.008018] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e9274965-3b8a-49a7-9276-3a0e6a892115 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2114.008429] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-e9274965-3b8a-49a7-9276-3a0e6a892115 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 363177c0-dc40-429a-a74b-e690da133edb/9e2a7d30-212d-4ab8-9606-c5c6d52629e8-rescue.vmdk. {{(pid=62824) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 2114.008792] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-484fd7c1-993d-44da-a0b4-a96eb87faa53 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2114.015017] env[62824]: DEBUG oslo_vmware.api [None req-e9274965-3b8a-49a7-9276-3a0e6a892115 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Waiting for the task: (returnval){ [ 2114.015017] env[62824]: value = "task-2146351" [ 2114.015017] env[62824]: _type = "Task" [ 2114.015017] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2114.022497] env[62824]: DEBUG oslo_vmware.api [None req-e9274965-3b8a-49a7-9276-3a0e6a892115 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146351, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2114.084506] env[62824]: INFO nova.compute.manager [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] Took 31.62 seconds to build instance. [ 2114.136891] env[62824]: DEBUG oslo_vmware.api [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146349, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063287} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2114.137185] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2114.138087] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c880e201-b034-44f0-86b5-a99ea186402c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2114.162325] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Reconfiguring VM instance instance-00000018 to attach disk [datastore1] 07ce931f-45ef-409b-b714-9f1cd47a3a88/07ce931f-45ef-409b-b714-9f1cd47a3a88.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2114.163694] env[62824]: DEBUG nova.compute.manager [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 49818910-61de-4fb8-bbab-d5d61d2f1ada] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2114.165958] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3d65180b-51ea-4bda-b439-d5ef87ef3f63 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2114.187848] env[62824]: DEBUG oslo_vmware.api [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for the task: (returnval){ [ 2114.187848] env[62824]: value = "task-2146352" [ 2114.187848] env[62824]: _type = "Task" [ 2114.187848] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2114.197775] env[62824]: DEBUG oslo_vmware.api [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146352, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2114.308301] env[62824]: DEBUG oslo_vmware.api [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2146350, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2114.345031] env[62824]: DEBUG nova.virt.hardware [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2114.345296] env[62824]: DEBUG nova.virt.hardware [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2114.345486] env[62824]: DEBUG nova.virt.hardware [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2114.345662] env[62824]: DEBUG nova.virt.hardware [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2114.345808] env[62824]: DEBUG nova.virt.hardware [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2114.346743] env[62824]: DEBUG nova.virt.hardware [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2114.346743] env[62824]: DEBUG nova.virt.hardware [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2114.346743] env[62824]: DEBUG nova.virt.hardware [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2114.346743] env[62824]: DEBUG nova.virt.hardware [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2114.346743] env[62824]: DEBUG nova.virt.hardware [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2114.347068] env[62824]: DEBUG nova.virt.hardware [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2114.347905] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2a4825c-60b4-4318-ae22-f2fa5255cdfe {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2114.356504] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1296dd55-4ccb-4068-a0c7-c105d62d03e2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2114.363890] env[62824]: DEBUG nova.scheduler.client.report [None req-b1f093ef-801c-4886-991f-869391717179 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2114.368862] env[62824]: DEBUG nova.network.neutron [None req-88bc485b-f4ba-418f-94b3-e4e74a1225cb tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: cc8b6a13-da40-4c68-afc3-8b2c6ba574d4] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2114.528636] env[62824]: DEBUG oslo_vmware.api [None req-e9274965-3b8a-49a7-9276-3a0e6a892115 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146351, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2114.586862] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a9a27205-809a-4a69-b569-2d75435b5bd6 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "ca42fef0-1e90-4ab6-9d60-8ef7e4997884" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 33.135s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2114.701065] env[62824]: DEBUG oslo_vmware.api [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146352, 'name': ReconfigVM_Task, 'duration_secs': 0.353373} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2114.701609] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Reconfigured VM instance instance-00000018 to attach disk [datastore1] 07ce931f-45ef-409b-b714-9f1cd47a3a88/07ce931f-45ef-409b-b714-9f1cd47a3a88.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2114.702629] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9b9bc710-d733-4f20-8184-89ec7bdb6bf4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2114.711439] env[62824]: DEBUG oslo_vmware.api [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for the task: (returnval){ [ 2114.711439] env[62824]: value = "task-2146353" [ 2114.711439] env[62824]: _type = "Task" [ 2114.711439] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2114.720064] env[62824]: DEBUG oslo_vmware.api [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146353, 'name': Rename_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2114.807577] env[62824]: DEBUG oslo_vmware.api [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2146350, 'name': PowerOnVM_Task, 'duration_secs': 0.77936} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2114.807969] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2114.808272] env[62824]: INFO nova.compute.manager [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Took 7.89 seconds to spawn the instance on the hypervisor. [ 2114.808473] env[62824]: DEBUG nova.compute.manager [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2114.809336] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-613cbe7c-9776-4b03-9f02-9b0d03e30f07 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2114.872296] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b1f093ef-801c-4886-991f-869391717179 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.234s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2114.875464] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ffe016d6-cd3f-454a-b9ac-22101a9a9bfd tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.178s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2114.875889] env[62824]: DEBUG nova.objects.instance [None req-ffe016d6-cd3f-454a-b9ac-22101a9a9bfd tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Lazy-loading 'resources' on Instance uuid 6a724332-a165-4a2b-9dd8-85f27e7b7637 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2114.877457] env[62824]: INFO nova.compute.manager [None req-88bc485b-f4ba-418f-94b3-e4e74a1225cb tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: cc8b6a13-da40-4c68-afc3-8b2c6ba574d4] Took 1.05 seconds to deallocate network for instance. [ 2114.904766] env[62824]: INFO nova.scheduler.client.report [None req-b1f093ef-801c-4886-991f-869391717179 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Deleted allocations for instance 8bb90a4d-93c9-4f54-b15e-48fe966e6c97 [ 2115.029560] env[62824]: DEBUG oslo_vmware.api [None req-e9274965-3b8a-49a7-9276-3a0e6a892115 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146351, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.518436} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2115.029897] env[62824]: INFO nova.virt.vmwareapi.ds_util [None req-e9274965-3b8a-49a7-9276-3a0e6a892115 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 363177c0-dc40-429a-a74b-e690da133edb/9e2a7d30-212d-4ab8-9606-c5c6d52629e8-rescue.vmdk. [ 2115.030718] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b88d3efc-6894-4347-bf5e-1cbe3159c8a3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2115.059431] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-e9274965-3b8a-49a7-9276-3a0e6a892115 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Reconfiguring VM instance instance-00000041 to attach disk [datastore2] 363177c0-dc40-429a-a74b-e690da133edb/9e2a7d30-212d-4ab8-9606-c5c6d52629e8-rescue.vmdk or device None with type thin {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2115.059913] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-635bccb5-be5a-4395-a4bf-0b108d86d316 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2115.078543] env[62824]: DEBUG oslo_vmware.api [None req-e9274965-3b8a-49a7-9276-3a0e6a892115 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Waiting for the task: (returnval){ [ 2115.078543] env[62824]: value = "task-2146354" [ 2115.078543] env[62824]: _type = "Task" [ 2115.078543] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2115.087535] env[62824]: DEBUG oslo_vmware.api [None req-e9274965-3b8a-49a7-9276-3a0e6a892115 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146354, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2115.089185] env[62824]: DEBUG nova.compute.manager [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 20f93b46-5e7e-4256-8cc1-e0a1b16740d5] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2115.232606] env[62824]: DEBUG oslo_vmware.api [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146353, 'name': Rename_Task, 'duration_secs': 0.149102} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2115.232965] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2115.233399] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b668ca67-5241-4509-8d3f-2f310b222403 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2115.240736] env[62824]: DEBUG oslo_vmware.api [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for the task: (returnval){ [ 2115.240736] env[62824]: value = "task-2146355" [ 2115.240736] env[62824]: _type = "Task" [ 2115.240736] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2115.254191] env[62824]: DEBUG oslo_vmware.api [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146355, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2115.283059] env[62824]: DEBUG nova.compute.manager [req-c6d3d973-9b9a-4682-8825-8a472589b611 req-df5b3b2e-48da-430f-9b05-7146272d6171 service nova] [instance: 49818910-61de-4fb8-bbab-d5d61d2f1ada] Received event network-vif-plugged-f26c1cce-02ce-4d96-b8b7-b206825081b3 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2115.283290] env[62824]: DEBUG oslo_concurrency.lockutils [req-c6d3d973-9b9a-4682-8825-8a472589b611 req-df5b3b2e-48da-430f-9b05-7146272d6171 service nova] Acquiring lock "49818910-61de-4fb8-bbab-d5d61d2f1ada-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2115.283507] env[62824]: DEBUG oslo_concurrency.lockutils [req-c6d3d973-9b9a-4682-8825-8a472589b611 req-df5b3b2e-48da-430f-9b05-7146272d6171 service nova] Lock "49818910-61de-4fb8-bbab-d5d61d2f1ada-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2115.283677] env[62824]: DEBUG oslo_concurrency.lockutils [req-c6d3d973-9b9a-4682-8825-8a472589b611 req-df5b3b2e-48da-430f-9b05-7146272d6171 service nova] Lock "49818910-61de-4fb8-bbab-d5d61d2f1ada-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2115.284129] env[62824]: DEBUG nova.compute.manager [req-c6d3d973-9b9a-4682-8825-8a472589b611 req-df5b3b2e-48da-430f-9b05-7146272d6171 service nova] [instance: 49818910-61de-4fb8-bbab-d5d61d2f1ada] No waiting events found dispatching network-vif-plugged-f26c1cce-02ce-4d96-b8b7-b206825081b3 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2115.284420] env[62824]: WARNING nova.compute.manager [req-c6d3d973-9b9a-4682-8825-8a472589b611 req-df5b3b2e-48da-430f-9b05-7146272d6171 service nova] [instance: 49818910-61de-4fb8-bbab-d5d61d2f1ada] Received unexpected event network-vif-plugged-f26c1cce-02ce-4d96-b8b7-b206825081b3 for instance with vm_state building and task_state spawning. [ 2115.334823] env[62824]: INFO nova.compute.manager [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Took 31.93 seconds to build instance. [ 2115.347152] env[62824]: DEBUG nova.network.neutron [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 49818910-61de-4fb8-bbab-d5d61d2f1ada] Successfully updated port: f26c1cce-02ce-4d96-b8b7-b206825081b3 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2115.413120] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b1f093ef-801c-4886-991f-869391717179 tempest-ListServerFiltersTestJSON-1467943214 tempest-ListServerFiltersTestJSON-1467943214-project-member] Lock "8bb90a4d-93c9-4f54-b15e-48fe966e6c97" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.575s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2115.590604] env[62824]: DEBUG oslo_vmware.api [None req-e9274965-3b8a-49a7-9276-3a0e6a892115 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146354, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2115.620024] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2115.664984] env[62824]: DEBUG nova.compute.manager [req-b1afb3c4-8cbc-408f-9587-e5f7271f0c04 req-585b1701-e33b-4e94-ae8f-bb483da40bdd service nova] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] Received event network-changed-3eb965bf-7ceb-4a51-bc6f-00803bc6b99e {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2115.664984] env[62824]: DEBUG nova.compute.manager [req-b1afb3c4-8cbc-408f-9587-e5f7271f0c04 req-585b1701-e33b-4e94-ae8f-bb483da40bdd service nova] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] Refreshing instance network info cache due to event network-changed-3eb965bf-7ceb-4a51-bc6f-00803bc6b99e. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2115.664984] env[62824]: DEBUG oslo_concurrency.lockutils [req-b1afb3c4-8cbc-408f-9587-e5f7271f0c04 req-585b1701-e33b-4e94-ae8f-bb483da40bdd service nova] Acquiring lock "refresh_cache-ca42fef0-1e90-4ab6-9d60-8ef7e4997884" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2115.664984] env[62824]: DEBUG oslo_concurrency.lockutils [req-b1afb3c4-8cbc-408f-9587-e5f7271f0c04 req-585b1701-e33b-4e94-ae8f-bb483da40bdd service nova] Acquired lock "refresh_cache-ca42fef0-1e90-4ab6-9d60-8ef7e4997884" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2115.664984] env[62824]: DEBUG nova.network.neutron [req-b1afb3c4-8cbc-408f-9587-e5f7271f0c04 req-585b1701-e33b-4e94-ae8f-bb483da40bdd service nova] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] Refreshing network info cache for port 3eb965bf-7ceb-4a51-bc6f-00803bc6b99e {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2115.757779] env[62824]: DEBUG oslo_vmware.api [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146355, 'name': PowerOnVM_Task, 'duration_secs': 0.496113} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2115.758541] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2115.759020] env[62824]: DEBUG nova.compute.manager [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2115.760123] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00d54669-df73-4446-980f-ded613e4a90b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2115.836796] env[62824]: DEBUG oslo_concurrency.lockutils [None req-876f467f-d557-4fec-a2d4-bcf5a6833ad3 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lock "5b3df31f-15fe-473f-992c-ddb272661c53" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 33.459s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2115.850718] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Acquiring lock "refresh_cache-49818910-61de-4fb8-bbab-d5d61d2f1ada" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2115.850718] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Acquired lock "refresh_cache-49818910-61de-4fb8-bbab-d5d61d2f1ada" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2115.850718] env[62824]: DEBUG nova.network.neutron [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 49818910-61de-4fb8-bbab-d5d61d2f1ada] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2115.900641] env[62824]: DEBUG oslo_concurrency.lockutils [None req-88bc485b-f4ba-418f-94b3-e4e74a1225cb tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "cc8b6a13-da40-4c68-afc3-8b2c6ba574d4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.116s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2116.007866] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5064956f-ab69-42f9-8e45-8c7a70b0e2ad {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2116.016313] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e3084f0-fc81-414b-a776-ac38f0a6b72c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2116.058444] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1503529-72db-4696-a93f-9008fe8d3001 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2116.070610] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d851df7-c5c4-4930-8a69-29f45dd20ec3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2116.088454] env[62824]: DEBUG nova.compute.provider_tree [None req-ffe016d6-cd3f-454a-b9ac-22101a9a9bfd tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2116.099840] env[62824]: DEBUG oslo_vmware.api [None req-e9274965-3b8a-49a7-9276-3a0e6a892115 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146354, 'name': ReconfigVM_Task, 'duration_secs': 0.690375} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2116.100272] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-e9274965-3b8a-49a7-9276-3a0e6a892115 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Reconfigured VM instance instance-00000041 to attach disk [datastore2] 363177c0-dc40-429a-a74b-e690da133edb/9e2a7d30-212d-4ab8-9606-c5c6d52629e8-rescue.vmdk or device None with type thin {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2116.101209] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-920fe797-4ad5-4cf3-84ef-54be637465a8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2116.131565] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-64e98b30-323e-48c3-b4b2-65cd9a78c5ba {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2116.149577] env[62824]: DEBUG oslo_vmware.api [None req-e9274965-3b8a-49a7-9276-3a0e6a892115 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Waiting for the task: (returnval){ [ 2116.149577] env[62824]: value = "task-2146356" [ 2116.149577] env[62824]: _type = "Task" [ 2116.149577] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2116.158741] env[62824]: DEBUG oslo_vmware.api [None req-e9274965-3b8a-49a7-9276-3a0e6a892115 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146356, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2116.283904] env[62824]: DEBUG oslo_concurrency.lockutils [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2116.340196] env[62824]: DEBUG nova.compute.manager [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2116.386713] env[62824]: DEBUG nova.network.neutron [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 49818910-61de-4fb8-bbab-d5d61d2f1ada] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2116.403363] env[62824]: DEBUG nova.compute.manager [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] [instance: bd7f2ed5-fa08-4a48-9755-60b61791ea0b] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2116.550245] env[62824]: DEBUG nova.network.neutron [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 49818910-61de-4fb8-bbab-d5d61d2f1ada] Updating instance_info_cache with network_info: [{"id": "f26c1cce-02ce-4d96-b8b7-b206825081b3", "address": "fa:16:3e:85:5b:0b", "network": {"id": "9e1885b6-0796-42b9-9cd8-8fa9ce9a3cfd", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-2118936609-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "64042a790d6c459186c68d73da32c019", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "afae6acd-1873-4228-9d5a-1cd5d4efe3e4", "external-id": "nsx-vlan-transportzone-183", "segmentation_id": 183, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf26c1cce-02", "ovs_interfaceid": "f26c1cce-02ce-4d96-b8b7-b206825081b3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2116.598067] env[62824]: DEBUG nova.scheduler.client.report [None req-ffe016d6-cd3f-454a-b9ac-22101a9a9bfd tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2116.664345] env[62824]: DEBUG oslo_vmware.api [None req-e9274965-3b8a-49a7-9276-3a0e6a892115 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146356, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2116.780671] env[62824]: DEBUG nova.network.neutron [req-b1afb3c4-8cbc-408f-9587-e5f7271f0c04 req-585b1701-e33b-4e94-ae8f-bb483da40bdd service nova] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] Updated VIF entry in instance network info cache for port 3eb965bf-7ceb-4a51-bc6f-00803bc6b99e. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2116.781112] env[62824]: DEBUG nova.network.neutron [req-b1afb3c4-8cbc-408f-9587-e5f7271f0c04 req-585b1701-e33b-4e94-ae8f-bb483da40bdd service nova] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] Updating instance_info_cache with network_info: [{"id": "3eb965bf-7ceb-4a51-bc6f-00803bc6b99e", "address": "fa:16:3e:3f:e9:09", "network": {"id": "b4ed285c-26a7-461b-8a71-d18e07582ab8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-997150974-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.185", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "815ecf9340934733a83d4dcb26612fd4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69f65356-c85e-4b7f-ad28-7c7b5e8cf50c", "external-id": "nsx-vlan-transportzone-281", "segmentation_id": 281, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3eb965bf-7c", "ovs_interfaceid": "3eb965bf-7ceb-4a51-bc6f-00803bc6b99e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2116.876955] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2116.926957] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2117.052855] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Releasing lock "refresh_cache-49818910-61de-4fb8-bbab-d5d61d2f1ada" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2117.052855] env[62824]: DEBUG nova.compute.manager [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 49818910-61de-4fb8-bbab-d5d61d2f1ada] Instance network_info: |[{"id": "f26c1cce-02ce-4d96-b8b7-b206825081b3", "address": "fa:16:3e:85:5b:0b", "network": {"id": "9e1885b6-0796-42b9-9cd8-8fa9ce9a3cfd", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-2118936609-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "64042a790d6c459186c68d73da32c019", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "afae6acd-1873-4228-9d5a-1cd5d4efe3e4", "external-id": "nsx-vlan-transportzone-183", "segmentation_id": 183, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf26c1cce-02", "ovs_interfaceid": "f26c1cce-02ce-4d96-b8b7-b206825081b3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2117.053371] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 49818910-61de-4fb8-bbab-d5d61d2f1ada] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:85:5b:0b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'afae6acd-1873-4228-9d5a-1cd5d4efe3e4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f26c1cce-02ce-4d96-b8b7-b206825081b3', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2117.064363] env[62824]: DEBUG oslo.service.loopingcall [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2117.064916] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 49818910-61de-4fb8-bbab-d5d61d2f1ada] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2117.064982] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7de5e5c1-af49-4693-b2a1-4f77d1f1ce25 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2117.087799] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2117.087799] env[62824]: value = "task-2146357" [ 2117.087799] env[62824]: _type = "Task" [ 2117.087799] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2117.097240] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146357, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2117.102297] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ffe016d6-cd3f-454a-b9ac-22101a9a9bfd tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.226s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2117.104328] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.402s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2117.105659] env[62824]: INFO nova.compute.claims [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 29140542-dc4b-411a-ac10-7d84086eabbb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2117.129204] env[62824]: INFO nova.scheduler.client.report [None req-ffe016d6-cd3f-454a-b9ac-22101a9a9bfd tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Deleted allocations for instance 6a724332-a165-4a2b-9dd8-85f27e7b7637 [ 2117.138392] env[62824]: DEBUG oslo_vmware.rw_handles [None req-4a76d89b-cbff-4ca0-a9aa-3ebc58bfda76 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52f9e045-1c7a-e2c9-c853-b2b6a239bfe8/disk-0.vmdk. {{(pid=62824) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 2117.138392] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cfe9c2d-f168-461f-b568-0b1a892d2cb4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2117.144738] env[62824]: DEBUG oslo_vmware.rw_handles [None req-4a76d89b-cbff-4ca0-a9aa-3ebc58bfda76 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52f9e045-1c7a-e2c9-c853-b2b6a239bfe8/disk-0.vmdk is in state: ready. {{(pid=62824) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 2117.144901] env[62824]: ERROR oslo_vmware.rw_handles [None req-4a76d89b-cbff-4ca0-a9aa-3ebc58bfda76 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52f9e045-1c7a-e2c9-c853-b2b6a239bfe8/disk-0.vmdk due to incomplete transfer. [ 2117.145174] env[62824]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-d9221d02-219a-4ecb-8a3c-93eb3c771580 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2117.157892] env[62824]: DEBUG oslo_vmware.rw_handles [None req-4a76d89b-cbff-4ca0-a9aa-3ebc58bfda76 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52f9e045-1c7a-e2c9-c853-b2b6a239bfe8/disk-0.vmdk. {{(pid=62824) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 2117.158147] env[62824]: DEBUG nova.virt.vmwareapi.images [None req-4a76d89b-cbff-4ca0-a9aa-3ebc58bfda76 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8] Uploaded image 758d9d98-91b8-46a1-9c95-5705c074be1a to the Glance image server {{(pid=62824) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 2117.161430] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a76d89b-cbff-4ca0-a9aa-3ebc58bfda76 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8] Destroying the VM {{(pid=62824) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 2117.161430] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-c5683920-4c42-45da-813f-81b3a682fffe {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2117.167499] env[62824]: DEBUG oslo_vmware.api [None req-e9274965-3b8a-49a7-9276-3a0e6a892115 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146356, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2117.172308] env[62824]: DEBUG oslo_vmware.api [None req-4a76d89b-cbff-4ca0-a9aa-3ebc58bfda76 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the task: (returnval){ [ 2117.172308] env[62824]: value = "task-2146358" [ 2117.172308] env[62824]: _type = "Task" [ 2117.172308] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2117.180522] env[62824]: DEBUG oslo_vmware.api [None req-4a76d89b-cbff-4ca0-a9aa-3ebc58bfda76 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146358, 'name': Destroy_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2117.285046] env[62824]: DEBUG oslo_concurrency.lockutils [req-b1afb3c4-8cbc-408f-9587-e5f7271f0c04 req-585b1701-e33b-4e94-ae8f-bb483da40bdd service nova] Releasing lock "refresh_cache-ca42fef0-1e90-4ab6-9d60-8ef7e4997884" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2117.320167] env[62824]: DEBUG nova.compute.manager [req-2b378cba-9866-4c6a-87a9-679968c8fc12 req-3ec4a3bd-b8cb-486d-b1d0-345a1fe58037 service nova] [instance: 49818910-61de-4fb8-bbab-d5d61d2f1ada] Received event network-changed-f26c1cce-02ce-4d96-b8b7-b206825081b3 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2117.320368] env[62824]: DEBUG nova.compute.manager [req-2b378cba-9866-4c6a-87a9-679968c8fc12 req-3ec4a3bd-b8cb-486d-b1d0-345a1fe58037 service nova] [instance: 49818910-61de-4fb8-bbab-d5d61d2f1ada] Refreshing instance network info cache due to event network-changed-f26c1cce-02ce-4d96-b8b7-b206825081b3. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2117.320588] env[62824]: DEBUG oslo_concurrency.lockutils [req-2b378cba-9866-4c6a-87a9-679968c8fc12 req-3ec4a3bd-b8cb-486d-b1d0-345a1fe58037 service nova] Acquiring lock "refresh_cache-49818910-61de-4fb8-bbab-d5d61d2f1ada" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2117.320731] env[62824]: DEBUG oslo_concurrency.lockutils [req-2b378cba-9866-4c6a-87a9-679968c8fc12 req-3ec4a3bd-b8cb-486d-b1d0-345a1fe58037 service nova] Acquired lock "refresh_cache-49818910-61de-4fb8-bbab-d5d61d2f1ada" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2117.320895] env[62824]: DEBUG nova.network.neutron [req-2b378cba-9866-4c6a-87a9-679968c8fc12 req-3ec4a3bd-b8cb-486d-b1d0-345a1fe58037 service nova] [instance: 49818910-61de-4fb8-bbab-d5d61d2f1ada] Refreshing network info cache for port f26c1cce-02ce-4d96-b8b7-b206825081b3 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2117.601517] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146357, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2117.638026] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ffe016d6-cd3f-454a-b9ac-22101a9a9bfd tempest-ServerShowV247Test-784026734 tempest-ServerShowV247Test-784026734-project-member] Lock "6a724332-a165-4a2b-9dd8-85f27e7b7637" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.824s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2117.663493] env[62824]: DEBUG oslo_vmware.api [None req-e9274965-3b8a-49a7-9276-3a0e6a892115 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146356, 'name': ReconfigVM_Task, 'duration_secs': 1.344866} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2117.663780] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9274965-3b8a-49a7-9276-3a0e6a892115 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2117.664051] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f05a63b3-4cdb-4389-a1ff-092653fef610 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2117.670951] env[62824]: DEBUG oslo_vmware.api [None req-e9274965-3b8a-49a7-9276-3a0e6a892115 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Waiting for the task: (returnval){ [ 2117.670951] env[62824]: value = "task-2146359" [ 2117.670951] env[62824]: _type = "Task" [ 2117.670951] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2117.683523] env[62824]: DEBUG oslo_vmware.api [None req-e9274965-3b8a-49a7-9276-3a0e6a892115 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146359, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2117.686607] env[62824]: DEBUG oslo_vmware.api [None req-4a76d89b-cbff-4ca0-a9aa-3ebc58bfda76 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146358, 'name': Destroy_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2118.044035] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7beba819-8d40-4b17-972a-05048027ddd5 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Acquiring lock "02f34565-6675-4c79-ac47-b131ceba9df8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2118.044035] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7beba819-8d40-4b17-972a-05048027ddd5 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Lock "02f34565-6675-4c79-ac47-b131ceba9df8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2118.044035] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7beba819-8d40-4b17-972a-05048027ddd5 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Acquiring lock "02f34565-6675-4c79-ac47-b131ceba9df8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2118.044185] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7beba819-8d40-4b17-972a-05048027ddd5 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Lock "02f34565-6675-4c79-ac47-b131ceba9df8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2118.044318] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7beba819-8d40-4b17-972a-05048027ddd5 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Lock "02f34565-6675-4c79-ac47-b131ceba9df8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2118.054327] env[62824]: INFO nova.compute.manager [None req-7beba819-8d40-4b17-972a-05048027ddd5 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02f34565-6675-4c79-ac47-b131ceba9df8] Terminating instance [ 2118.102380] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146357, 'name': CreateVM_Task, 'duration_secs': 0.938034} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2118.103315] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 49818910-61de-4fb8-bbab-d5d61d2f1ada] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2118.104198] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2118.104551] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2118.105028] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2118.105506] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-763a8b2e-f35f-4466-96f0-67041b1b3dca {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2118.110548] env[62824]: DEBUG oslo_vmware.api [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Waiting for the task: (returnval){ [ 2118.110548] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]527f5a8a-f776-c465-14bd-6612fb7e6f7a" [ 2118.110548] env[62824]: _type = "Task" [ 2118.110548] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2118.122318] env[62824]: DEBUG oslo_vmware.api [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]527f5a8a-f776-c465-14bd-6612fb7e6f7a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2118.185909] env[62824]: DEBUG oslo_vmware.api [None req-e9274965-3b8a-49a7-9276-3a0e6a892115 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146359, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2118.191038] env[62824]: DEBUG oslo_vmware.api [None req-4a76d89b-cbff-4ca0-a9aa-3ebc58bfda76 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146358, 'name': Destroy_Task, 'duration_secs': 0.909142} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2118.191038] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-4a76d89b-cbff-4ca0-a9aa-3ebc58bfda76 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8] Destroyed the VM [ 2118.191038] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-4a76d89b-cbff-4ca0-a9aa-3ebc58bfda76 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8] Deleting Snapshot of the VM instance {{(pid=62824) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 2118.191038] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-e7798e28-6bdb-4549-b25e-3e42ad020082 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2118.199355] env[62824]: DEBUG oslo_vmware.api [None req-4a76d89b-cbff-4ca0-a9aa-3ebc58bfda76 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the task: (returnval){ [ 2118.199355] env[62824]: value = "task-2146360" [ 2118.199355] env[62824]: _type = "Task" [ 2118.199355] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2118.209321] env[62824]: DEBUG oslo_vmware.api [None req-4a76d89b-cbff-4ca0-a9aa-3ebc58bfda76 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146360, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2118.335872] env[62824]: DEBUG nova.network.neutron [req-2b378cba-9866-4c6a-87a9-679968c8fc12 req-3ec4a3bd-b8cb-486d-b1d0-345a1fe58037 service nova] [instance: 49818910-61de-4fb8-bbab-d5d61d2f1ada] Updated VIF entry in instance network info cache for port f26c1cce-02ce-4d96-b8b7-b206825081b3. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2118.336321] env[62824]: DEBUG nova.network.neutron [req-2b378cba-9866-4c6a-87a9-679968c8fc12 req-3ec4a3bd-b8cb-486d-b1d0-345a1fe58037 service nova] [instance: 49818910-61de-4fb8-bbab-d5d61d2f1ada] Updating instance_info_cache with network_info: [{"id": "f26c1cce-02ce-4d96-b8b7-b206825081b3", "address": "fa:16:3e:85:5b:0b", "network": {"id": "9e1885b6-0796-42b9-9cd8-8fa9ce9a3cfd", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-2118936609-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "64042a790d6c459186c68d73da32c019", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "afae6acd-1873-4228-9d5a-1cd5d4efe3e4", "external-id": "nsx-vlan-transportzone-183", "segmentation_id": 183, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf26c1cce-02", "ovs_interfaceid": "f26c1cce-02ce-4d96-b8b7-b206825081b3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2118.562424] env[62824]: DEBUG nova.compute.manager [None req-7beba819-8d40-4b17-972a-05048027ddd5 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02f34565-6675-4c79-ac47-b131ceba9df8] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2118.562424] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-7beba819-8d40-4b17-972a-05048027ddd5 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02f34565-6675-4c79-ac47-b131ceba9df8] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2118.562424] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fab04a5-f6e6-4159-8a91-32030244aa12 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2118.570671] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-7beba819-8d40-4b17-972a-05048027ddd5 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02f34565-6675-4c79-ac47-b131ceba9df8] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2118.570955] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ceee8fa6-919c-4828-9535-05856fd444ad {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2118.579433] env[62824]: DEBUG oslo_vmware.api [None req-7beba819-8d40-4b17-972a-05048027ddd5 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for the task: (returnval){ [ 2118.579433] env[62824]: value = "task-2146361" [ 2118.579433] env[62824]: _type = "Task" [ 2118.579433] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2118.593992] env[62824]: DEBUG oslo_vmware.api [None req-7beba819-8d40-4b17-972a-05048027ddd5 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146361, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2118.632324] env[62824]: DEBUG oslo_vmware.api [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]527f5a8a-f776-c465-14bd-6612fb7e6f7a, 'name': SearchDatastore_Task, 'duration_secs': 0.012508} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2118.635266] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2118.635734] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 49818910-61de-4fb8-bbab-d5d61d2f1ada] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2118.635874] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2118.635998] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2118.636197] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2118.636941] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-26768d40-34f8-4eee-8cd3-7429055a3ba0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2118.647371] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2118.649016] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2118.649016] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e88cf0ef-0544-4804-957d-b6794c7ab2a2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2118.658472] env[62824]: DEBUG oslo_vmware.api [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Waiting for the task: (returnval){ [ 2118.658472] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5283a37b-d459-0ecc-f062-94f5339f0e4f" [ 2118.658472] env[62824]: _type = "Task" [ 2118.658472] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2118.671060] env[62824]: DEBUG oslo_vmware.api [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5283a37b-d459-0ecc-f062-94f5339f0e4f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2118.682669] env[62824]: DEBUG oslo_vmware.api [None req-e9274965-3b8a-49a7-9276-3a0e6a892115 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146359, 'name': PowerOnVM_Task, 'duration_secs': 0.53026} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2118.682966] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9274965-3b8a-49a7-9276-3a0e6a892115 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2118.685972] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-297850f8-4628-4709-9401-c678172d2be2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2118.689392] env[62824]: DEBUG nova.compute.manager [None req-e9274965-3b8a-49a7-9276-3a0e6a892115 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2118.690157] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e548980e-f9d5-4082-86ac-cfb1de6679f6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2118.701055] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f74d53cc-8d00-41bd-b81b-0297678a480b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2118.713600] env[62824]: DEBUG oslo_vmware.api [None req-4a76d89b-cbff-4ca0-a9aa-3ebc58bfda76 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146360, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2118.741267] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69b22f0b-73c3-4c55-929d-e7efb3b5cc1d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2118.749478] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-733c3802-187a-4947-8e26-dca1c665278d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2118.764245] env[62824]: DEBUG nova.compute.provider_tree [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2118.839702] env[62824]: DEBUG oslo_concurrency.lockutils [req-2b378cba-9866-4c6a-87a9-679968c8fc12 req-3ec4a3bd-b8cb-486d-b1d0-345a1fe58037 service nova] Releasing lock "refresh_cache-49818910-61de-4fb8-bbab-d5d61d2f1ada" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2118.839972] env[62824]: DEBUG nova.compute.manager [req-2b378cba-9866-4c6a-87a9-679968c8fc12 req-3ec4a3bd-b8cb-486d-b1d0-345a1fe58037 service nova] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Received event network-changed-7be95c5e-6e98-4fec-8550-1c22a787b95b {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2118.840155] env[62824]: DEBUG nova.compute.manager [req-2b378cba-9866-4c6a-87a9-679968c8fc12 req-3ec4a3bd-b8cb-486d-b1d0-345a1fe58037 service nova] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Refreshing instance network info cache due to event network-changed-7be95c5e-6e98-4fec-8550-1c22a787b95b. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2118.840401] env[62824]: DEBUG oslo_concurrency.lockutils [req-2b378cba-9866-4c6a-87a9-679968c8fc12 req-3ec4a3bd-b8cb-486d-b1d0-345a1fe58037 service nova] Acquiring lock "refresh_cache-5b3df31f-15fe-473f-992c-ddb272661c53" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2118.840504] env[62824]: DEBUG oslo_concurrency.lockutils [req-2b378cba-9866-4c6a-87a9-679968c8fc12 req-3ec4a3bd-b8cb-486d-b1d0-345a1fe58037 service nova] Acquired lock "refresh_cache-5b3df31f-15fe-473f-992c-ddb272661c53" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2118.840662] env[62824]: DEBUG nova.network.neutron [req-2b378cba-9866-4c6a-87a9-679968c8fc12 req-3ec4a3bd-b8cb-486d-b1d0-345a1fe58037 service nova] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Refreshing network info cache for port 7be95c5e-6e98-4fec-8550-1c22a787b95b {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2119.091028] env[62824]: DEBUG oslo_vmware.api [None req-7beba819-8d40-4b17-972a-05048027ddd5 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146361, 'name': PowerOffVM_Task, 'duration_secs': 0.347682} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2119.091028] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-7beba819-8d40-4b17-972a-05048027ddd5 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02f34565-6675-4c79-ac47-b131ceba9df8] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2119.091028] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-7beba819-8d40-4b17-972a-05048027ddd5 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02f34565-6675-4c79-ac47-b131ceba9df8] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2119.091028] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9fd878ca-a6f3-4f84-8864-86ed5c725efb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2119.137379] env[62824]: DEBUG oslo_concurrency.lockutils [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Acquiring lock "da1ad739-b252-4e29-a22a-ab3bcab173ec" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2119.137778] env[62824]: DEBUG oslo_concurrency.lockutils [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Lock "da1ad739-b252-4e29-a22a-ab3bcab173ec" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2119.172028] env[62824]: DEBUG oslo_vmware.api [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5283a37b-d459-0ecc-f062-94f5339f0e4f, 'name': SearchDatastore_Task, 'duration_secs': 0.012287} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2119.172028] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9fe9243a-38ae-4a61-814c-c2693c166203 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2119.175889] env[62824]: DEBUG oslo_vmware.api [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Waiting for the task: (returnval){ [ 2119.175889] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]524734c7-6153-dfac-9b35-25e3f5187433" [ 2119.175889] env[62824]: _type = "Task" [ 2119.175889] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2119.184398] env[62824]: DEBUG oslo_vmware.api [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]524734c7-6153-dfac-9b35-25e3f5187433, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2119.217401] env[62824]: DEBUG oslo_vmware.api [None req-4a76d89b-cbff-4ca0-a9aa-3ebc58bfda76 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146360, 'name': RemoveSnapshot_Task, 'duration_secs': 0.988625} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2119.218758] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-4a76d89b-cbff-4ca0-a9aa-3ebc58bfda76 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8] Deleted Snapshot of the VM instance {{(pid=62824) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 2119.221273] env[62824]: INFO nova.compute.manager [None req-4a76d89b-cbff-4ca0-a9aa-3ebc58bfda76 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8] Took 17.98 seconds to snapshot the instance on the hypervisor. [ 2119.270172] env[62824]: DEBUG nova.scheduler.client.report [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2119.320904] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-7beba819-8d40-4b17-972a-05048027ddd5 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02f34565-6675-4c79-ac47-b131ceba9df8] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2119.320904] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-7beba819-8d40-4b17-972a-05048027ddd5 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02f34565-6675-4c79-ac47-b131ceba9df8] Deleting contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2119.320904] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-7beba819-8d40-4b17-972a-05048027ddd5 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Deleting the datastore file [datastore1] 02f34565-6675-4c79-ac47-b131ceba9df8 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2119.320904] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e55993cd-c642-4ce1-b10a-1233ae0f31a6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2119.328747] env[62824]: DEBUG oslo_vmware.api [None req-7beba819-8d40-4b17-972a-05048027ddd5 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for the task: (returnval){ [ 2119.328747] env[62824]: value = "task-2146363" [ 2119.328747] env[62824]: _type = "Task" [ 2119.328747] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2119.338849] env[62824]: DEBUG oslo_vmware.api [None req-7beba819-8d40-4b17-972a-05048027ddd5 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146363, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2119.685115] env[62824]: DEBUG nova.network.neutron [req-2b378cba-9866-4c6a-87a9-679968c8fc12 req-3ec4a3bd-b8cb-486d-b1d0-345a1fe58037 service nova] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Updated VIF entry in instance network info cache for port 7be95c5e-6e98-4fec-8550-1c22a787b95b. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2119.686437] env[62824]: DEBUG nova.network.neutron [req-2b378cba-9866-4c6a-87a9-679968c8fc12 req-3ec4a3bd-b8cb-486d-b1d0-345a1fe58037 service nova] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Updating instance_info_cache with network_info: [{"id": "7be95c5e-6e98-4fec-8550-1c22a787b95b", "address": "fa:16:3e:e8:69:3a", "network": {"id": "bba9b98c-c7eb-4890-8de5-eafacfdebd7d", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1928394728-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.160", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7fcdbfc29df64a1ba8d982bdcc667b64", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b49c5024-2ced-42ca-90cc-6066766d43e6", "external-id": "nsx-vlan-transportzone-239", "segmentation_id": 239, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7be95c5e-6e", "ovs_interfaceid": "7be95c5e-6e98-4fec-8550-1c22a787b95b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2119.691419] env[62824]: DEBUG oslo_vmware.api [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]524734c7-6153-dfac-9b35-25e3f5187433, 'name': SearchDatastore_Task, 'duration_secs': 0.010037} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2119.691929] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2119.692198] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 49818910-61de-4fb8-bbab-d5d61d2f1ada/49818910-61de-4fb8-bbab-d5d61d2f1ada.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2119.692464] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c10bf608-ecb0-40c6-a050-ccfc73d502c3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2119.703614] env[62824]: DEBUG oslo_vmware.api [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Waiting for the task: (returnval){ [ 2119.703614] env[62824]: value = "task-2146364" [ 2119.703614] env[62824]: _type = "Task" [ 2119.703614] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2119.713808] env[62824]: DEBUG oslo_vmware.api [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146364, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2119.773400] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.669s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2119.773967] env[62824]: DEBUG nova.compute.manager [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 29140542-dc4b-411a-ac10-7d84086eabbb] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2119.777100] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.073s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2119.778634] env[62824]: INFO nova.compute.claims [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2119.840655] env[62824]: DEBUG oslo_vmware.api [None req-7beba819-8d40-4b17-972a-05048027ddd5 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146363, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.154359} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2119.840920] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-7beba819-8d40-4b17-972a-05048027ddd5 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2119.841190] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-7beba819-8d40-4b17-972a-05048027ddd5 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02f34565-6675-4c79-ac47-b131ceba9df8] Deleted contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2119.841607] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-7beba819-8d40-4b17-972a-05048027ddd5 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02f34565-6675-4c79-ac47-b131ceba9df8] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2119.841944] env[62824]: INFO nova.compute.manager [None req-7beba819-8d40-4b17-972a-05048027ddd5 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02f34565-6675-4c79-ac47-b131ceba9df8] Took 1.28 seconds to destroy the instance on the hypervisor. [ 2119.842247] env[62824]: DEBUG oslo.service.loopingcall [None req-7beba819-8d40-4b17-972a-05048027ddd5 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2119.842442] env[62824]: DEBUG nova.compute.manager [-] [instance: 02f34565-6675-4c79-ac47-b131ceba9df8] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2119.842549] env[62824]: DEBUG nova.network.neutron [-] [instance: 02f34565-6675-4c79-ac47-b131ceba9df8] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2120.122204] env[62824]: DEBUG oslo_concurrency.lockutils [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Acquiring lock "da46d1d5-f75f-4ef1-b571-fbebab89c2a6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2120.122518] env[62824]: DEBUG oslo_concurrency.lockutils [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Lock "da46d1d5-f75f-4ef1-b571-fbebab89c2a6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2120.193349] env[62824]: DEBUG oslo_concurrency.lockutils [req-2b378cba-9866-4c6a-87a9-679968c8fc12 req-3ec4a3bd-b8cb-486d-b1d0-345a1fe58037 service nova] Releasing lock "refresh_cache-5b3df31f-15fe-473f-992c-ddb272661c53" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2120.214596] env[62824]: DEBUG oslo_vmware.api [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146364, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.465893} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2120.214864] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 49818910-61de-4fb8-bbab-d5d61d2f1ada/49818910-61de-4fb8-bbab-d5d61d2f1ada.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2120.215113] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 49818910-61de-4fb8-bbab-d5d61d2f1ada] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2120.215374] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a87d10ef-706e-46a2-9747-4d6839969717 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2120.222317] env[62824]: DEBUG oslo_vmware.api [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Waiting for the task: (returnval){ [ 2120.222317] env[62824]: value = "task-2146365" [ 2120.222317] env[62824]: _type = "Task" [ 2120.222317] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2120.232765] env[62824]: DEBUG oslo_vmware.api [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146365, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2120.282847] env[62824]: DEBUG nova.compute.utils [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2120.287170] env[62824]: DEBUG nova.compute.manager [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 29140542-dc4b-411a-ac10-7d84086eabbb] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2120.287353] env[62824]: DEBUG nova.network.neutron [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 29140542-dc4b-411a-ac10-7d84086eabbb] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2120.380997] env[62824]: DEBUG nova.policy [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8bb15b0904de4f50b9d5f0ec23c4f7d4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '25bb8a4df0b740c0a1282ee61ad2f58c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2120.468396] env[62824]: DEBUG nova.compute.manager [req-6bd6b9cf-542b-40ec-9941-20e3bf1dc612 req-3005260d-dd5d-4403-9c1a-492f8aca20dd service nova] [instance: 02f34565-6675-4c79-ac47-b131ceba9df8] Received event network-vif-deleted-be36fef6-17a7-4c25-bf33-98cc6a037b53 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2120.468396] env[62824]: INFO nova.compute.manager [req-6bd6b9cf-542b-40ec-9941-20e3bf1dc612 req-3005260d-dd5d-4403-9c1a-492f8aca20dd service nova] [instance: 02f34565-6675-4c79-ac47-b131ceba9df8] Neutron deleted interface be36fef6-17a7-4c25-bf33-98cc6a037b53; detaching it from the instance and deleting it from the info cache [ 2120.468396] env[62824]: DEBUG nova.network.neutron [req-6bd6b9cf-542b-40ec-9941-20e3bf1dc612 req-3005260d-dd5d-4403-9c1a-492f8aca20dd service nova] [instance: 02f34565-6675-4c79-ac47-b131ceba9df8] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2120.735812] env[62824]: DEBUG oslo_vmware.api [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146365, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.269276} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2120.736921] env[62824]: DEBUG nova.network.neutron [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 29140542-dc4b-411a-ac10-7d84086eabbb] Successfully created port: 8ad74d1d-5091-4557-bb57-ba65d9b98b45 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2120.739129] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 49818910-61de-4fb8-bbab-d5d61d2f1ada] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2120.739608] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7480f35c-7105-4f80-a90c-1028f263ce8f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2120.757951] env[62824]: INFO nova.compute.manager [None req-711a313c-0171-44f0-b62d-de994a10eaf5 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Unrescuing [ 2120.758441] env[62824]: DEBUG oslo_concurrency.lockutils [None req-711a313c-0171-44f0-b62d-de994a10eaf5 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Acquiring lock "refresh_cache-363177c0-dc40-429a-a74b-e690da133edb" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2120.758607] env[62824]: DEBUG oslo_concurrency.lockutils [None req-711a313c-0171-44f0-b62d-de994a10eaf5 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Acquired lock "refresh_cache-363177c0-dc40-429a-a74b-e690da133edb" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2120.758787] env[62824]: DEBUG nova.network.neutron [None req-711a313c-0171-44f0-b62d-de994a10eaf5 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2120.771636] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 49818910-61de-4fb8-bbab-d5d61d2f1ada] Reconfiguring VM instance instance-00000044 to attach disk [datastore2] 49818910-61de-4fb8-bbab-d5d61d2f1ada/49818910-61de-4fb8-bbab-d5d61d2f1ada.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2120.771636] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f1f23e98-e28f-43dc-a46e-94d384fbf31a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2120.787108] env[62824]: DEBUG nova.network.neutron [-] [instance: 02f34565-6675-4c79-ac47-b131ceba9df8] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2120.788460] env[62824]: DEBUG nova.compute.manager [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 29140542-dc4b-411a-ac10-7d84086eabbb] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2120.796332] env[62824]: DEBUG oslo_vmware.api [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Waiting for the task: (returnval){ [ 2120.796332] env[62824]: value = "task-2146366" [ 2120.796332] env[62824]: _type = "Task" [ 2120.796332] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2120.807773] env[62824]: DEBUG oslo_vmware.api [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146366, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2120.972635] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1966abf7-de26-4ea6-9fda-34db57e65067 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2120.986024] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc844dfc-f3e8-46f3-8b4d-e36655ae7e80 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2121.035550] env[62824]: DEBUG nova.compute.manager [req-6bd6b9cf-542b-40ec-9941-20e3bf1dc612 req-3005260d-dd5d-4403-9c1a-492f8aca20dd service nova] [instance: 02f34565-6675-4c79-ac47-b131ceba9df8] Detach interface failed, port_id=be36fef6-17a7-4c25-bf33-98cc6a037b53, reason: Instance 02f34565-6675-4c79-ac47-b131ceba9df8 could not be found. {{(pid=62824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11482}} [ 2121.293078] env[62824]: INFO nova.compute.manager [-] [instance: 02f34565-6675-4c79-ac47-b131ceba9df8] Took 1.45 seconds to deallocate network for instance. [ 2121.318221] env[62824]: DEBUG oslo_vmware.api [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146366, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2121.334712] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7994db80-8c96-4d3b-9a45-930926146d01 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2121.343384] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9dfd6f6-1b07-410e-86d6-77a5d9e3fbb7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2121.385163] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-439f0c30-9047-4f4e-b13f-de8de6bcf624 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2121.393879] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f5bdbd3-c48c-4b6d-af0c-4231b4960df4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2121.408840] env[62824]: DEBUG nova.compute.provider_tree [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2121.520297] env[62824]: DEBUG nova.network.neutron [None req-711a313c-0171-44f0-b62d-de994a10eaf5 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Updating instance_info_cache with network_info: [{"id": "8e8b16ef-e3f8-48c1-aab6-51d94d356c77", "address": "fa:16:3e:48:83:8c", "network": {"id": "bcf30d77-cac1-4deb-971d-efb7c1fb8354", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1030043097-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "0503b434636c4ef5bae8db1b0c74d2ce", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3cc0a33d-17c0-4b87-b48f-413a87a4cc6a", "external-id": "nsx-vlan-transportzone-865", "segmentation_id": 865, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8e8b16ef-e3", "ovs_interfaceid": "8e8b16ef-e3f8-48c1-aab6-51d94d356c77", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2121.803096] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7beba819-8d40-4b17-972a-05048027ddd5 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2121.809995] env[62824]: DEBUG nova.compute.manager [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 29140542-dc4b-411a-ac10-7d84086eabbb] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2121.814887] env[62824]: DEBUG oslo_vmware.api [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146366, 'name': ReconfigVM_Task, 'duration_secs': 0.909805} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2121.815180] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 49818910-61de-4fb8-bbab-d5d61d2f1ada] Reconfigured VM instance instance-00000044 to attach disk [datastore2] 49818910-61de-4fb8-bbab-d5d61d2f1ada/49818910-61de-4fb8-bbab-d5d61d2f1ada.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2121.815767] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-af8200be-82ac-4e7a-b32a-badeefb4aa7d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2121.822833] env[62824]: DEBUG oslo_vmware.api [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Waiting for the task: (returnval){ [ 2121.822833] env[62824]: value = "task-2146367" [ 2121.822833] env[62824]: _type = "Task" [ 2121.822833] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2121.835494] env[62824]: DEBUG oslo_vmware.api [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146367, 'name': Rename_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2121.837734] env[62824]: DEBUG nova.virt.hardware [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2121.838183] env[62824]: DEBUG nova.virt.hardware [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2121.838255] env[62824]: DEBUG nova.virt.hardware [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2121.838400] env[62824]: DEBUG nova.virt.hardware [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2121.838543] env[62824]: DEBUG nova.virt.hardware [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2121.838688] env[62824]: DEBUG nova.virt.hardware [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2121.838894] env[62824]: DEBUG nova.virt.hardware [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2121.839101] env[62824]: DEBUG nova.virt.hardware [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2121.839264] env[62824]: DEBUG nova.virt.hardware [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2121.839425] env[62824]: DEBUG nova.virt.hardware [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2121.839593] env[62824]: DEBUG nova.virt.hardware [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2121.840383] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01dffe58-168a-47b6-a8bb-09b66a61d485 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2121.848050] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38121f37-72dd-40d6-a75c-6f811bada40c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2121.914163] env[62824]: DEBUG nova.scheduler.client.report [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2122.023557] env[62824]: DEBUG oslo_concurrency.lockutils [None req-711a313c-0171-44f0-b62d-de994a10eaf5 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Releasing lock "refresh_cache-363177c0-dc40-429a-a74b-e690da133edb" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2122.024294] env[62824]: DEBUG nova.objects.instance [None req-711a313c-0171-44f0-b62d-de994a10eaf5 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Lazy-loading 'flavor' on Instance uuid 363177c0-dc40-429a-a74b-e690da133edb {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2122.333471] env[62824]: DEBUG oslo_vmware.api [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146367, 'name': Rename_Task, 'duration_secs': 0.347574} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2122.333799] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 49818910-61de-4fb8-bbab-d5d61d2f1ada] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2122.333995] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-33330165-54a0-4caf-ae68-898ef6729a0e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2122.344864] env[62824]: DEBUG oslo_vmware.api [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Waiting for the task: (returnval){ [ 2122.344864] env[62824]: value = "task-2146368" [ 2122.344864] env[62824]: _type = "Task" [ 2122.344864] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2122.356594] env[62824]: DEBUG oslo_vmware.api [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146368, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2122.420273] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.643s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2122.421016] env[62824]: DEBUG nova.compute.manager [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2122.425106] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.706s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2122.427366] env[62824]: INFO nova.compute.claims [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 4c2c2068-664d-404f-a99e-8fc7719f43e5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2122.493552] env[62824]: DEBUG nova.compute.manager [req-f3541485-d000-47a4-9890-b0a7fce352ba req-c28cbd2b-81d7-497d-87e5-09bfac5ac555 service nova] [instance: 29140542-dc4b-411a-ac10-7d84086eabbb] Received event network-vif-plugged-8ad74d1d-5091-4557-bb57-ba65d9b98b45 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2122.493720] env[62824]: DEBUG oslo_concurrency.lockutils [req-f3541485-d000-47a4-9890-b0a7fce352ba req-c28cbd2b-81d7-497d-87e5-09bfac5ac555 service nova] Acquiring lock "29140542-dc4b-411a-ac10-7d84086eabbb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2122.494082] env[62824]: DEBUG oslo_concurrency.lockutils [req-f3541485-d000-47a4-9890-b0a7fce352ba req-c28cbd2b-81d7-497d-87e5-09bfac5ac555 service nova] Lock "29140542-dc4b-411a-ac10-7d84086eabbb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2122.494275] env[62824]: DEBUG oslo_concurrency.lockutils [req-f3541485-d000-47a4-9890-b0a7fce352ba req-c28cbd2b-81d7-497d-87e5-09bfac5ac555 service nova] Lock "29140542-dc4b-411a-ac10-7d84086eabbb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2122.494442] env[62824]: DEBUG nova.compute.manager [req-f3541485-d000-47a4-9890-b0a7fce352ba req-c28cbd2b-81d7-497d-87e5-09bfac5ac555 service nova] [instance: 29140542-dc4b-411a-ac10-7d84086eabbb] No waiting events found dispatching network-vif-plugged-8ad74d1d-5091-4557-bb57-ba65d9b98b45 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2122.494670] env[62824]: WARNING nova.compute.manager [req-f3541485-d000-47a4-9890-b0a7fce352ba req-c28cbd2b-81d7-497d-87e5-09bfac5ac555 service nova] [instance: 29140542-dc4b-411a-ac10-7d84086eabbb] Received unexpected event network-vif-plugged-8ad74d1d-5091-4557-bb57-ba65d9b98b45 for instance with vm_state building and task_state spawning. [ 2122.529665] env[62824]: DEBUG nova.network.neutron [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 29140542-dc4b-411a-ac10-7d84086eabbb] Successfully updated port: 8ad74d1d-5091-4557-bb57-ba65d9b98b45 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2122.531634] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-543d38d4-750f-4f42-b24c-71925ba871f1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2122.556502] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-711a313c-0171-44f0-b62d-de994a10eaf5 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2122.557545] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7af26161-4d64-4cf7-8427-7afd45da9b15 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2122.566911] env[62824]: DEBUG oslo_vmware.api [None req-711a313c-0171-44f0-b62d-de994a10eaf5 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Waiting for the task: (returnval){ [ 2122.566911] env[62824]: value = "task-2146369" [ 2122.566911] env[62824]: _type = "Task" [ 2122.566911] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2122.577858] env[62824]: DEBUG oslo_vmware.api [None req-711a313c-0171-44f0-b62d-de994a10eaf5 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146369, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2122.855919] env[62824]: DEBUG oslo_vmware.api [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146368, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2122.938538] env[62824]: DEBUG nova.compute.utils [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2122.938538] env[62824]: DEBUG nova.compute.manager [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2122.938538] env[62824]: DEBUG nova.network.neutron [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2123.006613] env[62824]: DEBUG nova.policy [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '959d127a7d144b33a0cae94db5c11846', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dfff08982dad4790bf4d555e2b4db5e4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2123.035738] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Acquiring lock "refresh_cache-29140542-dc4b-411a-ac10-7d84086eabbb" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2123.035865] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Acquired lock "refresh_cache-29140542-dc4b-411a-ac10-7d84086eabbb" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2123.036172] env[62824]: DEBUG nova.network.neutron [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 29140542-dc4b-411a-ac10-7d84086eabbb] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2123.077495] env[62824]: DEBUG oslo_vmware.api [None req-711a313c-0171-44f0-b62d-de994a10eaf5 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146369, 'name': PowerOffVM_Task, 'duration_secs': 0.299865} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2123.077845] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-711a313c-0171-44f0-b62d-de994a10eaf5 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2123.083678] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-711a313c-0171-44f0-b62d-de994a10eaf5 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Reconfiguring VM instance instance-00000041 to detach disk 2001 {{(pid=62824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 2123.084368] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3e1ceb8d-ef53-4b94-bf5d-f868243c4cdf {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2123.104011] env[62824]: DEBUG oslo_vmware.api [None req-711a313c-0171-44f0-b62d-de994a10eaf5 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Waiting for the task: (returnval){ [ 2123.104011] env[62824]: value = "task-2146370" [ 2123.104011] env[62824]: _type = "Task" [ 2123.104011] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2123.111760] env[62824]: DEBUG oslo_vmware.api [None req-711a313c-0171-44f0-b62d-de994a10eaf5 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146370, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2123.356070] env[62824]: DEBUG oslo_vmware.api [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146368, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2123.387975] env[62824]: DEBUG nova.network.neutron [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e] Successfully created port: 9ad5243a-5d04-419d-bfd7-e4b27536ee65 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2123.444716] env[62824]: DEBUG nova.compute.manager [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2123.584971] env[62824]: DEBUG nova.network.neutron [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 29140542-dc4b-411a-ac10-7d84086eabbb] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2123.614910] env[62824]: DEBUG oslo_vmware.api [None req-711a313c-0171-44f0-b62d-de994a10eaf5 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146370, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2123.732778] env[62824]: DEBUG nova.network.neutron [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 29140542-dc4b-411a-ac10-7d84086eabbb] Updating instance_info_cache with network_info: [{"id": "8ad74d1d-5091-4557-bb57-ba65d9b98b45", "address": "fa:16:3e:bd:76:32", "network": {"id": "4ee2cade-6795-4887-a1df-f3990219e980", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1173612309-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "25bb8a4df0b740c0a1282ee61ad2f58c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3202e2b9-c4a7-4f78-9476-12ed92fabe61", "external-id": "nsx-vlan-transportzone-846", "segmentation_id": 846, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8ad74d1d-50", "ovs_interfaceid": "8ad74d1d-5091-4557-bb57-ba65d9b98b45", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2123.859133] env[62824]: DEBUG oslo_vmware.api [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146368, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2123.920405] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41623997-c1fc-4e2d-aff7-2d8f0281b1cd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2123.928012] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a2dc7b9-7dda-4954-a4c4-338df5b0cbdf {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2123.963554] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c981e03-458a-488a-ae6d-c7bb90d4baba {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2123.972316] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ba9e84c-89b7-47d8-810e-e7fc16dc5fa8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2123.987390] env[62824]: DEBUG nova.compute.provider_tree [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2124.117268] env[62824]: DEBUG oslo_vmware.api [None req-711a313c-0171-44f0-b62d-de994a10eaf5 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146370, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2124.235669] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Releasing lock "refresh_cache-29140542-dc4b-411a-ac10-7d84086eabbb" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2124.236024] env[62824]: DEBUG nova.compute.manager [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 29140542-dc4b-411a-ac10-7d84086eabbb] Instance network_info: |[{"id": "8ad74d1d-5091-4557-bb57-ba65d9b98b45", "address": "fa:16:3e:bd:76:32", "network": {"id": "4ee2cade-6795-4887-a1df-f3990219e980", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1173612309-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "25bb8a4df0b740c0a1282ee61ad2f58c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3202e2b9-c4a7-4f78-9476-12ed92fabe61", "external-id": "nsx-vlan-transportzone-846", "segmentation_id": 846, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8ad74d1d-50", "ovs_interfaceid": "8ad74d1d-5091-4557-bb57-ba65d9b98b45", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2124.236494] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 29140542-dc4b-411a-ac10-7d84086eabbb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:bd:76:32', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3202e2b9-c4a7-4f78-9476-12ed92fabe61', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8ad74d1d-5091-4557-bb57-ba65d9b98b45', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2124.244029] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Creating folder: Project (25bb8a4df0b740c0a1282ee61ad2f58c). Parent ref: group-v438503. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 2124.244333] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8c1f897c-5340-44ad-927f-da24f71ad746 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2124.256366] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Created folder: Project (25bb8a4df0b740c0a1282ee61ad2f58c) in parent group-v438503. [ 2124.256562] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Creating folder: Instances. Parent ref: group-v438684. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 2124.256853] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c6facbd4-3b59-4f40-89f4-02f5ee6f9eb9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2124.266092] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Created folder: Instances in parent group-v438684. [ 2124.266328] env[62824]: DEBUG oslo.service.loopingcall [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2124.266514] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 29140542-dc4b-411a-ac10-7d84086eabbb] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2124.266731] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-32a6b84e-fc01-470e-b576-9e07dcb51a17 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2124.284417] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2124.284417] env[62824]: value = "task-2146373" [ 2124.284417] env[62824]: _type = "Task" [ 2124.284417] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2124.291805] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146373, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2124.357407] env[62824]: DEBUG oslo_vmware.api [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146368, 'name': PowerOnVM_Task, 'duration_secs': 1.625545} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2124.357686] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 49818910-61de-4fb8-bbab-d5d61d2f1ada] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2124.357845] env[62824]: INFO nova.compute.manager [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 49818910-61de-4fb8-bbab-d5d61d2f1ada] Took 10.19 seconds to spawn the instance on the hypervisor. [ 2124.358095] env[62824]: DEBUG nova.compute.manager [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 49818910-61de-4fb8-bbab-d5d61d2f1ada] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2124.358888] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aed793c0-fa4b-4307-9233-699a507841c7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2124.468871] env[62824]: DEBUG nova.compute.manager [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2124.491937] env[62824]: DEBUG nova.scheduler.client.report [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2124.502721] env[62824]: DEBUG nova.virt.hardware [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2124.502888] env[62824]: DEBUG nova.virt.hardware [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2124.503450] env[62824]: DEBUG nova.virt.hardware [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2124.503450] env[62824]: DEBUG nova.virt.hardware [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2124.503450] env[62824]: DEBUG nova.virt.hardware [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2124.503634] env[62824]: DEBUG nova.virt.hardware [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2124.503808] env[62824]: DEBUG nova.virt.hardware [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2124.503991] env[62824]: DEBUG nova.virt.hardware [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2124.504163] env[62824]: DEBUG nova.virt.hardware [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2124.504329] env[62824]: DEBUG nova.virt.hardware [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2124.504512] env[62824]: DEBUG nova.virt.hardware [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2124.505429] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-581745e8-5f4f-4651-ba1d-60299d3cbcb6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2124.516189] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e684ea2-77eb-4700-ad9f-d851bc7dc060 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2124.600551] env[62824]: DEBUG nova.compute.manager [req-4744f1e9-f129-4f4c-a708-e86488682d7a req-f55d67c6-7164-4a12-96e3-5ac0a15f6923 service nova] [instance: 29140542-dc4b-411a-ac10-7d84086eabbb] Received event network-changed-8ad74d1d-5091-4557-bb57-ba65d9b98b45 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2124.600767] env[62824]: DEBUG nova.compute.manager [req-4744f1e9-f129-4f4c-a708-e86488682d7a req-f55d67c6-7164-4a12-96e3-5ac0a15f6923 service nova] [instance: 29140542-dc4b-411a-ac10-7d84086eabbb] Refreshing instance network info cache due to event network-changed-8ad74d1d-5091-4557-bb57-ba65d9b98b45. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2124.601213] env[62824]: DEBUG oslo_concurrency.lockutils [req-4744f1e9-f129-4f4c-a708-e86488682d7a req-f55d67c6-7164-4a12-96e3-5ac0a15f6923 service nova] Acquiring lock "refresh_cache-29140542-dc4b-411a-ac10-7d84086eabbb" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2124.601423] env[62824]: DEBUG oslo_concurrency.lockutils [req-4744f1e9-f129-4f4c-a708-e86488682d7a req-f55d67c6-7164-4a12-96e3-5ac0a15f6923 service nova] Acquired lock "refresh_cache-29140542-dc4b-411a-ac10-7d84086eabbb" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2124.601641] env[62824]: DEBUG nova.network.neutron [req-4744f1e9-f129-4f4c-a708-e86488682d7a req-f55d67c6-7164-4a12-96e3-5ac0a15f6923 service nova] [instance: 29140542-dc4b-411a-ac10-7d84086eabbb] Refreshing network info cache for port 8ad74d1d-5091-4557-bb57-ba65d9b98b45 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2124.618077] env[62824]: DEBUG oslo_vmware.api [None req-711a313c-0171-44f0-b62d-de994a10eaf5 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146370, 'name': ReconfigVM_Task, 'duration_secs': 1.299342} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2124.618393] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-711a313c-0171-44f0-b62d-de994a10eaf5 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Reconfigured VM instance instance-00000041 to detach disk 2001 {{(pid=62824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 2124.618589] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-711a313c-0171-44f0-b62d-de994a10eaf5 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2124.618819] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1e09442a-cb82-4968-b4c2-630e63840513 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2124.628216] env[62824]: DEBUG oslo_vmware.api [None req-711a313c-0171-44f0-b62d-de994a10eaf5 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Waiting for the task: (returnval){ [ 2124.628216] env[62824]: value = "task-2146374" [ 2124.628216] env[62824]: _type = "Task" [ 2124.628216] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2124.636928] env[62824]: DEBUG oslo_vmware.api [None req-711a313c-0171-44f0-b62d-de994a10eaf5 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146374, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2124.794490] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146373, 'name': CreateVM_Task, 'duration_secs': 0.382387} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2124.794607] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 29140542-dc4b-411a-ac10-7d84086eabbb] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2124.795379] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2124.795541] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2124.795877] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2124.796149] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a4ad0e11-ba82-4c25-a49c-253d6cf8db22 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2124.800848] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Waiting for the task: (returnval){ [ 2124.800848] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52b465ff-f6c7-3031-b9f4-fc03b6448733" [ 2124.800848] env[62824]: _type = "Task" [ 2124.800848] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2124.809036] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52b465ff-f6c7-3031-b9f4-fc03b6448733, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2124.878910] env[62824]: INFO nova.compute.manager [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 49818910-61de-4fb8-bbab-d5d61d2f1ada] Took 37.42 seconds to build instance. [ 2124.997216] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.572s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2124.998027] env[62824]: DEBUG nova.compute.manager [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 4c2c2068-664d-404f-a99e-8fc7719f43e5] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2125.002922] env[62824]: DEBUG oslo_concurrency.lockutils [None req-123477fc-8c3e-4f2a-83de-cced4f011020 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 33.065s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2125.003268] env[62824]: DEBUG nova.objects.instance [None req-123477fc-8c3e-4f2a-83de-cced4f011020 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Lazy-loading 'pci_requests' on Instance uuid b8cc8cd3-ea03-40bf-b867-7ad193365552 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2125.068120] env[62824]: DEBUG nova.network.neutron [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e] Successfully updated port: 9ad5243a-5d04-419d-bfd7-e4b27536ee65 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2125.139252] env[62824]: DEBUG oslo_vmware.api [None req-711a313c-0171-44f0-b62d-de994a10eaf5 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146374, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2125.311276] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52b465ff-f6c7-3031-b9f4-fc03b6448733, 'name': SearchDatastore_Task, 'duration_secs': 0.009711} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2125.313694] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2125.313975] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 29140542-dc4b-411a-ac10-7d84086eabbb] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2125.314247] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2125.314399] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2125.314585] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2125.314879] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-de792078-f8a2-4023-a090-b6d1af969a9f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2125.323330] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2125.323507] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2125.324249] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2c9c170c-67b6-44c0-ac6a-158d7a02b2f6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2125.330600] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Waiting for the task: (returnval){ [ 2125.330600] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52c4a36f-d8f4-0b74-5fa6-218ac0d43254" [ 2125.330600] env[62824]: _type = "Task" [ 2125.330600] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2125.338152] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52c4a36f-d8f4-0b74-5fa6-218ac0d43254, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2125.383753] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b5dd8f11-213b-434c-b09d-59df9ec595d1 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Lock "49818910-61de-4fb8-bbab-d5d61d2f1ada" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 38.938s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2125.487493] env[62824]: DEBUG nova.network.neutron [req-4744f1e9-f129-4f4c-a708-e86488682d7a req-f55d67c6-7164-4a12-96e3-5ac0a15f6923 service nova] [instance: 29140542-dc4b-411a-ac10-7d84086eabbb] Updated VIF entry in instance network info cache for port 8ad74d1d-5091-4557-bb57-ba65d9b98b45. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2125.488297] env[62824]: DEBUG nova.network.neutron [req-4744f1e9-f129-4f4c-a708-e86488682d7a req-f55d67c6-7164-4a12-96e3-5ac0a15f6923 service nova] [instance: 29140542-dc4b-411a-ac10-7d84086eabbb] Updating instance_info_cache with network_info: [{"id": "8ad74d1d-5091-4557-bb57-ba65d9b98b45", "address": "fa:16:3e:bd:76:32", "network": {"id": "4ee2cade-6795-4887-a1df-f3990219e980", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1173612309-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "25bb8a4df0b740c0a1282ee61ad2f58c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3202e2b9-c4a7-4f78-9476-12ed92fabe61", "external-id": "nsx-vlan-transportzone-846", "segmentation_id": 846, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8ad74d1d-50", "ovs_interfaceid": "8ad74d1d-5091-4557-bb57-ba65d9b98b45", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2125.506596] env[62824]: DEBUG nova.compute.utils [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2125.508156] env[62824]: DEBUG nova.compute.manager [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 4c2c2068-664d-404f-a99e-8fc7719f43e5] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2125.508348] env[62824]: DEBUG nova.network.neutron [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 4c2c2068-664d-404f-a99e-8fc7719f43e5] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2125.511610] env[62824]: DEBUG nova.objects.instance [None req-123477fc-8c3e-4f2a-83de-cced4f011020 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Lazy-loading 'numa_topology' on Instance uuid b8cc8cd3-ea03-40bf-b867-7ad193365552 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2125.562689] env[62824]: DEBUG nova.policy [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8bb15b0904de4f50b9d5f0ec23c4f7d4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '25bb8a4df0b740c0a1282ee61ad2f58c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2125.575207] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquiring lock "refresh_cache-bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2125.575357] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquired lock "refresh_cache-bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2125.575510] env[62824]: DEBUG nova.network.neutron [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2125.641442] env[62824]: DEBUG oslo_vmware.api [None req-711a313c-0171-44f0-b62d-de994a10eaf5 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146374, 'name': PowerOnVM_Task, 'duration_secs': 0.544701} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2125.641807] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-711a313c-0171-44f0-b62d-de994a10eaf5 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2125.642063] env[62824]: DEBUG nova.compute.manager [None req-711a313c-0171-44f0-b62d-de994a10eaf5 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2125.642919] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-034275d6-7d95-4045-80e6-6d175edc237d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2125.841923] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52c4a36f-d8f4-0b74-5fa6-218ac0d43254, 'name': SearchDatastore_Task, 'duration_secs': 0.008448} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2125.842745] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-da666357-d704-4c5b-b854-c30d831468ed {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2125.848348] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Waiting for the task: (returnval){ [ 2125.848348] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52fb2036-634a-81fd-6ad6-0975e2c29b1a" [ 2125.848348] env[62824]: _type = "Task" [ 2125.848348] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2125.856555] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52fb2036-634a-81fd-6ad6-0975e2c29b1a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2125.887304] env[62824]: DEBUG nova.compute.manager [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 2afd1f18-234c-40b5-9ede-7413ad30dafe] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2125.992683] env[62824]: DEBUG oslo_concurrency.lockutils [req-4744f1e9-f129-4f4c-a708-e86488682d7a req-f55d67c6-7164-4a12-96e3-5ac0a15f6923 service nova] Releasing lock "refresh_cache-29140542-dc4b-411a-ac10-7d84086eabbb" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2126.014100] env[62824]: DEBUG nova.compute.manager [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 4c2c2068-664d-404f-a99e-8fc7719f43e5] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2126.016891] env[62824]: INFO nova.compute.claims [None req-123477fc-8c3e-4f2a-83de-cced4f011020 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2126.044291] env[62824]: DEBUG nova.network.neutron [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 4c2c2068-664d-404f-a99e-8fc7719f43e5] Successfully created port: 96a12f9d-8a1e-4df8-b5fa-24c8b92c413f {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2126.109399] env[62824]: DEBUG nova.network.neutron [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2126.194174] env[62824]: DEBUG nova.compute.manager [None req-9c371f91-b712-4ee9-8e58-6993dc153ed4 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 49818910-61de-4fb8-bbab-d5d61d2f1ada] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2126.195021] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-182c7f85-a0b2-4b61-8938-426d233673a9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2126.254182] env[62824]: DEBUG nova.network.neutron [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e] Updating instance_info_cache with network_info: [{"id": "9ad5243a-5d04-419d-bfd7-e4b27536ee65", "address": "fa:16:3e:c0:de:cb", "network": {"id": "0f512108-670f-4a85-9381-108e18b3b5ed", "bridge": "br-int", "label": "tempest-ServersTestJSON-766397782-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dfff08982dad4790bf4d555e2b4db5e4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f3a2eb5-353f-45c5-a73b-869626f4bb13", "external-id": "nsx-vlan-transportzone-411", "segmentation_id": 411, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9ad5243a-5d", "ovs_interfaceid": "9ad5243a-5d04-419d-bfd7-e4b27536ee65", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2126.358020] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52fb2036-634a-81fd-6ad6-0975e2c29b1a, 'name': SearchDatastore_Task, 'duration_secs': 0.009995} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2126.358315] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2126.358582] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 29140542-dc4b-411a-ac10-7d84086eabbb/29140542-dc4b-411a-ac10-7d84086eabbb.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2126.358846] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6c9c4446-ae80-456f-b7a0-e997ce53d709 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2126.365598] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Waiting for the task: (returnval){ [ 2126.365598] env[62824]: value = "task-2146375" [ 2126.365598] env[62824]: _type = "Task" [ 2126.365598] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2126.373261] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Task: {'id': task-2146375, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2126.411666] env[62824]: DEBUG oslo_concurrency.lockutils [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2126.666394] env[62824]: DEBUG nova.compute.manager [req-1e414d60-f6ee-4ac8-9b04-267d54eeebc4 req-5c53f72c-2013-4334-aabd-eabdb7ab791e service nova] [instance: bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e] Received event network-vif-plugged-9ad5243a-5d04-419d-bfd7-e4b27536ee65 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2126.666656] env[62824]: DEBUG oslo_concurrency.lockutils [req-1e414d60-f6ee-4ac8-9b04-267d54eeebc4 req-5c53f72c-2013-4334-aabd-eabdb7ab791e service nova] Acquiring lock "bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2126.667009] env[62824]: DEBUG oslo_concurrency.lockutils [req-1e414d60-f6ee-4ac8-9b04-267d54eeebc4 req-5c53f72c-2013-4334-aabd-eabdb7ab791e service nova] Lock "bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2126.667189] env[62824]: DEBUG oslo_concurrency.lockutils [req-1e414d60-f6ee-4ac8-9b04-267d54eeebc4 req-5c53f72c-2013-4334-aabd-eabdb7ab791e service nova] Lock "bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2126.667368] env[62824]: DEBUG nova.compute.manager [req-1e414d60-f6ee-4ac8-9b04-267d54eeebc4 req-5c53f72c-2013-4334-aabd-eabdb7ab791e service nova] [instance: bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e] No waiting events found dispatching network-vif-plugged-9ad5243a-5d04-419d-bfd7-e4b27536ee65 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2126.667567] env[62824]: WARNING nova.compute.manager [req-1e414d60-f6ee-4ac8-9b04-267d54eeebc4 req-5c53f72c-2013-4334-aabd-eabdb7ab791e service nova] [instance: bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e] Received unexpected event network-vif-plugged-9ad5243a-5d04-419d-bfd7-e4b27536ee65 for instance with vm_state building and task_state spawning. [ 2126.667750] env[62824]: DEBUG nova.compute.manager [req-1e414d60-f6ee-4ac8-9b04-267d54eeebc4 req-5c53f72c-2013-4334-aabd-eabdb7ab791e service nova] [instance: bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e] Received event network-changed-9ad5243a-5d04-419d-bfd7-e4b27536ee65 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2126.667961] env[62824]: DEBUG nova.compute.manager [req-1e414d60-f6ee-4ac8-9b04-267d54eeebc4 req-5c53f72c-2013-4334-aabd-eabdb7ab791e service nova] [instance: bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e] Refreshing instance network info cache due to event network-changed-9ad5243a-5d04-419d-bfd7-e4b27536ee65. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2126.668171] env[62824]: DEBUG oslo_concurrency.lockutils [req-1e414d60-f6ee-4ac8-9b04-267d54eeebc4 req-5c53f72c-2013-4334-aabd-eabdb7ab791e service nova] Acquiring lock "refresh_cache-bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2126.707118] env[62824]: INFO nova.compute.manager [None req-9c371f91-b712-4ee9-8e58-6993dc153ed4 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 49818910-61de-4fb8-bbab-d5d61d2f1ada] instance snapshotting [ 2126.710995] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f81e0896-b041-434f-8f62-b926862394c7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2126.740334] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-016f8267-6be4-4b4e-98cb-c4132649b9dd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2126.757602] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Releasing lock "refresh_cache-bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2126.757947] env[62824]: DEBUG nova.compute.manager [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e] Instance network_info: |[{"id": "9ad5243a-5d04-419d-bfd7-e4b27536ee65", "address": "fa:16:3e:c0:de:cb", "network": {"id": "0f512108-670f-4a85-9381-108e18b3b5ed", "bridge": "br-int", "label": "tempest-ServersTestJSON-766397782-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dfff08982dad4790bf4d555e2b4db5e4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f3a2eb5-353f-45c5-a73b-869626f4bb13", "external-id": "nsx-vlan-transportzone-411", "segmentation_id": 411, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9ad5243a-5d", "ovs_interfaceid": "9ad5243a-5d04-419d-bfd7-e4b27536ee65", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2126.758491] env[62824]: DEBUG oslo_concurrency.lockutils [req-1e414d60-f6ee-4ac8-9b04-267d54eeebc4 req-5c53f72c-2013-4334-aabd-eabdb7ab791e service nova] Acquired lock "refresh_cache-bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2126.758677] env[62824]: DEBUG nova.network.neutron [req-1e414d60-f6ee-4ac8-9b04-267d54eeebc4 req-5c53f72c-2013-4334-aabd-eabdb7ab791e service nova] [instance: bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e] Refreshing network info cache for port 9ad5243a-5d04-419d-bfd7-e4b27536ee65 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2126.759960] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c0:de:cb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9f3a2eb5-353f-45c5-a73b-869626f4bb13', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9ad5243a-5d04-419d-bfd7-e4b27536ee65', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2126.768290] env[62824]: DEBUG oslo.service.loopingcall [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2126.768848] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2126.769086] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d79c9f8a-0fde-4ba3-95e4-43957c54a96c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2126.797436] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2126.797436] env[62824]: value = "task-2146376" [ 2126.797436] env[62824]: _type = "Task" [ 2126.797436] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2126.811312] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146376, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2126.881559] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Task: {'id': task-2146375, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2126.995710] env[62824]: DEBUG oslo_concurrency.lockutils [None req-66e16106-e2f9-4347-968f-3ba76efdbbd5 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Acquiring lock "363177c0-dc40-429a-a74b-e690da133edb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2126.995965] env[62824]: DEBUG oslo_concurrency.lockutils [None req-66e16106-e2f9-4347-968f-3ba76efdbbd5 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Lock "363177c0-dc40-429a-a74b-e690da133edb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2126.996193] env[62824]: DEBUG oslo_concurrency.lockutils [None req-66e16106-e2f9-4347-968f-3ba76efdbbd5 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Acquiring lock "363177c0-dc40-429a-a74b-e690da133edb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2126.996379] env[62824]: DEBUG oslo_concurrency.lockutils [None req-66e16106-e2f9-4347-968f-3ba76efdbbd5 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Lock "363177c0-dc40-429a-a74b-e690da133edb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2126.996549] env[62824]: DEBUG oslo_concurrency.lockutils [None req-66e16106-e2f9-4347-968f-3ba76efdbbd5 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Lock "363177c0-dc40-429a-a74b-e690da133edb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2127.001781] env[62824]: INFO nova.compute.manager [None req-66e16106-e2f9-4347-968f-3ba76efdbbd5 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Terminating instance [ 2127.027835] env[62824]: DEBUG nova.compute.manager [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 4c2c2068-664d-404f-a99e-8fc7719f43e5] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2127.051155] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-292ff58b-0455-4151-9b20-13188766d083 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2127.056579] env[62824]: DEBUG nova.virt.hardware [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2127.056579] env[62824]: DEBUG nova.virt.hardware [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2127.056579] env[62824]: DEBUG nova.virt.hardware [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2127.056579] env[62824]: DEBUG nova.virt.hardware [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2127.056828] env[62824]: DEBUG nova.virt.hardware [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2127.056860] env[62824]: DEBUG nova.virt.hardware [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2127.057075] env[62824]: DEBUG nova.virt.hardware [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2127.057244] env[62824]: DEBUG nova.virt.hardware [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2127.057416] env[62824]: DEBUG nova.virt.hardware [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2127.057580] env[62824]: DEBUG nova.virt.hardware [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2127.058540] env[62824]: DEBUG nova.virt.hardware [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2127.058617] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d51c1fe0-70b0-482a-83dd-3ab9b7c1c82b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2127.068217] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f3436f9-c634-4876-b2a4-8f623c67565b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2127.072328] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ba694cd-a22b-4cd6-8e75-823819b68708 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2127.112334] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b1374c3-b004-4c07-98c5-5582d520d6dc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2127.120064] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7d05213-9db2-4edd-8383-19ddf6a1e753 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2127.133676] env[62824]: DEBUG nova.compute.provider_tree [None req-123477fc-8c3e-4f2a-83de-cced4f011020 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2127.255693] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-9c371f91-b712-4ee9-8e58-6993dc153ed4 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 49818910-61de-4fb8-bbab-d5d61d2f1ada] Creating Snapshot of the VM instance {{(pid=62824) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 2127.256024] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-bfd03d16-5c6f-4de1-8c20-5744084f0c72 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2127.263438] env[62824]: DEBUG oslo_vmware.api [None req-9c371f91-b712-4ee9-8e58-6993dc153ed4 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Waiting for the task: (returnval){ [ 2127.263438] env[62824]: value = "task-2146377" [ 2127.263438] env[62824]: _type = "Task" [ 2127.263438] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2127.273538] env[62824]: DEBUG oslo_vmware.api [None req-9c371f91-b712-4ee9-8e58-6993dc153ed4 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146377, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2127.308259] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146376, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2127.377621] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Task: {'id': task-2146375, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.59468} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2127.380818] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 29140542-dc4b-411a-ac10-7d84086eabbb/29140542-dc4b-411a-ac10-7d84086eabbb.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2127.381177] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 29140542-dc4b-411a-ac10-7d84086eabbb] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2127.381544] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7589eb0f-106e-42b9-a060-5ec334d1bd6c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2127.389276] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Waiting for the task: (returnval){ [ 2127.389276] env[62824]: value = "task-2146378" [ 2127.389276] env[62824]: _type = "Task" [ 2127.389276] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2127.400925] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Task: {'id': task-2146378, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2127.507031] env[62824]: DEBUG nova.compute.manager [None req-66e16106-e2f9-4347-968f-3ba76efdbbd5 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2127.507361] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-66e16106-e2f9-4347-968f-3ba76efdbbd5 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2127.508291] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36217d9c-4c0b-4254-b57f-d24915a5349a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2127.520086] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-66e16106-e2f9-4347-968f-3ba76efdbbd5 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2127.520342] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-16541388-0a46-46a9-a785-6e4a2ac8921c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2127.527815] env[62824]: DEBUG oslo_vmware.api [None req-66e16106-e2f9-4347-968f-3ba76efdbbd5 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Waiting for the task: (returnval){ [ 2127.527815] env[62824]: value = "task-2146379" [ 2127.527815] env[62824]: _type = "Task" [ 2127.527815] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2127.532260] env[62824]: DEBUG nova.network.neutron [req-1e414d60-f6ee-4ac8-9b04-267d54eeebc4 req-5c53f72c-2013-4334-aabd-eabdb7ab791e service nova] [instance: bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e] Updated VIF entry in instance network info cache for port 9ad5243a-5d04-419d-bfd7-e4b27536ee65. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2127.532520] env[62824]: DEBUG nova.network.neutron [req-1e414d60-f6ee-4ac8-9b04-267d54eeebc4 req-5c53f72c-2013-4334-aabd-eabdb7ab791e service nova] [instance: bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e] Updating instance_info_cache with network_info: [{"id": "9ad5243a-5d04-419d-bfd7-e4b27536ee65", "address": "fa:16:3e:c0:de:cb", "network": {"id": "0f512108-670f-4a85-9381-108e18b3b5ed", "bridge": "br-int", "label": "tempest-ServersTestJSON-766397782-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dfff08982dad4790bf4d555e2b4db5e4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f3a2eb5-353f-45c5-a73b-869626f4bb13", "external-id": "nsx-vlan-transportzone-411", "segmentation_id": 411, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9ad5243a-5d", "ovs_interfaceid": "9ad5243a-5d04-419d-bfd7-e4b27536ee65", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2127.545141] env[62824]: DEBUG oslo_vmware.api [None req-66e16106-e2f9-4347-968f-3ba76efdbbd5 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146379, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2127.565205] env[62824]: DEBUG nova.network.neutron [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 4c2c2068-664d-404f-a99e-8fc7719f43e5] Successfully updated port: 96a12f9d-8a1e-4df8-b5fa-24c8b92c413f {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2127.637024] env[62824]: DEBUG nova.scheduler.client.report [None req-123477fc-8c3e-4f2a-83de-cced4f011020 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2127.773131] env[62824]: DEBUG oslo_vmware.api [None req-9c371f91-b712-4ee9-8e58-6993dc153ed4 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146377, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2127.808294] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146376, 'name': CreateVM_Task, 'duration_secs': 0.624929} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2127.808466] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2127.809215] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2127.809385] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2127.809698] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2127.809946] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-12d1de2b-661c-45dc-a4d6-38fb38dff521 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2127.814598] env[62824]: DEBUG oslo_vmware.api [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2127.814598] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52aa0679-c861-d8bc-0ea8-fa42e2d73485" [ 2127.814598] env[62824]: _type = "Task" [ 2127.814598] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2127.822446] env[62824]: DEBUG oslo_vmware.api [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52aa0679-c861-d8bc-0ea8-fa42e2d73485, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2127.898742] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Task: {'id': task-2146378, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06965} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2127.899029] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 29140542-dc4b-411a-ac10-7d84086eabbb] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2127.899817] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ce263d2-d165-48fa-bb11-1e63f4dbce45 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2127.921664] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 29140542-dc4b-411a-ac10-7d84086eabbb] Reconfiguring VM instance instance-00000046 to attach disk [datastore2] 29140542-dc4b-411a-ac10-7d84086eabbb/29140542-dc4b-411a-ac10-7d84086eabbb.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2127.922298] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6cb6b303-182a-441d-befe-6dbf1c4c5f3b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2127.941073] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Waiting for the task: (returnval){ [ 2127.941073] env[62824]: value = "task-2146380" [ 2127.941073] env[62824]: _type = "Task" [ 2127.941073] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2127.949725] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Task: {'id': task-2146380, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2128.038632] env[62824]: DEBUG oslo_vmware.api [None req-66e16106-e2f9-4347-968f-3ba76efdbbd5 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146379, 'name': PowerOffVM_Task, 'duration_secs': 0.484071} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2128.038934] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-66e16106-e2f9-4347-968f-3ba76efdbbd5 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2128.039127] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-66e16106-e2f9-4347-968f-3ba76efdbbd5 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2128.039388] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0063c08e-41ee-40d1-8992-9a198e32351a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2128.042610] env[62824]: DEBUG oslo_concurrency.lockutils [req-1e414d60-f6ee-4ac8-9b04-267d54eeebc4 req-5c53f72c-2013-4334-aabd-eabdb7ab791e service nova] Releasing lock "refresh_cache-bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2128.067726] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Acquiring lock "refresh_cache-4c2c2068-664d-404f-a99e-8fc7719f43e5" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2128.067726] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Acquired lock "refresh_cache-4c2c2068-664d-404f-a99e-8fc7719f43e5" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2128.067894] env[62824]: DEBUG nova.network.neutron [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 4c2c2068-664d-404f-a99e-8fc7719f43e5] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2128.145973] env[62824]: DEBUG oslo_concurrency.lockutils [None req-123477fc-8c3e-4f2a-83de-cced4f011020 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.143s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2128.147699] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1254224f-0bf3-41dd-a390-a83d0850a18a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 34.875s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2128.147963] env[62824]: DEBUG nova.objects.instance [None req-1254224f-0bf3-41dd-a390-a83d0850a18a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Lazy-loading 'resources' on Instance uuid bbf1440b-8681-48c8-a178-9a83b925c695 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2128.190599] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-66e16106-e2f9-4347-968f-3ba76efdbbd5 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2128.190886] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-66e16106-e2f9-4347-968f-3ba76efdbbd5 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Deleting contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2128.191138] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-66e16106-e2f9-4347-968f-3ba76efdbbd5 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Deleting the datastore file [datastore2] 363177c0-dc40-429a-a74b-e690da133edb {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2128.192166] env[62824]: INFO nova.network.neutron [None req-123477fc-8c3e-4f2a-83de-cced4f011020 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Updating port 60a4fce0-0e63-45af-890c-46ca44ea9a0c with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 2128.194288] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1c63038e-ab48-4ff2-b261-4148988e6f74 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2128.202097] env[62824]: DEBUG oslo_vmware.api [None req-66e16106-e2f9-4347-968f-3ba76efdbbd5 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Waiting for the task: (returnval){ [ 2128.202097] env[62824]: value = "task-2146382" [ 2128.202097] env[62824]: _type = "Task" [ 2128.202097] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2128.210230] env[62824]: DEBUG oslo_vmware.api [None req-66e16106-e2f9-4347-968f-3ba76efdbbd5 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146382, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2128.274889] env[62824]: DEBUG oslo_vmware.api [None req-9c371f91-b712-4ee9-8e58-6993dc153ed4 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146377, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2128.325546] env[62824]: DEBUG oslo_vmware.api [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52aa0679-c861-d8bc-0ea8-fa42e2d73485, 'name': SearchDatastore_Task, 'duration_secs': 0.009765} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2128.325879] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2128.326175] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2128.326437] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2128.326613] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2128.326880] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2128.327226] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-183dd55c-dda6-4686-8e31-ff3b5c0e0614 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2128.343791] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2128.343931] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2128.344693] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d98a0266-c48c-4ff7-b7b8-73abc56b6762 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2128.350999] env[62824]: DEBUG oslo_vmware.api [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2128.350999] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52229920-e60f-74be-7093-3dffe79eac22" [ 2128.350999] env[62824]: _type = "Task" [ 2128.350999] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2128.358622] env[62824]: DEBUG oslo_vmware.api [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52229920-e60f-74be-7093-3dffe79eac22, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2128.452486] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Task: {'id': task-2146380, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2128.600835] env[62824]: DEBUG nova.network.neutron [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 4c2c2068-664d-404f-a99e-8fc7719f43e5] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2128.714127] env[62824]: DEBUG oslo_vmware.api [None req-66e16106-e2f9-4347-968f-3ba76efdbbd5 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146382, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.142303} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2128.716446] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-66e16106-e2f9-4347-968f-3ba76efdbbd5 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2128.716654] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-66e16106-e2f9-4347-968f-3ba76efdbbd5 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Deleted contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2128.716924] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-66e16106-e2f9-4347-968f-3ba76efdbbd5 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2128.717123] env[62824]: INFO nova.compute.manager [None req-66e16106-e2f9-4347-968f-3ba76efdbbd5 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Took 1.21 seconds to destroy the instance on the hypervisor. [ 2128.717382] env[62824]: DEBUG oslo.service.loopingcall [None req-66e16106-e2f9-4347-968f-3ba76efdbbd5 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2128.717774] env[62824]: DEBUG nova.compute.manager [-] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2128.717953] env[62824]: DEBUG nova.network.neutron [-] [instance: 363177c0-dc40-429a-a74b-e690da133edb] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2128.776634] env[62824]: DEBUG oslo_vmware.api [None req-9c371f91-b712-4ee9-8e58-6993dc153ed4 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146377, 'name': CreateSnapshot_Task, 'duration_secs': 1.066039} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2128.778056] env[62824]: DEBUG nova.network.neutron [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 4c2c2068-664d-404f-a99e-8fc7719f43e5] Updating instance_info_cache with network_info: [{"id": "96a12f9d-8a1e-4df8-b5fa-24c8b92c413f", "address": "fa:16:3e:90:f7:44", "network": {"id": "4ee2cade-6795-4887-a1df-f3990219e980", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1173612309-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "25bb8a4df0b740c0a1282ee61ad2f58c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3202e2b9-c4a7-4f78-9476-12ed92fabe61", "external-id": "nsx-vlan-transportzone-846", "segmentation_id": 846, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap96a12f9d-8a", "ovs_interfaceid": "96a12f9d-8a1e-4df8-b5fa-24c8b92c413f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2128.778905] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-9c371f91-b712-4ee9-8e58-6993dc153ed4 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 49818910-61de-4fb8-bbab-d5d61d2f1ada] Created Snapshot of the VM instance {{(pid=62824) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 2128.783468] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-240163fd-34b4-4471-946a-a3a6565e5a8f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2128.839069] env[62824]: DEBUG nova.compute.manager [req-d089dbce-dda4-4d23-8e25-ea050f0c7e7b req-6b51b19e-789a-41ea-a7eb-ad59ae1951ea service nova] [instance: 4c2c2068-664d-404f-a99e-8fc7719f43e5] Received event network-vif-plugged-96a12f9d-8a1e-4df8-b5fa-24c8b92c413f {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2128.841360] env[62824]: DEBUG oslo_concurrency.lockutils [req-d089dbce-dda4-4d23-8e25-ea050f0c7e7b req-6b51b19e-789a-41ea-a7eb-ad59ae1951ea service nova] Acquiring lock "4c2c2068-664d-404f-a99e-8fc7719f43e5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2128.841360] env[62824]: DEBUG oslo_concurrency.lockutils [req-d089dbce-dda4-4d23-8e25-ea050f0c7e7b req-6b51b19e-789a-41ea-a7eb-ad59ae1951ea service nova] Lock "4c2c2068-664d-404f-a99e-8fc7719f43e5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2128.841360] env[62824]: DEBUG oslo_concurrency.lockutils [req-d089dbce-dda4-4d23-8e25-ea050f0c7e7b req-6b51b19e-789a-41ea-a7eb-ad59ae1951ea service nova] Lock "4c2c2068-664d-404f-a99e-8fc7719f43e5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2128.841360] env[62824]: DEBUG nova.compute.manager [req-d089dbce-dda4-4d23-8e25-ea050f0c7e7b req-6b51b19e-789a-41ea-a7eb-ad59ae1951ea service nova] [instance: 4c2c2068-664d-404f-a99e-8fc7719f43e5] No waiting events found dispatching network-vif-plugged-96a12f9d-8a1e-4df8-b5fa-24c8b92c413f {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2128.841360] env[62824]: WARNING nova.compute.manager [req-d089dbce-dda4-4d23-8e25-ea050f0c7e7b req-6b51b19e-789a-41ea-a7eb-ad59ae1951ea service nova] [instance: 4c2c2068-664d-404f-a99e-8fc7719f43e5] Received unexpected event network-vif-plugged-96a12f9d-8a1e-4df8-b5fa-24c8b92c413f for instance with vm_state building and task_state spawning. [ 2128.842609] env[62824]: DEBUG nova.compute.manager [req-d089dbce-dda4-4d23-8e25-ea050f0c7e7b req-6b51b19e-789a-41ea-a7eb-ad59ae1951ea service nova] [instance: 4c2c2068-664d-404f-a99e-8fc7719f43e5] Received event network-changed-96a12f9d-8a1e-4df8-b5fa-24c8b92c413f {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2128.842806] env[62824]: DEBUG nova.compute.manager [req-d089dbce-dda4-4d23-8e25-ea050f0c7e7b req-6b51b19e-789a-41ea-a7eb-ad59ae1951ea service nova] [instance: 4c2c2068-664d-404f-a99e-8fc7719f43e5] Refreshing instance network info cache due to event network-changed-96a12f9d-8a1e-4df8-b5fa-24c8b92c413f. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2128.842982] env[62824]: DEBUG oslo_concurrency.lockutils [req-d089dbce-dda4-4d23-8e25-ea050f0c7e7b req-6b51b19e-789a-41ea-a7eb-ad59ae1951ea service nova] Acquiring lock "refresh_cache-4c2c2068-664d-404f-a99e-8fc7719f43e5" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2128.864385] env[62824]: DEBUG oslo_vmware.api [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52229920-e60f-74be-7093-3dffe79eac22, 'name': SearchDatastore_Task, 'duration_secs': 0.008103} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2128.865324] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-997c513c-f473-4949-989b-23c55d844ad2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2128.870856] env[62824]: DEBUG oslo_vmware.api [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2128.870856] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52077645-4aa2-0cb2-9687-fa31914f2985" [ 2128.870856] env[62824]: _type = "Task" [ 2128.870856] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2128.878633] env[62824]: DEBUG oslo_vmware.api [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52077645-4aa2-0cb2-9687-fa31914f2985, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2128.950527] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Task: {'id': task-2146380, 'name': ReconfigVM_Task, 'duration_secs': 0.775758} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2128.952938] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 29140542-dc4b-411a-ac10-7d84086eabbb] Reconfigured VM instance instance-00000046 to attach disk [datastore2] 29140542-dc4b-411a-ac10-7d84086eabbb/29140542-dc4b-411a-ac10-7d84086eabbb.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2128.954378] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3d05e9b5-b8b5-4a0e-ad12-4df9bcc82614 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2128.960256] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Waiting for the task: (returnval){ [ 2128.960256] env[62824]: value = "task-2146383" [ 2128.960256] env[62824]: _type = "Task" [ 2128.960256] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2128.972193] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Task: {'id': task-2146383, 'name': Rename_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2129.059937] env[62824]: DEBUG nova.compute.manager [req-96bacd0c-ce3e-42c3-8823-2e069d13f524 req-9f289380-4d01-412e-9dbd-8c7868d5a98a service nova] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Received event network-vif-deleted-8e8b16ef-e3f8-48c1-aab6-51d94d356c77 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2129.060152] env[62824]: INFO nova.compute.manager [req-96bacd0c-ce3e-42c3-8823-2e069d13f524 req-9f289380-4d01-412e-9dbd-8c7868d5a98a service nova] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Neutron deleted interface 8e8b16ef-e3f8-48c1-aab6-51d94d356c77; detaching it from the instance and deleting it from the info cache [ 2129.060322] env[62824]: DEBUG nova.network.neutron [req-96bacd0c-ce3e-42c3-8823-2e069d13f524 req-9f289380-4d01-412e-9dbd-8c7868d5a98a service nova] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2129.157665] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45376a34-01e4-4a44-a919-927886c4a57e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2129.165663] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-907fb854-0688-4e4b-9865-a0efed516059 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2129.199247] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-963837dc-78bd-4835-a39e-2bdf17b346c9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2129.207637] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dbbf881-2a3d-427c-ae8e-37655c6f8d18 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2129.221475] env[62824]: DEBUG nova.compute.provider_tree [None req-1254224f-0bf3-41dd-a390-a83d0850a18a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2129.283851] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Releasing lock "refresh_cache-4c2c2068-664d-404f-a99e-8fc7719f43e5" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2129.284214] env[62824]: DEBUG nova.compute.manager [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 4c2c2068-664d-404f-a99e-8fc7719f43e5] Instance network_info: |[{"id": "96a12f9d-8a1e-4df8-b5fa-24c8b92c413f", "address": "fa:16:3e:90:f7:44", "network": {"id": "4ee2cade-6795-4887-a1df-f3990219e980", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1173612309-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "25bb8a4df0b740c0a1282ee61ad2f58c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3202e2b9-c4a7-4f78-9476-12ed92fabe61", "external-id": "nsx-vlan-transportzone-846", "segmentation_id": 846, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap96a12f9d-8a", "ovs_interfaceid": "96a12f9d-8a1e-4df8-b5fa-24c8b92c413f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2129.284519] env[62824]: DEBUG oslo_concurrency.lockutils [req-d089dbce-dda4-4d23-8e25-ea050f0c7e7b req-6b51b19e-789a-41ea-a7eb-ad59ae1951ea service nova] Acquired lock "refresh_cache-4c2c2068-664d-404f-a99e-8fc7719f43e5" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2129.284699] env[62824]: DEBUG nova.network.neutron [req-d089dbce-dda4-4d23-8e25-ea050f0c7e7b req-6b51b19e-789a-41ea-a7eb-ad59ae1951ea service nova] [instance: 4c2c2068-664d-404f-a99e-8fc7719f43e5] Refreshing network info cache for port 96a12f9d-8a1e-4df8-b5fa-24c8b92c413f {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2129.285927] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 4c2c2068-664d-404f-a99e-8fc7719f43e5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:90:f7:44', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3202e2b9-c4a7-4f78-9476-12ed92fabe61', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '96a12f9d-8a1e-4df8-b5fa-24c8b92c413f', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2129.293847] env[62824]: DEBUG oslo.service.loopingcall [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2129.294823] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4c2c2068-664d-404f-a99e-8fc7719f43e5] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2129.295078] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6b241d22-d001-4a78-9ce6-e6e40ed75fbc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2129.315766] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-9c371f91-b712-4ee9-8e58-6993dc153ed4 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 49818910-61de-4fb8-bbab-d5d61d2f1ada] Creating linked-clone VM from snapshot {{(pid=62824) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 2129.316270] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-f1e2bdab-37e2-40f8-8ab5-d22c2c26f1f6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2129.325242] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2129.325242] env[62824]: value = "task-2146385" [ 2129.325242] env[62824]: _type = "Task" [ 2129.325242] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2129.326427] env[62824]: DEBUG oslo_vmware.api [None req-9c371f91-b712-4ee9-8e58-6993dc153ed4 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Waiting for the task: (returnval){ [ 2129.326427] env[62824]: value = "task-2146384" [ 2129.326427] env[62824]: _type = "Task" [ 2129.326427] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2129.339140] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146385, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2129.382207] env[62824]: DEBUG oslo_vmware.api [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52077645-4aa2-0cb2-9687-fa31914f2985, 'name': SearchDatastore_Task, 'duration_secs': 0.009153} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2129.382876] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2129.382996] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e/bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2129.383353] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-96e89061-0aca-4719-9dd2-ae69c0a8ea6b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2129.390646] env[62824]: DEBUG oslo_vmware.api [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2129.390646] env[62824]: value = "task-2146386" [ 2129.390646] env[62824]: _type = "Task" [ 2129.390646] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2129.402012] env[62824]: DEBUG oslo_vmware.api [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146386, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2129.469766] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Task: {'id': task-2146383, 'name': Rename_Task, 'duration_secs': 0.133531} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2129.470106] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 29140542-dc4b-411a-ac10-7d84086eabbb] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2129.471140] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6f08ae1e-75d7-4bde-8282-d9ddda1150ea {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2129.477633] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Waiting for the task: (returnval){ [ 2129.477633] env[62824]: value = "task-2146387" [ 2129.477633] env[62824]: _type = "Task" [ 2129.477633] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2129.486120] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Task: {'id': task-2146387, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2129.505763] env[62824]: DEBUG nova.network.neutron [-] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2129.566308] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fee7c4be-1efc-4211-9d52-fe4ef9953c1f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2129.578402] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dded5562-c24b-44c2-8e90-2c5c905ff21c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2129.617716] env[62824]: DEBUG nova.compute.manager [req-96bacd0c-ce3e-42c3-8823-2e069d13f524 req-9f289380-4d01-412e-9dbd-8c7868d5a98a service nova] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Detach interface failed, port_id=8e8b16ef-e3f8-48c1-aab6-51d94d356c77, reason: Instance 363177c0-dc40-429a-a74b-e690da133edb could not be found. {{(pid=62824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11482}} [ 2129.725096] env[62824]: DEBUG nova.scheduler.client.report [None req-1254224f-0bf3-41dd-a390-a83d0850a18a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2129.841697] env[62824]: DEBUG oslo_vmware.api [None req-9c371f91-b712-4ee9-8e58-6993dc153ed4 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146384, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2129.847748] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146385, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2129.900573] env[62824]: DEBUG oslo_vmware.api [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146386, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.467509} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2129.900863] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e/bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2129.901111] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2129.901409] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7b5d0a6b-c8ae-4507-81bd-ee0be9428e88 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2129.909872] env[62824]: DEBUG oslo_vmware.api [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2129.909872] env[62824]: value = "task-2146388" [ 2129.909872] env[62824]: _type = "Task" [ 2129.909872] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2129.918888] env[62824]: DEBUG oslo_vmware.api [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146388, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2129.990481] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Task: {'id': task-2146387, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2130.009154] env[62824]: INFO nova.compute.manager [-] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Took 1.29 seconds to deallocate network for instance. [ 2130.032018] env[62824]: DEBUG nova.network.neutron [req-d089dbce-dda4-4d23-8e25-ea050f0c7e7b req-6b51b19e-789a-41ea-a7eb-ad59ae1951ea service nova] [instance: 4c2c2068-664d-404f-a99e-8fc7719f43e5] Updated VIF entry in instance network info cache for port 96a12f9d-8a1e-4df8-b5fa-24c8b92c413f. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2130.032427] env[62824]: DEBUG nova.network.neutron [req-d089dbce-dda4-4d23-8e25-ea050f0c7e7b req-6b51b19e-789a-41ea-a7eb-ad59ae1951ea service nova] [instance: 4c2c2068-664d-404f-a99e-8fc7719f43e5] Updating instance_info_cache with network_info: [{"id": "96a12f9d-8a1e-4df8-b5fa-24c8b92c413f", "address": "fa:16:3e:90:f7:44", "network": {"id": "4ee2cade-6795-4887-a1df-f3990219e980", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1173612309-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "25bb8a4df0b740c0a1282ee61ad2f58c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3202e2b9-c4a7-4f78-9476-12ed92fabe61", "external-id": "nsx-vlan-transportzone-846", "segmentation_id": 846, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap96a12f9d-8a", "ovs_interfaceid": "96a12f9d-8a1e-4df8-b5fa-24c8b92c413f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2130.230284] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1254224f-0bf3-41dd-a390-a83d0850a18a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.082s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2130.232616] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b5cb0b06-e659-44b5-a4c3-505f639c37f7 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.019s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2130.232990] env[62824]: DEBUG nova.objects.instance [None req-b5cb0b06-e659-44b5-a4c3-505f639c37f7 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lazy-loading 'resources' on Instance uuid 919e8a45-7810-4a8d-a1aa-5046b5ab059c {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2130.253320] env[62824]: INFO nova.scheduler.client.report [None req-1254224f-0bf3-41dd-a390-a83d0850a18a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Deleted allocations for instance bbf1440b-8681-48c8-a178-9a83b925c695 [ 2130.341640] env[62824]: DEBUG oslo_vmware.api [None req-9c371f91-b712-4ee9-8e58-6993dc153ed4 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146384, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2130.344732] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146385, 'name': CreateVM_Task, 'duration_secs': 0.757285} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2130.344916] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4c2c2068-664d-404f-a99e-8fc7719f43e5] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2130.345587] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2130.345747] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2130.346074] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2130.346317] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-caf7a977-261d-4dd2-86dc-6ee79653e59c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2130.350987] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Waiting for the task: (returnval){ [ 2130.350987] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]521571a6-c05c-0b2b-0ec0-1225eede3e44" [ 2130.350987] env[62824]: _type = "Task" [ 2130.350987] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2130.359434] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]521571a6-c05c-0b2b-0ec0-1225eede3e44, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2130.419439] env[62824]: DEBUG oslo_vmware.api [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146388, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074914} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2130.419589] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2130.420369] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da5fd5ba-e956-4d32-94b7-3918bf08df9e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2130.441502] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e] Reconfiguring VM instance instance-00000049 to attach disk [datastore2] bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e/bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2130.441766] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-81367be8-8409-462e-8565-74564d297f03 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2130.462159] env[62824]: DEBUG oslo_vmware.api [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2130.462159] env[62824]: value = "task-2146389" [ 2130.462159] env[62824]: _type = "Task" [ 2130.462159] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2130.469808] env[62824]: DEBUG oslo_vmware.api [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146389, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2130.486455] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Task: {'id': task-2146387, 'name': PowerOnVM_Task, 'duration_secs': 0.58902} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2130.486742] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 29140542-dc4b-411a-ac10-7d84086eabbb] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2130.486984] env[62824]: INFO nova.compute.manager [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 29140542-dc4b-411a-ac10-7d84086eabbb] Took 8.68 seconds to spawn the instance on the hypervisor. [ 2130.487191] env[62824]: DEBUG nova.compute.manager [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 29140542-dc4b-411a-ac10-7d84086eabbb] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2130.487980] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5533093c-83ad-450a-b9d5-944be40e5036 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2130.519816] env[62824]: DEBUG oslo_concurrency.lockutils [None req-66e16106-e2f9-4347-968f-3ba76efdbbd5 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2130.535710] env[62824]: DEBUG oslo_concurrency.lockutils [req-d089dbce-dda4-4d23-8e25-ea050f0c7e7b req-6b51b19e-789a-41ea-a7eb-ad59ae1951ea service nova] Releasing lock "refresh_cache-4c2c2068-664d-404f-a99e-8fc7719f43e5" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2130.762730] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1254224f-0bf3-41dd-a390-a83d0850a18a tempest-ListImageFiltersTestJSON-208399242 tempest-ListImageFiltersTestJSON-208399242-project-member] Lock "bbf1440b-8681-48c8-a178-9a83b925c695" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 42.037s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2130.843169] env[62824]: DEBUG oslo_vmware.api [None req-9c371f91-b712-4ee9-8e58-6993dc153ed4 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146384, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2130.861968] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]521571a6-c05c-0b2b-0ec0-1225eede3e44, 'name': SearchDatastore_Task, 'duration_secs': 0.009859} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2130.862349] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2130.862542] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 4c2c2068-664d-404f-a99e-8fc7719f43e5] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2130.862764] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2130.862926] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2130.863125] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2130.863393] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c64f0490-e655-43c4-a009-0b9c63fe156e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2130.871567] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2130.871758] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2130.874858] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-954fb8c8-df30-457d-aab7-b4eadda23cc5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2130.879928] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Waiting for the task: (returnval){ [ 2130.879928] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52e56339-ee76-f2cf-fe32-6f50cb6c1999" [ 2130.879928] env[62824]: _type = "Task" [ 2130.879928] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2130.887768] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52e56339-ee76-f2cf-fe32-6f50cb6c1999, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2130.974182] env[62824]: DEBUG oslo_vmware.api [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146389, 'name': ReconfigVM_Task, 'duration_secs': 0.309179} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2130.974513] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e] Reconfigured VM instance instance-00000049 to attach disk [datastore2] bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e/bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2130.975584] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5c982c24-0c43-4a9f-a73e-401c3ef1b69a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2130.982242] env[62824]: DEBUG oslo_vmware.api [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2130.982242] env[62824]: value = "task-2146390" [ 2130.982242] env[62824]: _type = "Task" [ 2130.982242] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2130.994477] env[62824]: DEBUG oslo_vmware.api [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146390, 'name': Rename_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2131.009676] env[62824]: INFO nova.compute.manager [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 29140542-dc4b-411a-ac10-7d84086eabbb] Took 39.34 seconds to build instance. [ 2131.211911] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4c2e867-fe0a-4eb7-af8f-ffc89dcc217e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2131.220482] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66a80bae-0fde-49ab-8a14-65a1951409d8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2131.253127] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74bd6202-7347-4319-89e9-3a5e1c55c2b6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2131.261081] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdaab91e-6958-468b-832d-8b2fecdfae89 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2131.276023] env[62824]: DEBUG nova.compute.provider_tree [None req-b5cb0b06-e659-44b5-a4c3-505f639c37f7 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2131.344445] env[62824]: DEBUG oslo_vmware.api [None req-9c371f91-b712-4ee9-8e58-6993dc153ed4 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146384, 'name': CloneVM_Task, 'duration_secs': 1.949642} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2131.344718] env[62824]: INFO nova.virt.vmwareapi.vmops [None req-9c371f91-b712-4ee9-8e58-6993dc153ed4 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 49818910-61de-4fb8-bbab-d5d61d2f1ada] Created linked-clone VM from snapshot [ 2131.345476] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ee8665d-89b0-45cf-ad33-6842b058d5e5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2131.353548] env[62824]: DEBUG nova.virt.vmwareapi.images [None req-9c371f91-b712-4ee9-8e58-6993dc153ed4 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 49818910-61de-4fb8-bbab-d5d61d2f1ada] Uploading image 0840fcae-3ca2-4ec3-846b-617aedfb9fd7 {{(pid=62824) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 2131.364907] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c371f91-b712-4ee9-8e58-6993dc153ed4 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 49818910-61de-4fb8-bbab-d5d61d2f1ada] Destroying the VM {{(pid=62824) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 2131.365221] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-69515311-983b-4494-acc1-0383a1c26e2a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2131.372793] env[62824]: DEBUG oslo_vmware.api [None req-9c371f91-b712-4ee9-8e58-6993dc153ed4 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Waiting for the task: (returnval){ [ 2131.372793] env[62824]: value = "task-2146391" [ 2131.372793] env[62824]: _type = "Task" [ 2131.372793] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2131.381320] env[62824]: DEBUG oslo_vmware.api [None req-9c371f91-b712-4ee9-8e58-6993dc153ed4 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146391, 'name': Destroy_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2131.391608] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52e56339-ee76-f2cf-fe32-6f50cb6c1999, 'name': SearchDatastore_Task, 'duration_secs': 0.008786} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2131.392413] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-44eddcc9-c5a1-4011-81e2-67c5fd8d55c0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2131.397547] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Waiting for the task: (returnval){ [ 2131.397547] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52f8473b-75be-514c-f8b6-e1ab174693fb" [ 2131.397547] env[62824]: _type = "Task" [ 2131.397547] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2131.405356] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52f8473b-75be-514c-f8b6-e1ab174693fb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2131.494589] env[62824]: DEBUG oslo_vmware.api [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146390, 'name': Rename_Task, 'duration_secs': 0.136284} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2131.494985] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2131.495300] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5850e0d0-e3e4-4b73-9d3b-ac3b7112ab0b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2131.502693] env[62824]: DEBUG oslo_vmware.api [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2131.502693] env[62824]: value = "task-2146392" [ 2131.502693] env[62824]: _type = "Task" [ 2131.502693] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2131.512204] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Lock "29140542-dc4b-411a-ac10-7d84086eabbb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 40.866s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2131.512416] env[62824]: DEBUG oslo_vmware.api [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146392, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2131.779403] env[62824]: DEBUG nova.scheduler.client.report [None req-b5cb0b06-e659-44b5-a4c3-505f639c37f7 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2131.883754] env[62824]: DEBUG oslo_vmware.api [None req-9c371f91-b712-4ee9-8e58-6993dc153ed4 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146391, 'name': Destroy_Task, 'duration_secs': 0.463232} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2131.883999] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-9c371f91-b712-4ee9-8e58-6993dc153ed4 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 49818910-61de-4fb8-bbab-d5d61d2f1ada] Destroyed the VM [ 2131.884254] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-9c371f91-b712-4ee9-8e58-6993dc153ed4 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 49818910-61de-4fb8-bbab-d5d61d2f1ada] Deleting Snapshot of the VM instance {{(pid=62824) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 2131.884515] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-e5077332-faab-4a74-9ddc-daa97732508f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2131.891380] env[62824]: DEBUG oslo_vmware.api [None req-9c371f91-b712-4ee9-8e58-6993dc153ed4 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Waiting for the task: (returnval){ [ 2131.891380] env[62824]: value = "task-2146393" [ 2131.891380] env[62824]: _type = "Task" [ 2131.891380] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2131.899537] env[62824]: DEBUG oslo_vmware.api [None req-9c371f91-b712-4ee9-8e58-6993dc153ed4 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146393, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2131.908891] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52f8473b-75be-514c-f8b6-e1ab174693fb, 'name': SearchDatastore_Task, 'duration_secs': 0.00825} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2131.909395] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2131.909509] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 4c2c2068-664d-404f-a99e-8fc7719f43e5/4c2c2068-664d-404f-a99e-8fc7719f43e5.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2131.909747] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0e9c7011-a0dd-46f1-99c9-91357210bedd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2131.917316] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Waiting for the task: (returnval){ [ 2131.917316] env[62824]: value = "task-2146394" [ 2131.917316] env[62824]: _type = "Task" [ 2131.917316] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2131.925313] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Task: {'id': task-2146394, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2132.013222] env[62824]: DEBUG oslo_vmware.api [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146392, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2132.014919] env[62824]: DEBUG nova.compute.manager [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] [instance: ab37b1a2-0012-40fb-9341-b613525e89cf] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2132.286925] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b5cb0b06-e659-44b5-a4c3-505f639c37f7 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.054s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2132.289693] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 33.004s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2132.291925] env[62824]: INFO nova.compute.claims [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: d067fa97-cedc-4e3d-9be4-d860a79a7723] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2132.317267] env[62824]: INFO nova.scheduler.client.report [None req-b5cb0b06-e659-44b5-a4c3-505f639c37f7 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Deleted allocations for instance 919e8a45-7810-4a8d-a1aa-5046b5ab059c [ 2132.403647] env[62824]: DEBUG oslo_vmware.api [None req-9c371f91-b712-4ee9-8e58-6993dc153ed4 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146393, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2132.427899] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Task: {'id': task-2146394, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2132.512296] env[62824]: DEBUG oslo_vmware.api [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146392, 'name': PowerOnVM_Task, 'duration_secs': 0.644114} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2132.512563] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2132.512766] env[62824]: INFO nova.compute.manager [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e] Took 8.04 seconds to spawn the instance on the hypervisor. [ 2132.512943] env[62824]: DEBUG nova.compute.manager [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2132.513742] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-864227a7-9044-46a6-bfbb-6dc7e5b42f9d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2132.536828] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2132.827558] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b5cb0b06-e659-44b5-a4c3-505f639c37f7 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "919e8a45-7810-4a8d-a1aa-5046b5ab059c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 39.632s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2132.904235] env[62824]: DEBUG oslo_vmware.api [None req-9c371f91-b712-4ee9-8e58-6993dc153ed4 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146393, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2132.927706] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Task: {'id': task-2146394, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.62764} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2132.928032] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 4c2c2068-664d-404f-a99e-8fc7719f43e5/4c2c2068-664d-404f-a99e-8fc7719f43e5.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2132.928272] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 4c2c2068-664d-404f-a99e-8fc7719f43e5] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2132.928552] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c9e299dd-5da2-4483-ab71-9d53cf2aa9d8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2132.935319] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Waiting for the task: (returnval){ [ 2132.935319] env[62824]: value = "task-2146395" [ 2132.935319] env[62824]: _type = "Task" [ 2132.935319] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2132.943561] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Task: {'id': task-2146395, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2133.032125] env[62824]: INFO nova.compute.manager [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e] Took 41.36 seconds to build instance. [ 2133.402311] env[62824]: DEBUG oslo_vmware.api [None req-9c371f91-b712-4ee9-8e58-6993dc153ed4 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146393, 'name': RemoveSnapshot_Task, 'duration_secs': 1.276992} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2133.404967] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-9c371f91-b712-4ee9-8e58-6993dc153ed4 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 49818910-61de-4fb8-bbab-d5d61d2f1ada] Deleted Snapshot of the VM instance {{(pid=62824) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 2133.449016] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Task: {'id': task-2146395, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.179036} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2133.449505] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 4c2c2068-664d-404f-a99e-8fc7719f43e5] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2133.451285] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09c9f0e8-5c7b-4200-994f-054112a0c161 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2133.476724] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 4c2c2068-664d-404f-a99e-8fc7719f43e5] Reconfiguring VM instance instance-00000047 to attach disk [datastore2] 4c2c2068-664d-404f-a99e-8fc7719f43e5/4c2c2068-664d-404f-a99e-8fc7719f43e5.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2133.479888] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8ab670f5-507e-4f28-835e-68cc67e09ae6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2133.502298] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Waiting for the task: (returnval){ [ 2133.502298] env[62824]: value = "task-2146396" [ 2133.502298] env[62824]: _type = "Task" [ 2133.502298] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2133.514013] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Task: {'id': task-2146396, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2133.534883] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7d8673b1-7f38-4aab-ba0f-9175b6e26593 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.889s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2133.755921] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db784cbb-ecdc-461d-8f6f-da37ef01643e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2133.766996] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ff9ba37-8a15-471f-a2a5-6d08c02bf528 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2133.798495] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22b36510-df8c-414f-9600-43333ea4076e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2133.805677] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9005355f-5e9e-4ddf-9312-ee4fb5fbcb7e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2133.819308] env[62824]: DEBUG nova.compute.provider_tree [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2133.909498] env[62824]: WARNING nova.compute.manager [None req-9c371f91-b712-4ee9-8e58-6993dc153ed4 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 49818910-61de-4fb8-bbab-d5d61d2f1ada] Image not found during snapshot: nova.exception.ImageNotFound: Image 0840fcae-3ca2-4ec3-846b-617aedfb9fd7 could not be found. [ 2134.014318] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Task: {'id': task-2146396, 'name': ReconfigVM_Task, 'duration_secs': 0.270598} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2134.015332] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 4c2c2068-664d-404f-a99e-8fc7719f43e5] Reconfigured VM instance instance-00000047 to attach disk [datastore2] 4c2c2068-664d-404f-a99e-8fc7719f43e5/4c2c2068-664d-404f-a99e-8fc7719f43e5.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2134.015617] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e517c83c-5cd2-49d5-8fc3-29c797fb0271 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2134.022670] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Waiting for the task: (returnval){ [ 2134.022670] env[62824]: value = "task-2146397" [ 2134.022670] env[62824]: _type = "Task" [ 2134.022670] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2134.030789] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Task: {'id': task-2146397, 'name': Rename_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2134.033243] env[62824]: DEBUG oslo_concurrency.lockutils [None req-61bb4e80-84d9-4e19-b9c0-195ff35ecb41 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquiring lock "bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2134.033481] env[62824]: DEBUG oslo_concurrency.lockutils [None req-61bb4e80-84d9-4e19-b9c0-195ff35ecb41 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2134.033660] env[62824]: DEBUG oslo_concurrency.lockutils [None req-61bb4e80-84d9-4e19-b9c0-195ff35ecb41 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquiring lock "bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2134.033997] env[62824]: DEBUG oslo_concurrency.lockutils [None req-61bb4e80-84d9-4e19-b9c0-195ff35ecb41 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2134.033997] env[62824]: DEBUG oslo_concurrency.lockutils [None req-61bb4e80-84d9-4e19-b9c0-195ff35ecb41 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2134.036090] env[62824]: INFO nova.compute.manager [None req-61bb4e80-84d9-4e19-b9c0-195ff35ecb41 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e] Terminating instance [ 2134.037657] env[62824]: DEBUG nova.compute.manager [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: da1ad739-b252-4e29-a22a-ab3bcab173ec] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2134.322888] env[62824]: DEBUG nova.scheduler.client.report [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2134.534018] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Task: {'id': task-2146397, 'name': Rename_Task, 'duration_secs': 0.127979} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2134.534524] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 4c2c2068-664d-404f-a99e-8fc7719f43e5] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2134.534898] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-49f62dee-07bd-4ba8-92d0-b176fb5f544d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2134.542029] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Waiting for the task: (returnval){ [ 2134.542029] env[62824]: value = "task-2146398" [ 2134.542029] env[62824]: _type = "Task" [ 2134.542029] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2134.548712] env[62824]: DEBUG nova.compute.manager [None req-61bb4e80-84d9-4e19-b9c0-195ff35ecb41 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2134.548908] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-61bb4e80-84d9-4e19-b9c0-195ff35ecb41 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2134.549219] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b30d3e55-ba09-4502-8b58-e6c2d9aaba20 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Acquiring lock "49818910-61de-4fb8-bbab-d5d61d2f1ada" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2134.549430] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b30d3e55-ba09-4502-8b58-e6c2d9aaba20 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Lock "49818910-61de-4fb8-bbab-d5d61d2f1ada" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2134.549611] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b30d3e55-ba09-4502-8b58-e6c2d9aaba20 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Acquiring lock "49818910-61de-4fb8-bbab-d5d61d2f1ada-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2134.549788] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b30d3e55-ba09-4502-8b58-e6c2d9aaba20 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Lock "49818910-61de-4fb8-bbab-d5d61d2f1ada-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2134.549951] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b30d3e55-ba09-4502-8b58-e6c2d9aaba20 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Lock "49818910-61de-4fb8-bbab-d5d61d2f1ada-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2134.553881] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa8787f5-7598-43e7-ba7b-c8ac08dd99da {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2134.557474] env[62824]: INFO nova.compute.manager [None req-b30d3e55-ba09-4502-8b58-e6c2d9aaba20 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 49818910-61de-4fb8-bbab-d5d61d2f1ada] Terminating instance [ 2134.563391] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Task: {'id': task-2146398, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2134.566747] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-61bb4e80-84d9-4e19-b9c0-195ff35ecb41 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2134.567245] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5859fd46-e3b6-44ec-b395-3ca2d477974f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2134.570619] env[62824]: DEBUG oslo_concurrency.lockutils [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2134.573937] env[62824]: DEBUG oslo_vmware.api [None req-61bb4e80-84d9-4e19-b9c0-195ff35ecb41 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2134.573937] env[62824]: value = "task-2146399" [ 2134.573937] env[62824]: _type = "Task" [ 2134.573937] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2134.585019] env[62824]: DEBUG oslo_vmware.api [None req-61bb4e80-84d9-4e19-b9c0-195ff35ecb41 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146399, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2134.829191] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.539s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2134.829794] env[62824]: DEBUG nova.compute.manager [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: d067fa97-cedc-4e3d-9be4-d860a79a7723] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2134.833008] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 32.128s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2134.835340] env[62824]: INFO nova.compute.claims [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] [instance: dadf590a-7288-4dd2-90de-125fe272f4ba] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2135.051990] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Task: {'id': task-2146398, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2135.067889] env[62824]: DEBUG nova.compute.manager [None req-b30d3e55-ba09-4502-8b58-e6c2d9aaba20 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 49818910-61de-4fb8-bbab-d5d61d2f1ada] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2135.068169] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-b30d3e55-ba09-4502-8b58-e6c2d9aaba20 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 49818910-61de-4fb8-bbab-d5d61d2f1ada] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2135.069048] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03f05f4c-a26d-47b8-a906-8fc171928e52 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2135.077776] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-b30d3e55-ba09-4502-8b58-e6c2d9aaba20 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 49818910-61de-4fb8-bbab-d5d61d2f1ada] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2135.080730] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ba356a94-9bf9-493e-ac99-2b5bfcac207f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2135.087218] env[62824]: DEBUG oslo_vmware.api [None req-61bb4e80-84d9-4e19-b9c0-195ff35ecb41 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146399, 'name': PowerOffVM_Task, 'duration_secs': 0.430126} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2135.088506] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-61bb4e80-84d9-4e19-b9c0-195ff35ecb41 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2135.088685] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-61bb4e80-84d9-4e19-b9c0-195ff35ecb41 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2135.088978] env[62824]: DEBUG oslo_vmware.api [None req-b30d3e55-ba09-4502-8b58-e6c2d9aaba20 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Waiting for the task: (returnval){ [ 2135.088978] env[62824]: value = "task-2146400" [ 2135.088978] env[62824]: _type = "Task" [ 2135.088978] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2135.089177] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a1a030cd-9bb8-411f-bb9a-686a28656f5b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2135.098835] env[62824]: DEBUG oslo_vmware.api [None req-b30d3e55-ba09-4502-8b58-e6c2d9aaba20 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146400, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2135.222457] env[62824]: DEBUG oslo_concurrency.lockutils [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquiring lock "58243465-ba24-4b37-a640-04ab24963230" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2135.222457] env[62824]: DEBUG oslo_concurrency.lockutils [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "58243465-ba24-4b37-a640-04ab24963230" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2135.311901] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-61bb4e80-84d9-4e19-b9c0-195ff35ecb41 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2135.312139] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-61bb4e80-84d9-4e19-b9c0-195ff35ecb41 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e] Deleting contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2135.312320] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-61bb4e80-84d9-4e19-b9c0-195ff35ecb41 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Deleting the datastore file [datastore2] bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2135.312592] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e63fe075-c1a5-4a4f-9135-c751992e59b5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2135.319394] env[62824]: DEBUG oslo_vmware.api [None req-61bb4e80-84d9-4e19-b9c0-195ff35ecb41 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2135.319394] env[62824]: value = "task-2146402" [ 2135.319394] env[62824]: _type = "Task" [ 2135.319394] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2135.327407] env[62824]: DEBUG oslo_vmware.api [None req-61bb4e80-84d9-4e19-b9c0-195ff35ecb41 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146402, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2135.340838] env[62824]: DEBUG nova.compute.utils [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2135.345881] env[62824]: DEBUG nova.compute.manager [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: d067fa97-cedc-4e3d-9be4-d860a79a7723] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2135.345881] env[62824]: DEBUG nova.network.neutron [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: d067fa97-cedc-4e3d-9be4-d860a79a7723] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2135.392663] env[62824]: DEBUG nova.policy [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8bb15b0904de4f50b9d5f0ec23c4f7d4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '25bb8a4df0b740c0a1282ee61ad2f58c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2135.552817] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Task: {'id': task-2146398, 'name': PowerOnVM_Task, 'duration_secs': 0.803865} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2135.553189] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 4c2c2068-664d-404f-a99e-8fc7719f43e5] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2135.553415] env[62824]: INFO nova.compute.manager [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 4c2c2068-664d-404f-a99e-8fc7719f43e5] Took 8.53 seconds to spawn the instance on the hypervisor. [ 2135.553596] env[62824]: DEBUG nova.compute.manager [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 4c2c2068-664d-404f-a99e-8fc7719f43e5] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2135.554428] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1993b24-2c44-461a-94a9-4b73f01271d9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2135.602109] env[62824]: DEBUG oslo_vmware.api [None req-b30d3e55-ba09-4502-8b58-e6c2d9aaba20 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146400, 'name': PowerOffVM_Task, 'duration_secs': 0.219215} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2135.602392] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-b30d3e55-ba09-4502-8b58-e6c2d9aaba20 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 49818910-61de-4fb8-bbab-d5d61d2f1ada] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2135.602561] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-b30d3e55-ba09-4502-8b58-e6c2d9aaba20 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 49818910-61de-4fb8-bbab-d5d61d2f1ada] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2135.602811] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-efc5b5d6-31f9-4b7a-9052-db24f201bb6f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2135.640603] env[62824]: DEBUG nova.network.neutron [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: d067fa97-cedc-4e3d-9be4-d860a79a7723] Successfully created port: 2b52b217-d85d-4239-a02a-d97e3c82d93d {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2135.686866] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-b30d3e55-ba09-4502-8b58-e6c2d9aaba20 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 49818910-61de-4fb8-bbab-d5d61d2f1ada] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2135.687123] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-b30d3e55-ba09-4502-8b58-e6c2d9aaba20 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 49818910-61de-4fb8-bbab-d5d61d2f1ada] Deleting contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2135.687333] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-b30d3e55-ba09-4502-8b58-e6c2d9aaba20 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Deleting the datastore file [datastore2] 49818910-61de-4fb8-bbab-d5d61d2f1ada {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2135.687595] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1dcb0748-8410-46f4-9a5a-0560fc105869 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2135.696201] env[62824]: DEBUG oslo_vmware.api [None req-b30d3e55-ba09-4502-8b58-e6c2d9aaba20 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Waiting for the task: (returnval){ [ 2135.696201] env[62824]: value = "task-2146404" [ 2135.696201] env[62824]: _type = "Task" [ 2135.696201] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2135.703436] env[62824]: DEBUG oslo_vmware.api [None req-b30d3e55-ba09-4502-8b58-e6c2d9aaba20 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146404, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2135.829684] env[62824]: DEBUG oslo_vmware.api [None req-61bb4e80-84d9-4e19-b9c0-195ff35ecb41 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146402, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.143859} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2135.829954] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-61bb4e80-84d9-4e19-b9c0-195ff35ecb41 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2135.830166] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-61bb4e80-84d9-4e19-b9c0-195ff35ecb41 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e] Deleted contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2135.830352] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-61bb4e80-84d9-4e19-b9c0-195ff35ecb41 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2135.830525] env[62824]: INFO nova.compute.manager [None req-61bb4e80-84d9-4e19-b9c0-195ff35ecb41 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e] Took 1.28 seconds to destroy the instance on the hypervisor. [ 2135.830766] env[62824]: DEBUG oslo.service.loopingcall [None req-61bb4e80-84d9-4e19-b9c0-195ff35ecb41 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2135.830966] env[62824]: DEBUG nova.compute.manager [-] [instance: bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2135.831062] env[62824]: DEBUG nova.network.neutron [-] [instance: bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2135.845537] env[62824]: DEBUG nova.compute.manager [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: d067fa97-cedc-4e3d-9be4-d860a79a7723] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2136.076962] env[62824]: INFO nova.compute.manager [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 4c2c2068-664d-404f-a99e-8fc7719f43e5] Took 44.40 seconds to build instance. [ 2136.209265] env[62824]: DEBUG oslo_vmware.api [None req-b30d3e55-ba09-4502-8b58-e6c2d9aaba20 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146404, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.146367} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2136.209792] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-b30d3e55-ba09-4502-8b58-e6c2d9aaba20 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2136.209792] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-b30d3e55-ba09-4502-8b58-e6c2d9aaba20 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 49818910-61de-4fb8-bbab-d5d61d2f1ada] Deleted contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2136.209917] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-b30d3e55-ba09-4502-8b58-e6c2d9aaba20 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 49818910-61de-4fb8-bbab-d5d61d2f1ada] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2136.210106] env[62824]: INFO nova.compute.manager [None req-b30d3e55-ba09-4502-8b58-e6c2d9aaba20 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 49818910-61de-4fb8-bbab-d5d61d2f1ada] Took 1.14 seconds to destroy the instance on the hypervisor. [ 2136.210937] env[62824]: DEBUG oslo.service.loopingcall [None req-b30d3e55-ba09-4502-8b58-e6c2d9aaba20 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2136.212956] env[62824]: DEBUG nova.compute.manager [-] [instance: 49818910-61de-4fb8-bbab-d5d61d2f1ada] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2136.213138] env[62824]: DEBUG nova.network.neutron [-] [instance: 49818910-61de-4fb8-bbab-d5d61d2f1ada] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2136.379908] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51706e6a-6ba4-499c-a76e-5ceb8601bfb8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2136.392875] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22d5653b-aaa3-4995-8501-b63398ab1e02 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2136.433207] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b738c18-1fdd-4933-b290-2160ddfb2b3b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2136.443258] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5871fa7b-27ef-44ca-9b87-f5756389466e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2136.458466] env[62824]: DEBUG nova.compute.provider_tree [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2136.570973] env[62824]: DEBUG nova.compute.manager [req-9d75c889-405b-4b3e-8bef-cf953916b619 req-0d31ba67-6e37-4cdf-a67a-b39b6828207a service nova] [instance: bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e] Received event network-vif-deleted-9ad5243a-5d04-419d-bfd7-e4b27536ee65 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2136.571199] env[62824]: INFO nova.compute.manager [req-9d75c889-405b-4b3e-8bef-cf953916b619 req-0d31ba67-6e37-4cdf-a67a-b39b6828207a service nova] [instance: bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e] Neutron deleted interface 9ad5243a-5d04-419d-bfd7-e4b27536ee65; detaching it from the instance and deleting it from the info cache [ 2136.571373] env[62824]: DEBUG nova.network.neutron [req-9d75c889-405b-4b3e-8bef-cf953916b619 req-0d31ba67-6e37-4cdf-a67a-b39b6828207a service nova] [instance: bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2136.579472] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Lock "4c2c2068-664d-404f-a99e-8fc7719f43e5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 45.930s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2136.699917] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d39ad21e-d65d-4fc4-9829-53a13e0b4169 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Acquiring lock "59b5b883-4188-471c-8862-444f3ce08cb0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2136.699917] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d39ad21e-d65d-4fc4-9829-53a13e0b4169 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Lock "59b5b883-4188-471c-8862-444f3ce08cb0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2136.699917] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d39ad21e-d65d-4fc4-9829-53a13e0b4169 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Acquiring lock "59b5b883-4188-471c-8862-444f3ce08cb0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2136.699917] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d39ad21e-d65d-4fc4-9829-53a13e0b4169 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Lock "59b5b883-4188-471c-8862-444f3ce08cb0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2136.699917] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d39ad21e-d65d-4fc4-9829-53a13e0b4169 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Lock "59b5b883-4188-471c-8862-444f3ce08cb0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2136.701581] env[62824]: INFO nova.compute.manager [None req-d39ad21e-d65d-4fc4-9829-53a13e0b4169 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Terminating instance [ 2136.857399] env[62824]: DEBUG nova.compute.manager [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: d067fa97-cedc-4e3d-9be4-d860a79a7723] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2136.883454] env[62824]: DEBUG nova.virt.hardware [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2136.883708] env[62824]: DEBUG nova.virt.hardware [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2136.883864] env[62824]: DEBUG nova.virt.hardware [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2136.884059] env[62824]: DEBUG nova.virt.hardware [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2136.884213] env[62824]: DEBUG nova.virt.hardware [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2136.884362] env[62824]: DEBUG nova.virt.hardware [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2136.884571] env[62824]: DEBUG nova.virt.hardware [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2136.884729] env[62824]: DEBUG nova.virt.hardware [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2136.884896] env[62824]: DEBUG nova.virt.hardware [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2136.885073] env[62824]: DEBUG nova.virt.hardware [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2136.885259] env[62824]: DEBUG nova.virt.hardware [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2136.886162] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a02b597-3dde-4b4b-9d12-838fc2d9a8f7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2136.894291] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-742c10ef-e050-4d17-b808-fd2f9bdbc443 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2136.937061] env[62824]: DEBUG nova.network.neutron [-] [instance: bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2136.961541] env[62824]: DEBUG nova.scheduler.client.report [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2137.028677] env[62824]: DEBUG nova.network.neutron [-] [instance: 49818910-61de-4fb8-bbab-d5d61d2f1ada] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2137.076057] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-090989e5-c332-4c8a-a8b8-8f0a018533cc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2137.081611] env[62824]: DEBUG nova.compute.manager [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2137.086802] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dbd7a83-56f4-43b7-b70c-0c75dfa4d926 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2137.122982] env[62824]: DEBUG nova.compute.manager [req-9d75c889-405b-4b3e-8bef-cf953916b619 req-0d31ba67-6e37-4cdf-a67a-b39b6828207a service nova] [instance: bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e] Detach interface failed, port_id=9ad5243a-5d04-419d-bfd7-e4b27536ee65, reason: Instance bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e could not be found. {{(pid=62824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11482}} [ 2137.188482] env[62824]: DEBUG nova.network.neutron [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: d067fa97-cedc-4e3d-9be4-d860a79a7723] Successfully updated port: 2b52b217-d85d-4239-a02a-d97e3c82d93d {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2137.205680] env[62824]: DEBUG nova.compute.manager [None req-d39ad21e-d65d-4fc4-9829-53a13e0b4169 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2137.205951] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-d39ad21e-d65d-4fc4-9829-53a13e0b4169 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2137.207726] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba84d0b8-599d-4248-95dc-a8aaf0ee51c5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2137.215268] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d39ad21e-d65d-4fc4-9829-53a13e0b4169 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2137.215519] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4eaf1231-b206-4fc0-87c2-f71289875c1d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2137.221722] env[62824]: DEBUG oslo_vmware.api [None req-d39ad21e-d65d-4fc4-9829-53a13e0b4169 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Waiting for the task: (returnval){ [ 2137.221722] env[62824]: value = "task-2146405" [ 2137.221722] env[62824]: _type = "Task" [ 2137.221722] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2137.229790] env[62824]: DEBUG oslo_vmware.api [None req-d39ad21e-d65d-4fc4-9829-53a13e0b4169 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Task: {'id': task-2146405, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2137.439954] env[62824]: INFO nova.compute.manager [-] [instance: bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e] Took 1.61 seconds to deallocate network for instance. [ 2137.469213] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.636s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2137.469616] env[62824]: DEBUG nova.compute.manager [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] [instance: dadf590a-7288-4dd2-90de-125fe272f4ba] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2137.473197] env[62824]: DEBUG oslo_concurrency.lockutils [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 31.373s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2137.473464] env[62824]: DEBUG nova.objects.instance [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62824) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 2137.532158] env[62824]: INFO nova.compute.manager [-] [instance: 49818910-61de-4fb8-bbab-d5d61d2f1ada] Took 1.32 seconds to deallocate network for instance. [ 2137.619715] env[62824]: DEBUG oslo_concurrency.lockutils [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2137.691806] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Acquiring lock "refresh_cache-d067fa97-cedc-4e3d-9be4-d860a79a7723" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2137.691806] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Acquired lock "refresh_cache-d067fa97-cedc-4e3d-9be4-d860a79a7723" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2137.691806] env[62824]: DEBUG nova.network.neutron [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: d067fa97-cedc-4e3d-9be4-d860a79a7723] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2137.732696] env[62824]: DEBUG oslo_vmware.api [None req-d39ad21e-d65d-4fc4-9829-53a13e0b4169 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Task: {'id': task-2146405, 'name': PowerOffVM_Task, 'duration_secs': 0.215725} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2137.733341] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d39ad21e-d65d-4fc4-9829-53a13e0b4169 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2137.733341] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-d39ad21e-d65d-4fc4-9829-53a13e0b4169 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2137.733482] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a46c40be-b452-44a0-8e29-c64e6837f904 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2137.849691] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-d39ad21e-d65d-4fc4-9829-53a13e0b4169 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2137.850067] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-d39ad21e-d65d-4fc4-9829-53a13e0b4169 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Deleting contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2137.850151] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-d39ad21e-d65d-4fc4-9829-53a13e0b4169 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Deleting the datastore file [datastore1] 59b5b883-4188-471c-8862-444f3ce08cb0 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2137.850431] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-040bb1d9-1ee8-46ed-bee6-e59bf2e393a5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2137.859805] env[62824]: DEBUG oslo_vmware.api [None req-d39ad21e-d65d-4fc4-9829-53a13e0b4169 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Waiting for the task: (returnval){ [ 2137.859805] env[62824]: value = "task-2146407" [ 2137.859805] env[62824]: _type = "Task" [ 2137.859805] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2137.869749] env[62824]: DEBUG oslo_vmware.api [None req-d39ad21e-d65d-4fc4-9829-53a13e0b4169 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Task: {'id': task-2146407, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2137.946490] env[62824]: DEBUG oslo_concurrency.lockutils [None req-61bb4e80-84d9-4e19-b9c0-195ff35ecb41 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2137.978754] env[62824]: DEBUG nova.compute.utils [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2137.980271] env[62824]: DEBUG nova.compute.manager [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] [instance: dadf590a-7288-4dd2-90de-125fe272f4ba] Not allocating networking since 'none' was specified. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1984}} [ 2138.042061] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b30d3e55-ba09-4502-8b58-e6c2d9aaba20 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2138.224345] env[62824]: DEBUG nova.network.neutron [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: d067fa97-cedc-4e3d-9be4-d860a79a7723] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2138.370547] env[62824]: DEBUG oslo_vmware.api [None req-d39ad21e-d65d-4fc4-9829-53a13e0b4169 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Task: {'id': task-2146407, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.168897} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2138.370811] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-d39ad21e-d65d-4fc4-9829-53a13e0b4169 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2138.370996] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-d39ad21e-d65d-4fc4-9829-53a13e0b4169 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Deleted contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2138.371419] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-d39ad21e-d65d-4fc4-9829-53a13e0b4169 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2138.371609] env[62824]: INFO nova.compute.manager [None req-d39ad21e-d65d-4fc4-9829-53a13e0b4169 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Took 1.17 seconds to destroy the instance on the hypervisor. [ 2138.371855] env[62824]: DEBUG oslo.service.loopingcall [None req-d39ad21e-d65d-4fc4-9829-53a13e0b4169 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2138.372162] env[62824]: DEBUG nova.compute.manager [-] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2138.372267] env[62824]: DEBUG nova.network.neutron [-] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2138.374606] env[62824]: DEBUG nova.network.neutron [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: d067fa97-cedc-4e3d-9be4-d860a79a7723] Updating instance_info_cache with network_info: [{"id": "2b52b217-d85d-4239-a02a-d97e3c82d93d", "address": "fa:16:3e:36:39:80", "network": {"id": "4ee2cade-6795-4887-a1df-f3990219e980", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1173612309-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "25bb8a4df0b740c0a1282ee61ad2f58c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3202e2b9-c4a7-4f78-9476-12ed92fabe61", "external-id": "nsx-vlan-transportzone-846", "segmentation_id": 846, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b52b217-d8", "ovs_interfaceid": "2b52b217-d85d-4239-a02a-d97e3c82d93d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2138.486208] env[62824]: DEBUG nova.compute.manager [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] [instance: dadf590a-7288-4dd2-90de-125fe272f4ba] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2138.489636] env[62824]: DEBUG oslo_concurrency.lockutils [None req-90ba2e87-c9de-452c-aa83-251e08f86592 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.016s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2138.491347] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.250s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2138.492894] env[62824]: INFO nova.compute.claims [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7a2a51c4-e558-4f5f-b82c-718bc12c1df5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2138.602130] env[62824]: DEBUG nova.compute.manager [req-7d135403-0d56-47b0-bf53-383b639d068e req-eb361a13-4a20-4f6e-96ea-17082708c322 service nova] [instance: 49818910-61de-4fb8-bbab-d5d61d2f1ada] Received event network-vif-deleted-f26c1cce-02ce-4d96-b8b7-b206825081b3 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2138.602346] env[62824]: DEBUG nova.compute.manager [req-7d135403-0d56-47b0-bf53-383b639d068e req-eb361a13-4a20-4f6e-96ea-17082708c322 service nova] [instance: d067fa97-cedc-4e3d-9be4-d860a79a7723] Received event network-vif-plugged-2b52b217-d85d-4239-a02a-d97e3c82d93d {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2138.602536] env[62824]: DEBUG oslo_concurrency.lockutils [req-7d135403-0d56-47b0-bf53-383b639d068e req-eb361a13-4a20-4f6e-96ea-17082708c322 service nova] Acquiring lock "d067fa97-cedc-4e3d-9be4-d860a79a7723-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2138.602746] env[62824]: DEBUG oslo_concurrency.lockutils [req-7d135403-0d56-47b0-bf53-383b639d068e req-eb361a13-4a20-4f6e-96ea-17082708c322 service nova] Lock "d067fa97-cedc-4e3d-9be4-d860a79a7723-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2138.602912] env[62824]: DEBUG oslo_concurrency.lockutils [req-7d135403-0d56-47b0-bf53-383b639d068e req-eb361a13-4a20-4f6e-96ea-17082708c322 service nova] Lock "d067fa97-cedc-4e3d-9be4-d860a79a7723-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2138.603318] env[62824]: DEBUG nova.compute.manager [req-7d135403-0d56-47b0-bf53-383b639d068e req-eb361a13-4a20-4f6e-96ea-17082708c322 service nova] [instance: d067fa97-cedc-4e3d-9be4-d860a79a7723] No waiting events found dispatching network-vif-plugged-2b52b217-d85d-4239-a02a-d97e3c82d93d {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2138.603527] env[62824]: WARNING nova.compute.manager [req-7d135403-0d56-47b0-bf53-383b639d068e req-eb361a13-4a20-4f6e-96ea-17082708c322 service nova] [instance: d067fa97-cedc-4e3d-9be4-d860a79a7723] Received unexpected event network-vif-plugged-2b52b217-d85d-4239-a02a-d97e3c82d93d for instance with vm_state building and task_state spawning. [ 2138.603760] env[62824]: DEBUG nova.compute.manager [req-7d135403-0d56-47b0-bf53-383b639d068e req-eb361a13-4a20-4f6e-96ea-17082708c322 service nova] [instance: d067fa97-cedc-4e3d-9be4-d860a79a7723] Received event network-changed-2b52b217-d85d-4239-a02a-d97e3c82d93d {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2138.603943] env[62824]: DEBUG nova.compute.manager [req-7d135403-0d56-47b0-bf53-383b639d068e req-eb361a13-4a20-4f6e-96ea-17082708c322 service nova] [instance: d067fa97-cedc-4e3d-9be4-d860a79a7723] Refreshing instance network info cache due to event network-changed-2b52b217-d85d-4239-a02a-d97e3c82d93d. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2138.604132] env[62824]: DEBUG oslo_concurrency.lockutils [req-7d135403-0d56-47b0-bf53-383b639d068e req-eb361a13-4a20-4f6e-96ea-17082708c322 service nova] Acquiring lock "refresh_cache-d067fa97-cedc-4e3d-9be4-d860a79a7723" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2138.880410] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Releasing lock "refresh_cache-d067fa97-cedc-4e3d-9be4-d860a79a7723" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2138.880740] env[62824]: DEBUG nova.compute.manager [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: d067fa97-cedc-4e3d-9be4-d860a79a7723] Instance network_info: |[{"id": "2b52b217-d85d-4239-a02a-d97e3c82d93d", "address": "fa:16:3e:36:39:80", "network": {"id": "4ee2cade-6795-4887-a1df-f3990219e980", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1173612309-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "25bb8a4df0b740c0a1282ee61ad2f58c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3202e2b9-c4a7-4f78-9476-12ed92fabe61", "external-id": "nsx-vlan-transportzone-846", "segmentation_id": 846, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b52b217-d8", "ovs_interfaceid": "2b52b217-d85d-4239-a02a-d97e3c82d93d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2138.881052] env[62824]: DEBUG oslo_concurrency.lockutils [req-7d135403-0d56-47b0-bf53-383b639d068e req-eb361a13-4a20-4f6e-96ea-17082708c322 service nova] Acquired lock "refresh_cache-d067fa97-cedc-4e3d-9be4-d860a79a7723" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2138.881249] env[62824]: DEBUG nova.network.neutron [req-7d135403-0d56-47b0-bf53-383b639d068e req-eb361a13-4a20-4f6e-96ea-17082708c322 service nova] [instance: d067fa97-cedc-4e3d-9be4-d860a79a7723] Refreshing network info cache for port 2b52b217-d85d-4239-a02a-d97e3c82d93d {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2138.882483] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: d067fa97-cedc-4e3d-9be4-d860a79a7723] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:36:39:80', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3202e2b9-c4a7-4f78-9476-12ed92fabe61', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2b52b217-d85d-4239-a02a-d97e3c82d93d', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2138.892826] env[62824]: DEBUG oslo.service.loopingcall [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2138.896284] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d067fa97-cedc-4e3d-9be4-d860a79a7723] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2138.896901] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dc4c207c-df66-4ba2-8ad8-8f013c1ff419 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2138.940055] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2138.940055] env[62824]: value = "task-2146408" [ 2138.940055] env[62824]: _type = "Task" [ 2138.940055] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2138.945308] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146408, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2139.166089] env[62824]: DEBUG nova.network.neutron [req-7d135403-0d56-47b0-bf53-383b639d068e req-eb361a13-4a20-4f6e-96ea-17082708c322 service nova] [instance: d067fa97-cedc-4e3d-9be4-d860a79a7723] Updated VIF entry in instance network info cache for port 2b52b217-d85d-4239-a02a-d97e3c82d93d. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2139.166477] env[62824]: DEBUG nova.network.neutron [req-7d135403-0d56-47b0-bf53-383b639d068e req-eb361a13-4a20-4f6e-96ea-17082708c322 service nova] [instance: d067fa97-cedc-4e3d-9be4-d860a79a7723] Updating instance_info_cache with network_info: [{"id": "2b52b217-d85d-4239-a02a-d97e3c82d93d", "address": "fa:16:3e:36:39:80", "network": {"id": "4ee2cade-6795-4887-a1df-f3990219e980", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1173612309-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "25bb8a4df0b740c0a1282ee61ad2f58c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3202e2b9-c4a7-4f78-9476-12ed92fabe61", "external-id": "nsx-vlan-transportzone-846", "segmentation_id": 846, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b52b217-d8", "ovs_interfaceid": "2b52b217-d85d-4239-a02a-d97e3c82d93d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2139.336479] env[62824]: DEBUG nova.network.neutron [-] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2139.424428] env[62824]: DEBUG oslo_concurrency.lockutils [None req-123477fc-8c3e-4f2a-83de-cced4f011020 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Acquiring lock "refresh_cache-b8cc8cd3-ea03-40bf-b867-7ad193365552" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2139.424614] env[62824]: DEBUG oslo_concurrency.lockutils [None req-123477fc-8c3e-4f2a-83de-cced4f011020 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Acquired lock "refresh_cache-b8cc8cd3-ea03-40bf-b867-7ad193365552" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2139.424791] env[62824]: DEBUG nova.network.neutron [None req-123477fc-8c3e-4f2a-83de-cced4f011020 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2139.449156] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146408, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2139.451388] env[62824]: DEBUG nova.compute.manager [req-7a5c1e9a-f9fd-4cd3-a4af-f61ee283b0ba req-e68a4a4a-875c-4433-84f4-4cdbd2afb486 service nova] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Received event network-changed-60a4fce0-0e63-45af-890c-46ca44ea9a0c {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2139.451602] env[62824]: DEBUG nova.compute.manager [req-7a5c1e9a-f9fd-4cd3-a4af-f61ee283b0ba req-e68a4a4a-875c-4433-84f4-4cdbd2afb486 service nova] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Refreshing instance network info cache due to event network-changed-60a4fce0-0e63-45af-890c-46ca44ea9a0c. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2139.451759] env[62824]: DEBUG oslo_concurrency.lockutils [req-7a5c1e9a-f9fd-4cd3-a4af-f61ee283b0ba req-e68a4a4a-875c-4433-84f4-4cdbd2afb486 service nova] Acquiring lock "refresh_cache-b8cc8cd3-ea03-40bf-b867-7ad193365552" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2139.500275] env[62824]: DEBUG nova.compute.manager [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] [instance: dadf590a-7288-4dd2-90de-125fe272f4ba] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2139.529915] env[62824]: DEBUG nova.virt.hardware [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2139.530219] env[62824]: DEBUG nova.virt.hardware [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2139.530383] env[62824]: DEBUG nova.virt.hardware [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2139.530681] env[62824]: DEBUG nova.virt.hardware [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2139.530762] env[62824]: DEBUG nova.virt.hardware [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2139.530901] env[62824]: DEBUG nova.virt.hardware [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2139.531131] env[62824]: DEBUG nova.virt.hardware [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2139.531299] env[62824]: DEBUG nova.virt.hardware [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2139.531471] env[62824]: DEBUG nova.virt.hardware [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2139.531637] env[62824]: DEBUG nova.virt.hardware [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2139.531811] env[62824]: DEBUG nova.virt.hardware [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2139.532709] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4b62bcd-04eb-47b9-87c4-4095fa978c84 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2139.542910] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec2b6210-9e8b-451e-beab-7e78d479c133 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2139.559364] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] [instance: dadf590a-7288-4dd2-90de-125fe272f4ba] Instance VIF info [] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2139.564893] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Creating folder: Project (f059d46a2ce64a6298ecc1fd4e37b34a). Parent ref: group-v438503. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 2139.567571] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-339adbfa-14d3-40f9-9778-65595e2243ca {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2139.577994] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Created folder: Project (f059d46a2ce64a6298ecc1fd4e37b34a) in parent group-v438503. [ 2139.578236] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Creating folder: Instances. Parent ref: group-v438692. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 2139.580788] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ce228edf-3830-4bda-8b8b-4cd79418a922 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2139.591550] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Created folder: Instances in parent group-v438692. [ 2139.591719] env[62824]: DEBUG oslo.service.loopingcall [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2139.594253] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dadf590a-7288-4dd2-90de-125fe272f4ba] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2139.594644] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5f0b08f0-01f6-4947-ad21-12bb751a4837 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2139.614505] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2139.614505] env[62824]: value = "task-2146411" [ 2139.614505] env[62824]: _type = "Task" [ 2139.614505] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2139.624654] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146411, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2139.669920] env[62824]: DEBUG oslo_concurrency.lockutils [req-7d135403-0d56-47b0-bf53-383b639d068e req-eb361a13-4a20-4f6e-96ea-17082708c322 service nova] Releasing lock "refresh_cache-d067fa97-cedc-4e3d-9be4-d860a79a7723" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2139.842266] env[62824]: INFO nova.compute.manager [-] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Took 1.47 seconds to deallocate network for instance. [ 2139.950567] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146408, 'name': CreateVM_Task, 'duration_secs': 0.700975} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2139.950756] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d067fa97-cedc-4e3d-9be4-d860a79a7723] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2139.951593] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2139.951763] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2139.952097] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2139.952366] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1f9a6dcc-3179-4e91-aedd-52156c3013f9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2139.958204] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Waiting for the task: (returnval){ [ 2139.958204] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5201794e-fa04-1c4c-341e-f4594cec7442" [ 2139.958204] env[62824]: _type = "Task" [ 2139.958204] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2139.966051] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5201794e-fa04-1c4c-341e-f4594cec7442, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2139.980878] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72ab0d43-835a-4698-9579-0c8aa9217fec {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2139.988242] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef784565-1534-41af-87a2-0810761e58ac {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2140.023675] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85330c21-b367-4c3b-b8bf-eaac3940cc09 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2140.031502] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b839095-4c63-4505-8831-7f320e7b0edb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2140.044950] env[62824]: DEBUG nova.compute.provider_tree [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2140.127862] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146411, 'name': CreateVM_Task, 'duration_secs': 0.284357} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2140.128085] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dadf590a-7288-4dd2-90de-125fe272f4ba] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2140.128525] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2140.128690] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2140.129035] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2140.129266] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-15872a55-a732-4342-8cd0-4856a6eadd5f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2140.133526] env[62824]: DEBUG oslo_vmware.api [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Waiting for the task: (returnval){ [ 2140.133526] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]523ba385-974a-9867-d3c4-ca100b7132e9" [ 2140.133526] env[62824]: _type = "Task" [ 2140.133526] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2140.141103] env[62824]: DEBUG oslo_vmware.api [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]523ba385-974a-9867-d3c4-ca100b7132e9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2140.257619] env[62824]: DEBUG nova.network.neutron [None req-123477fc-8c3e-4f2a-83de-cced4f011020 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Updating instance_info_cache with network_info: [{"id": "60a4fce0-0e63-45af-890c-46ca44ea9a0c", "address": "fa:16:3e:cf:18:eb", "network": {"id": "db233b0c-27f3-4a76-ae9c-2b5889bf3d3b", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-1825671448-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.231", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3240be4f32254f54aa87cb045e3d8f31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "binding_failed", "details": {}, "devname": "tap60a4fce0-0e", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2140.349745] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d39ad21e-d65d-4fc4-9829-53a13e0b4169 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2140.468760] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5201794e-fa04-1c4c-341e-f4594cec7442, 'name': SearchDatastore_Task, 'duration_secs': 0.010515} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2140.469066] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2140.469316] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: d067fa97-cedc-4e3d-9be4-d860a79a7723] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2140.469534] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2140.469680] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2140.469857] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2140.470129] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-aa39b572-0e88-4fb2-892f-69ed7b1a1c3a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2140.478115] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2140.478305] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2140.479000] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0e4eda94-2830-42a8-a4f6-0a06033a18c0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2140.484138] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Waiting for the task: (returnval){ [ 2140.484138] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52d9027c-ffc1-646f-e0b9-a2e4f9406143" [ 2140.484138] env[62824]: _type = "Task" [ 2140.484138] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2140.491144] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52d9027c-ffc1-646f-e0b9-a2e4f9406143, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2140.548824] env[62824]: DEBUG nova.scheduler.client.report [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2140.632116] env[62824]: DEBUG nova.compute.manager [req-ccf47faf-7760-499f-ab54-22ab660b9658 req-4c4e01a0-66fb-4c71-967e-857554fe0363 service nova] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Received event network-vif-deleted-1e74959f-64c2-4c37-85f8-fa8f81a30d60 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2140.644333] env[62824]: DEBUG oslo_vmware.api [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]523ba385-974a-9867-d3c4-ca100b7132e9, 'name': SearchDatastore_Task, 'duration_secs': 0.011182} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2140.644630] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2140.644859] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] [instance: dadf590a-7288-4dd2-90de-125fe272f4ba] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2140.645172] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2140.645331] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2140.645513] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2140.645801] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f7edc44b-c6cf-4937-9b16-48b5c3503469 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2140.653554] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2140.653723] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2140.654410] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-13655a75-0c02-4efa-a354-4e12cc5cacee {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2140.659584] env[62824]: DEBUG oslo_vmware.api [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Waiting for the task: (returnval){ [ 2140.659584] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52dade1b-f3b6-b998-5288-6061c08b17be" [ 2140.659584] env[62824]: _type = "Task" [ 2140.659584] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2140.667076] env[62824]: DEBUG oslo_vmware.api [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52dade1b-f3b6-b998-5288-6061c08b17be, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2140.760291] env[62824]: DEBUG oslo_concurrency.lockutils [None req-123477fc-8c3e-4f2a-83de-cced4f011020 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Releasing lock "refresh_cache-b8cc8cd3-ea03-40bf-b867-7ad193365552" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2140.763022] env[62824]: DEBUG oslo_concurrency.lockutils [req-7a5c1e9a-f9fd-4cd3-a4af-f61ee283b0ba req-e68a4a4a-875c-4433-84f4-4cdbd2afb486 service nova] Acquired lock "refresh_cache-b8cc8cd3-ea03-40bf-b867-7ad193365552" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2140.763022] env[62824]: DEBUG nova.network.neutron [req-7a5c1e9a-f9fd-4cd3-a4af-f61ee283b0ba req-e68a4a4a-875c-4433-84f4-4cdbd2afb486 service nova] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Refreshing network info cache for port 60a4fce0-0e63-45af-890c-46ca44ea9a0c {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2140.787947] env[62824]: DEBUG nova.virt.hardware [None req-123477fc-8c3e-4f2a-83de-cced4f011020 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='08b4b315b8956a22a2d81f53bbe808c3',container_format='bare',created_at=2025-01-16T14:54:11Z,direct_url=,disk_format='vmdk',id=cc65c6c8-036d-4ab6-a8a0-770a7853932b,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-1338477357-shelved',owner='3240be4f32254f54aa87cb045e3d8f31',properties=ImageMetaProps,protected=,size=31671808,status='active',tags=,updated_at=2025-01-16T14:54:28Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2140.788542] env[62824]: DEBUG nova.virt.hardware [None req-123477fc-8c3e-4f2a-83de-cced4f011020 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2140.788542] env[62824]: DEBUG nova.virt.hardware [None req-123477fc-8c3e-4f2a-83de-cced4f011020 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2140.788652] env[62824]: DEBUG nova.virt.hardware [None req-123477fc-8c3e-4f2a-83de-cced4f011020 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2140.788701] env[62824]: DEBUG nova.virt.hardware [None req-123477fc-8c3e-4f2a-83de-cced4f011020 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2140.788840] env[62824]: DEBUG nova.virt.hardware [None req-123477fc-8c3e-4f2a-83de-cced4f011020 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2140.789055] env[62824]: DEBUG nova.virt.hardware [None req-123477fc-8c3e-4f2a-83de-cced4f011020 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2140.789216] env[62824]: DEBUG nova.virt.hardware [None req-123477fc-8c3e-4f2a-83de-cced4f011020 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2140.789382] env[62824]: DEBUG nova.virt.hardware [None req-123477fc-8c3e-4f2a-83de-cced4f011020 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2140.789550] env[62824]: DEBUG nova.virt.hardware [None req-123477fc-8c3e-4f2a-83de-cced4f011020 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2140.789726] env[62824]: DEBUG nova.virt.hardware [None req-123477fc-8c3e-4f2a-83de-cced4f011020 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2140.790593] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ad26e62-78e5-484e-9d07-1b5e53e3b504 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2140.799457] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad8510b9-fe1c-47e0-aece-bbde34f2c4ff {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2140.814440] env[62824]: DEBUG nova.compute.claims [None req-123477fc-8c3e-4f2a-83de-cced4f011020 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Aborting claim: {{(pid=62824) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 2140.814645] env[62824]: DEBUG oslo_concurrency.lockutils [None req-123477fc-8c3e-4f2a-83de-cced4f011020 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2140.994386] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52d9027c-ffc1-646f-e0b9-a2e4f9406143, 'name': SearchDatastore_Task, 'duration_secs': 0.008268} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2140.995281] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4ec90b80-8fc9-450a-adae-c0ffebd5c7ea {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2141.000439] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Waiting for the task: (returnval){ [ 2141.000439] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52557b10-d1ba-2591-1e74-c3a9f7d4fdfb" [ 2141.000439] env[62824]: _type = "Task" [ 2141.000439] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2141.007867] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52557b10-d1ba-2591-1e74-c3a9f7d4fdfb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2141.053888] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.562s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2141.054135] env[62824]: DEBUG nova.compute.manager [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7a2a51c4-e558-4f5f-b82c-718bc12c1df5] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2141.056703] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.437s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2141.058150] env[62824]: INFO nova.compute.claims [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 20f93b46-5e7e-4256-8cc1-e0a1b16740d5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2141.169548] env[62824]: DEBUG oslo_vmware.api [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52dade1b-f3b6-b998-5288-6061c08b17be, 'name': SearchDatastore_Task, 'duration_secs': 0.008177} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2141.170300] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c72553b2-53f4-4ff2-85e2-6bc2171b0a07 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2141.175195] env[62824]: DEBUG oslo_vmware.api [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Waiting for the task: (returnval){ [ 2141.175195] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]528cfb62-89e7-0018-ffc7-3c0a987e38e3" [ 2141.175195] env[62824]: _type = "Task" [ 2141.175195] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2141.182681] env[62824]: DEBUG oslo_vmware.api [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]528cfb62-89e7-0018-ffc7-3c0a987e38e3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2141.510161] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52557b10-d1ba-2591-1e74-c3a9f7d4fdfb, 'name': SearchDatastore_Task, 'duration_secs': 0.008811} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2141.510470] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2141.510672] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] d067fa97-cedc-4e3d-9be4-d860a79a7723/d067fa97-cedc-4e3d-9be4-d860a79a7723.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2141.510921] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2907494a-a86d-44b5-96bd-c418ec33557a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2141.518048] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Waiting for the task: (returnval){ [ 2141.518048] env[62824]: value = "task-2146412" [ 2141.518048] env[62824]: _type = "Task" [ 2141.518048] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2141.525499] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Task: {'id': task-2146412, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2141.532541] env[62824]: DEBUG nova.network.neutron [req-7a5c1e9a-f9fd-4cd3-a4af-f61ee283b0ba req-e68a4a4a-875c-4433-84f4-4cdbd2afb486 service nova] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Updated VIF entry in instance network info cache for port 60a4fce0-0e63-45af-890c-46ca44ea9a0c. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2141.532858] env[62824]: DEBUG nova.network.neutron [req-7a5c1e9a-f9fd-4cd3-a4af-f61ee283b0ba req-e68a4a4a-875c-4433-84f4-4cdbd2afb486 service nova] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Updating instance_info_cache with network_info: [{"id": "60a4fce0-0e63-45af-890c-46ca44ea9a0c", "address": "fa:16:3e:cf:18:eb", "network": {"id": "db233b0c-27f3-4a76-ae9c-2b5889bf3d3b", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-1825671448-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.231", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3240be4f32254f54aa87cb045e3d8f31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "binding_failed", "details": {}, "devname": "tap60a4fce0-0e", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2141.562647] env[62824]: DEBUG nova.compute.utils [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2141.565850] env[62824]: DEBUG nova.compute.manager [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7a2a51c4-e558-4f5f-b82c-718bc12c1df5] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2141.566027] env[62824]: DEBUG nova.network.neutron [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7a2a51c4-e558-4f5f-b82c-718bc12c1df5] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2141.605678] env[62824]: DEBUG nova.policy [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8c6cd8cd9ac848c89218208d1c2290aa', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '461117110de7467fbf8ada5ee9978069', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2141.686235] env[62824]: DEBUG oslo_vmware.api [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]528cfb62-89e7-0018-ffc7-3c0a987e38e3, 'name': SearchDatastore_Task, 'duration_secs': 0.010641} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2141.686537] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2141.686799] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] dadf590a-7288-4dd2-90de-125fe272f4ba/dadf590a-7288-4dd2-90de-125fe272f4ba.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2141.687161] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ca39d761-a64c-4e8b-a6b7-eceb226b9cac {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2141.694898] env[62824]: DEBUG oslo_vmware.api [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Waiting for the task: (returnval){ [ 2141.694898] env[62824]: value = "task-2146413" [ 2141.694898] env[62824]: _type = "Task" [ 2141.694898] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2141.704150] env[62824]: DEBUG oslo_vmware.api [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Task: {'id': task-2146413, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2141.910585] env[62824]: DEBUG nova.network.neutron [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7a2a51c4-e558-4f5f-b82c-718bc12c1df5] Successfully created port: 9ad6111d-68a7-4700-9b92-1a7482806e37 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2142.030652] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Task: {'id': task-2146412, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.467915} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2142.031012] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] d067fa97-cedc-4e3d-9be4-d860a79a7723/d067fa97-cedc-4e3d-9be4-d860a79a7723.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2142.031303] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: d067fa97-cedc-4e3d-9be4-d860a79a7723] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2142.031603] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bcba5531-fbb2-4db8-891e-7d89ac0ac8d8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2142.035763] env[62824]: DEBUG oslo_concurrency.lockutils [req-7a5c1e9a-f9fd-4cd3-a4af-f61ee283b0ba req-e68a4a4a-875c-4433-84f4-4cdbd2afb486 service nova] Releasing lock "refresh_cache-b8cc8cd3-ea03-40bf-b867-7ad193365552" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2142.039362] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Waiting for the task: (returnval){ [ 2142.039362] env[62824]: value = "task-2146414" [ 2142.039362] env[62824]: _type = "Task" [ 2142.039362] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2142.049663] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Task: {'id': task-2146414, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2142.069081] env[62824]: DEBUG nova.compute.manager [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7a2a51c4-e558-4f5f-b82c-718bc12c1df5] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2142.211827] env[62824]: DEBUG oslo_vmware.api [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Task: {'id': task-2146413, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2142.548189] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-868e4f24-4944-4ede-8a8b-e6a55d50c815 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2142.553358] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Task: {'id': task-2146414, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.103384} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2142.553943] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: d067fa97-cedc-4e3d-9be4-d860a79a7723] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2142.554725] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a118bd10-80ac-4d1d-864f-d9a7e1692a2b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2142.559839] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9577df63-7dc4-437e-97bc-e158c3663122 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2142.579622] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: d067fa97-cedc-4e3d-9be4-d860a79a7723] Reconfiguring VM instance instance-00000048 to attach disk [datastore2] d067fa97-cedc-4e3d-9be4-d860a79a7723/d067fa97-cedc-4e3d-9be4-d860a79a7723.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2142.583261] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-00a14cec-71d4-404c-a06c-c637a4dafefa {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2142.622030] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6548a9f-eef9-4a88-9f19-cd190caa8e40 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2142.626928] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Waiting for the task: (returnval){ [ 2142.626928] env[62824]: value = "task-2146415" [ 2142.626928] env[62824]: _type = "Task" [ 2142.626928] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2142.633049] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7d94d8d-060a-4c06-8062-4d4ce09d652b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2142.648157] env[62824]: DEBUG nova.compute.provider_tree [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2142.652663] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Task: {'id': task-2146415, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2142.707036] env[62824]: DEBUG oslo_vmware.api [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Task: {'id': task-2146413, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.677693} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2142.707036] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] dadf590a-7288-4dd2-90de-125fe272f4ba/dadf590a-7288-4dd2-90de-125fe272f4ba.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2142.707036] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] [instance: dadf590a-7288-4dd2-90de-125fe272f4ba] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2142.707282] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6dc35efb-6298-4bed-bd2a-e654ba095230 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2142.713088] env[62824]: DEBUG oslo_vmware.api [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Waiting for the task: (returnval){ [ 2142.713088] env[62824]: value = "task-2146416" [ 2142.713088] env[62824]: _type = "Task" [ 2142.713088] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2142.720893] env[62824]: DEBUG oslo_vmware.api [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Task: {'id': task-2146416, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2143.099361] env[62824]: DEBUG nova.compute.manager [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7a2a51c4-e558-4f5f-b82c-718bc12c1df5] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2143.128118] env[62824]: DEBUG nova.virt.hardware [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2143.128475] env[62824]: DEBUG nova.virt.hardware [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2143.128702] env[62824]: DEBUG nova.virt.hardware [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2143.128986] env[62824]: DEBUG nova.virt.hardware [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2143.129252] env[62824]: DEBUG nova.virt.hardware [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2143.129534] env[62824]: DEBUG nova.virt.hardware [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2143.129823] env[62824]: DEBUG nova.virt.hardware [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2143.130046] env[62824]: DEBUG nova.virt.hardware [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2143.130260] env[62824]: DEBUG nova.virt.hardware [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2143.130446] env[62824]: DEBUG nova.virt.hardware [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2143.130705] env[62824]: DEBUG nova.virt.hardware [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2143.133638] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d701616-571c-40df-a25a-53e954f1f41a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2143.146912] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7944c16-b4af-4064-b6c9-42299a775b69 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2143.150837] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Task: {'id': task-2146415, 'name': ReconfigVM_Task, 'duration_secs': 0.284684} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2143.151123] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: d067fa97-cedc-4e3d-9be4-d860a79a7723] Reconfigured VM instance instance-00000048 to attach disk [datastore2] d067fa97-cedc-4e3d-9be4-d860a79a7723/d067fa97-cedc-4e3d-9be4-d860a79a7723.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2143.152029] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3222b64c-9818-4e73-a50d-7935f9f1f604 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2143.153940] env[62824]: DEBUG nova.scheduler.client.report [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2143.181502] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Waiting for the task: (returnval){ [ 2143.181502] env[62824]: value = "task-2146417" [ 2143.181502] env[62824]: _type = "Task" [ 2143.181502] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2143.191327] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Task: {'id': task-2146417, 'name': Rename_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2143.223285] env[62824]: DEBUG oslo_vmware.api [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Task: {'id': task-2146416, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.086817} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2143.223698] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] [instance: dadf590a-7288-4dd2-90de-125fe272f4ba] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2143.224543] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-589f710d-6d32-4696-97ac-68cc27767e45 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2143.245675] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] [instance: dadf590a-7288-4dd2-90de-125fe272f4ba] Reconfiguring VM instance instance-0000004a to attach disk [datastore1] dadf590a-7288-4dd2-90de-125fe272f4ba/dadf590a-7288-4dd2-90de-125fe272f4ba.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2143.246344] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c9512bce-4f3e-41f5-b1ad-f435c12d9cd5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2143.267185] env[62824]: DEBUG oslo_vmware.api [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Waiting for the task: (returnval){ [ 2143.267185] env[62824]: value = "task-2146418" [ 2143.267185] env[62824]: _type = "Task" [ 2143.267185] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2143.275978] env[62824]: DEBUG oslo_vmware.api [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Task: {'id': task-2146418, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2143.540816] env[62824]: DEBUG nova.compute.manager [req-0de71d29-076a-426c-bde2-72e2e436ad4b req-fe557929-4eb3-4467-afc1-d1e3cb06cc64 service nova] [instance: 7a2a51c4-e558-4f5f-b82c-718bc12c1df5] Received event network-vif-plugged-9ad6111d-68a7-4700-9b92-1a7482806e37 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2143.541061] env[62824]: DEBUG oslo_concurrency.lockutils [req-0de71d29-076a-426c-bde2-72e2e436ad4b req-fe557929-4eb3-4467-afc1-d1e3cb06cc64 service nova] Acquiring lock "7a2a51c4-e558-4f5f-b82c-718bc12c1df5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2143.541279] env[62824]: DEBUG oslo_concurrency.lockutils [req-0de71d29-076a-426c-bde2-72e2e436ad4b req-fe557929-4eb3-4467-afc1-d1e3cb06cc64 service nova] Lock "7a2a51c4-e558-4f5f-b82c-718bc12c1df5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2143.541451] env[62824]: DEBUG oslo_concurrency.lockutils [req-0de71d29-076a-426c-bde2-72e2e436ad4b req-fe557929-4eb3-4467-afc1-d1e3cb06cc64 service nova] Lock "7a2a51c4-e558-4f5f-b82c-718bc12c1df5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2143.541680] env[62824]: DEBUG nova.compute.manager [req-0de71d29-076a-426c-bde2-72e2e436ad4b req-fe557929-4eb3-4467-afc1-d1e3cb06cc64 service nova] [instance: 7a2a51c4-e558-4f5f-b82c-718bc12c1df5] No waiting events found dispatching network-vif-plugged-9ad6111d-68a7-4700-9b92-1a7482806e37 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2143.541858] env[62824]: WARNING nova.compute.manager [req-0de71d29-076a-426c-bde2-72e2e436ad4b req-fe557929-4eb3-4467-afc1-d1e3cb06cc64 service nova] [instance: 7a2a51c4-e558-4f5f-b82c-718bc12c1df5] Received unexpected event network-vif-plugged-9ad6111d-68a7-4700-9b92-1a7482806e37 for instance with vm_state building and task_state spawning. [ 2143.666506] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.610s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2143.667149] env[62824]: DEBUG nova.compute.manager [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 20f93b46-5e7e-4256-8cc1-e0a1b16740d5] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2143.670376] env[62824]: DEBUG oslo_concurrency.lockutils [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 27.387s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2143.670561] env[62824]: DEBUG nova.objects.instance [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62824) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 2143.693835] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Task: {'id': task-2146417, 'name': Rename_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2143.778310] env[62824]: DEBUG oslo_vmware.api [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Task: {'id': task-2146418, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2143.930155] env[62824]: DEBUG nova.network.neutron [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7a2a51c4-e558-4f5f-b82c-718bc12c1df5] Successfully updated port: 9ad6111d-68a7-4700-9b92-1a7482806e37 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2144.174446] env[62824]: DEBUG nova.compute.utils [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2144.175870] env[62824]: DEBUG nova.compute.manager [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 20f93b46-5e7e-4256-8cc1-e0a1b16740d5] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2144.176052] env[62824]: DEBUG nova.network.neutron [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 20f93b46-5e7e-4256-8cc1-e0a1b16740d5] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2144.192958] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Task: {'id': task-2146417, 'name': Rename_Task, 'duration_secs': 0.889776} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2144.193238] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: d067fa97-cedc-4e3d-9be4-d860a79a7723] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2144.193480] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-43485430-af72-42e2-b218-7ae59842651b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2144.205290] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Waiting for the task: (returnval){ [ 2144.205290] env[62824]: value = "task-2146419" [ 2144.205290] env[62824]: _type = "Task" [ 2144.205290] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2144.212991] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Task: {'id': task-2146419, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2144.217452] env[62824]: DEBUG nova.policy [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8c6cd8cd9ac848c89218208d1c2290aa', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '461117110de7467fbf8ada5ee9978069', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2144.276565] env[62824]: DEBUG oslo_vmware.api [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Task: {'id': task-2146418, 'name': ReconfigVM_Task, 'duration_secs': 0.876371} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2144.276946] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] [instance: dadf590a-7288-4dd2-90de-125fe272f4ba] Reconfigured VM instance instance-0000004a to attach disk [datastore1] dadf590a-7288-4dd2-90de-125fe272f4ba/dadf590a-7288-4dd2-90de-125fe272f4ba.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2144.277607] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dde3a913-a859-4dd5-a6f4-aa12f5306cf3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2144.283302] env[62824]: DEBUG oslo_vmware.api [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Waiting for the task: (returnval){ [ 2144.283302] env[62824]: value = "task-2146420" [ 2144.283302] env[62824]: _type = "Task" [ 2144.283302] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2144.291277] env[62824]: DEBUG oslo_vmware.api [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Task: {'id': task-2146420, 'name': Rename_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2144.433262] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Acquiring lock "refresh_cache-7a2a51c4-e558-4f5f-b82c-718bc12c1df5" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2144.433425] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Acquired lock "refresh_cache-7a2a51c4-e558-4f5f-b82c-718bc12c1df5" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2144.433583] env[62824]: DEBUG nova.network.neutron [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7a2a51c4-e558-4f5f-b82c-718bc12c1df5] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2144.604261] env[62824]: DEBUG nova.network.neutron [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 20f93b46-5e7e-4256-8cc1-e0a1b16740d5] Successfully created port: 32e4bbdc-8153-4bb8-9c2d-669a9c40b380 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2144.681339] env[62824]: DEBUG nova.compute.manager [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 20f93b46-5e7e-4256-8cc1-e0a1b16740d5] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2144.684833] env[62824]: DEBUG oslo_concurrency.lockutils [None req-cf91fb23-1957-4f94-afcf-03dc9f864960 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.014s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2144.685959] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.809s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2144.687439] env[62824]: INFO nova.compute.claims [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2144.715654] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Task: {'id': task-2146419, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2144.794173] env[62824]: DEBUG oslo_vmware.api [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Task: {'id': task-2146420, 'name': Rename_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2144.968859] env[62824]: DEBUG nova.network.neutron [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7a2a51c4-e558-4f5f-b82c-718bc12c1df5] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2145.113035] env[62824]: DEBUG nova.network.neutron [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7a2a51c4-e558-4f5f-b82c-718bc12c1df5] Updating instance_info_cache with network_info: [{"id": "9ad6111d-68a7-4700-9b92-1a7482806e37", "address": "fa:16:3e:79:00:9f", "network": {"id": "91425d93-4190-4e61-ae47-94efd278247d", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1364948114-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "461117110de7467fbf8ada5ee9978069", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c894ab55-c869-4530-9702-cb46d173ce94", "external-id": "nsx-vlan-transportzone-792", "segmentation_id": 792, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9ad6111d-68", "ovs_interfaceid": "9ad6111d-68a7-4700-9b92-1a7482806e37", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2145.221138] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Task: {'id': task-2146419, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2145.295160] env[62824]: DEBUG oslo_vmware.api [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Task: {'id': task-2146420, 'name': Rename_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2145.569272] env[62824]: DEBUG nova.compute.manager [req-680e1fdc-dc0d-4151-b6d3-5ed284133942 req-3ad60b72-f1eb-4c3d-8bc9-ae9693b490f8 service nova] [instance: 7a2a51c4-e558-4f5f-b82c-718bc12c1df5] Received event network-changed-9ad6111d-68a7-4700-9b92-1a7482806e37 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2145.569272] env[62824]: DEBUG nova.compute.manager [req-680e1fdc-dc0d-4151-b6d3-5ed284133942 req-3ad60b72-f1eb-4c3d-8bc9-ae9693b490f8 service nova] [instance: 7a2a51c4-e558-4f5f-b82c-718bc12c1df5] Refreshing instance network info cache due to event network-changed-9ad6111d-68a7-4700-9b92-1a7482806e37. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2145.569454] env[62824]: DEBUG oslo_concurrency.lockutils [req-680e1fdc-dc0d-4151-b6d3-5ed284133942 req-3ad60b72-f1eb-4c3d-8bc9-ae9693b490f8 service nova] Acquiring lock "refresh_cache-7a2a51c4-e558-4f5f-b82c-718bc12c1df5" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2145.615414] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Releasing lock "refresh_cache-7a2a51c4-e558-4f5f-b82c-718bc12c1df5" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2145.615941] env[62824]: DEBUG nova.compute.manager [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7a2a51c4-e558-4f5f-b82c-718bc12c1df5] Instance network_info: |[{"id": "9ad6111d-68a7-4700-9b92-1a7482806e37", "address": "fa:16:3e:79:00:9f", "network": {"id": "91425d93-4190-4e61-ae47-94efd278247d", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1364948114-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "461117110de7467fbf8ada5ee9978069", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c894ab55-c869-4530-9702-cb46d173ce94", "external-id": "nsx-vlan-transportzone-792", "segmentation_id": 792, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9ad6111d-68", "ovs_interfaceid": "9ad6111d-68a7-4700-9b92-1a7482806e37", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2145.617175] env[62824]: DEBUG oslo_concurrency.lockutils [req-680e1fdc-dc0d-4151-b6d3-5ed284133942 req-3ad60b72-f1eb-4c3d-8bc9-ae9693b490f8 service nova] Acquired lock "refresh_cache-7a2a51c4-e558-4f5f-b82c-718bc12c1df5" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2145.617175] env[62824]: DEBUG nova.network.neutron [req-680e1fdc-dc0d-4151-b6d3-5ed284133942 req-3ad60b72-f1eb-4c3d-8bc9-ae9693b490f8 service nova] [instance: 7a2a51c4-e558-4f5f-b82c-718bc12c1df5] Refreshing network info cache for port 9ad6111d-68a7-4700-9b92-1a7482806e37 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2145.617653] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7a2a51c4-e558-4f5f-b82c-718bc12c1df5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:79:00:9f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c894ab55-c869-4530-9702-cb46d173ce94', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9ad6111d-68a7-4700-9b92-1a7482806e37', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2145.626507] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Creating folder: Project (461117110de7467fbf8ada5ee9978069). Parent ref: group-v438503. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 2145.629746] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4fa40f82-eba1-470b-b9a5-2e59faf1b2e5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2145.641786] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Created folder: Project (461117110de7467fbf8ada5ee9978069) in parent group-v438503. [ 2145.641966] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Creating folder: Instances. Parent ref: group-v438695. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 2145.643021] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-946cd5ed-eb7d-4c46-9433-d7002609c947 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2145.651248] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Created folder: Instances in parent group-v438695. [ 2145.651496] env[62824]: DEBUG oslo.service.loopingcall [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2145.652210] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7a2a51c4-e558-4f5f-b82c-718bc12c1df5] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2145.652435] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e6d375fb-f2e9-46dd-b44a-029b2f964f0f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2145.674461] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2145.674461] env[62824]: value = "task-2146423" [ 2145.674461] env[62824]: _type = "Task" [ 2145.674461] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2145.681981] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146423, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2145.693201] env[62824]: DEBUG nova.compute.manager [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 20f93b46-5e7e-4256-8cc1-e0a1b16740d5] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2145.718176] env[62824]: DEBUG oslo_vmware.api [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Task: {'id': task-2146419, 'name': PowerOnVM_Task, 'duration_secs': 1.296765} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2145.718332] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: d067fa97-cedc-4e3d-9be4-d860a79a7723] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2145.718736] env[62824]: INFO nova.compute.manager [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: d067fa97-cedc-4e3d-9be4-d860a79a7723] Took 8.86 seconds to spawn the instance on the hypervisor. [ 2145.718736] env[62824]: DEBUG nova.compute.manager [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: d067fa97-cedc-4e3d-9be4-d860a79a7723] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2145.719459] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98c31884-9e1b-4faf-a094-607dfb1bf395 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2145.723967] env[62824]: DEBUG nova.virt.hardware [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2145.724225] env[62824]: DEBUG nova.virt.hardware [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2145.724393] env[62824]: DEBUG nova.virt.hardware [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2145.724624] env[62824]: DEBUG nova.virt.hardware [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2145.724855] env[62824]: DEBUG nova.virt.hardware [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2145.725066] env[62824]: DEBUG nova.virt.hardware [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2145.725327] env[62824]: DEBUG nova.virt.hardware [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2145.725565] env[62824]: DEBUG nova.virt.hardware [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2145.725788] env[62824]: DEBUG nova.virt.hardware [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2145.726107] env[62824]: DEBUG nova.virt.hardware [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2145.726339] env[62824]: DEBUG nova.virt.hardware [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2145.727188] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e867033-aef9-4f9a-b8f9-76f432abdaf1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2145.743955] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d0ded4e-ef48-4693-becf-a5cc581b6e7f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2145.796550] env[62824]: DEBUG oslo_vmware.api [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Task: {'id': task-2146420, 'name': Rename_Task, 'duration_secs': 1.020446} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2145.798760] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] [instance: dadf590a-7288-4dd2-90de-125fe272f4ba] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2145.799236] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3740d81e-3535-4b1e-81a1-825ec906abfc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2145.806166] env[62824]: DEBUG oslo_vmware.api [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Waiting for the task: (returnval){ [ 2145.806166] env[62824]: value = "task-2146424" [ 2145.806166] env[62824]: _type = "Task" [ 2145.806166] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2145.816043] env[62824]: DEBUG oslo_vmware.api [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Task: {'id': task-2146424, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2145.865148] env[62824]: DEBUG nova.network.neutron [req-680e1fdc-dc0d-4151-b6d3-5ed284133942 req-3ad60b72-f1eb-4c3d-8bc9-ae9693b490f8 service nova] [instance: 7a2a51c4-e558-4f5f-b82c-718bc12c1df5] Updated VIF entry in instance network info cache for port 9ad6111d-68a7-4700-9b92-1a7482806e37. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2145.865605] env[62824]: DEBUG nova.network.neutron [req-680e1fdc-dc0d-4151-b6d3-5ed284133942 req-3ad60b72-f1eb-4c3d-8bc9-ae9693b490f8 service nova] [instance: 7a2a51c4-e558-4f5f-b82c-718bc12c1df5] Updating instance_info_cache with network_info: [{"id": "9ad6111d-68a7-4700-9b92-1a7482806e37", "address": "fa:16:3e:79:00:9f", "network": {"id": "91425d93-4190-4e61-ae47-94efd278247d", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1364948114-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "461117110de7467fbf8ada5ee9978069", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c894ab55-c869-4530-9702-cb46d173ce94", "external-id": "nsx-vlan-transportzone-792", "segmentation_id": 792, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9ad6111d-68", "ovs_interfaceid": "9ad6111d-68a7-4700-9b92-1a7482806e37", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2146.144952] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d45ce449-3ea2-4bbc-8532-6bd2849cf978 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2146.152931] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8469cf5c-872e-433f-a892-155b65a72676 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2146.186942] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48d74317-9969-45e7-aed7-127fafabdc99 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2146.189903] env[62824]: DEBUG nova.network.neutron [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 20f93b46-5e7e-4256-8cc1-e0a1b16740d5] Successfully updated port: 32e4bbdc-8153-4bb8-9c2d-669a9c40b380 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2146.197669] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146423, 'name': CreateVM_Task, 'duration_secs': 0.423342} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2146.199608] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7a2a51c4-e558-4f5f-b82c-718bc12c1df5] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2146.200373] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2146.200694] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2146.201078] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2146.202480] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dff188e-14ff-440d-a3d7-d2ac588c8f7b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2146.207403] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b5e749f6-8b80-41a5-9ab1-bf073e451a63 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2146.213613] env[62824]: DEBUG oslo_vmware.api [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Waiting for the task: (returnval){ [ 2146.213613] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]523f4f19-60d0-1d15-f437-c4b5cc9380dc" [ 2146.213613] env[62824]: _type = "Task" [ 2146.213613] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2146.223424] env[62824]: DEBUG nova.compute.provider_tree [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2146.235101] env[62824]: DEBUG oslo_vmware.api [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]523f4f19-60d0-1d15-f437-c4b5cc9380dc, 'name': SearchDatastore_Task, 'duration_secs': 0.009413} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2146.236039] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2146.236350] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7a2a51c4-e558-4f5f-b82c-718bc12c1df5] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2146.238892] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2146.238892] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2146.238892] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2146.238892] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-380445db-ebfa-4690-a091-03c9bc6edb40 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2146.247337] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2146.247572] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2146.250981] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cc76b006-5c13-4fbb-8d6e-25f793bc6fa9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2146.252528] env[62824]: INFO nova.compute.manager [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: d067fa97-cedc-4e3d-9be4-d860a79a7723] Took 46.99 seconds to build instance. [ 2146.256838] env[62824]: DEBUG oslo_vmware.api [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Waiting for the task: (returnval){ [ 2146.256838] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]528f3508-e675-3369-afef-8b59d992fe67" [ 2146.256838] env[62824]: _type = "Task" [ 2146.256838] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2146.265403] env[62824]: DEBUG oslo_vmware.api [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]528f3508-e675-3369-afef-8b59d992fe67, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2146.318464] env[62824]: DEBUG oslo_vmware.api [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Task: {'id': task-2146424, 'name': PowerOnVM_Task, 'duration_secs': 0.449187} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2146.318746] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] [instance: dadf590a-7288-4dd2-90de-125fe272f4ba] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2146.318952] env[62824]: INFO nova.compute.manager [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] [instance: dadf590a-7288-4dd2-90de-125fe272f4ba] Took 6.82 seconds to spawn the instance on the hypervisor. [ 2146.319147] env[62824]: DEBUG nova.compute.manager [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] [instance: dadf590a-7288-4dd2-90de-125fe272f4ba] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2146.320048] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-661f37db-1ca6-4f82-9cdc-e277532fc2dc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2146.368415] env[62824]: DEBUG oslo_concurrency.lockutils [req-680e1fdc-dc0d-4151-b6d3-5ed284133942 req-3ad60b72-f1eb-4c3d-8bc9-ae9693b490f8 service nova] Releasing lock "refresh_cache-7a2a51c4-e558-4f5f-b82c-718bc12c1df5" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2146.693152] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Acquiring lock "refresh_cache-20f93b46-5e7e-4256-8cc1-e0a1b16740d5" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2146.693422] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Acquired lock "refresh_cache-20f93b46-5e7e-4256-8cc1-e0a1b16740d5" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2146.693508] env[62824]: DEBUG nova.network.neutron [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 20f93b46-5e7e-4256-8cc1-e0a1b16740d5] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2146.730275] env[62824]: DEBUG nova.scheduler.client.report [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2146.757762] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dcfa7940-3ed8-4656-b14f-b8a1fea20f51 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Lock "d067fa97-cedc-4e3d-9be4-d860a79a7723" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 56.108s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2146.769435] env[62824]: DEBUG oslo_vmware.api [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]528f3508-e675-3369-afef-8b59d992fe67, 'name': SearchDatastore_Task, 'duration_secs': 0.009505} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2146.771362] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9eee4780-c7dd-476f-b350-ed5b5473bc1d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2146.777370] env[62824]: DEBUG oslo_vmware.api [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Waiting for the task: (returnval){ [ 2146.777370] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52f083dd-4212-3e49-87e8-4c18e5dc9493" [ 2146.777370] env[62824]: _type = "Task" [ 2146.777370] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2146.785258] env[62824]: DEBUG oslo_vmware.api [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52f083dd-4212-3e49-87e8-4c18e5dc9493, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2146.841106] env[62824]: INFO nova.compute.manager [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] [instance: dadf590a-7288-4dd2-90de-125fe272f4ba] Took 44.16 seconds to build instance. [ 2147.235608] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.549s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2147.236433] env[62824]: DEBUG nova.compute.manager [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2147.239622] env[62824]: DEBUG nova.network.neutron [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 20f93b46-5e7e-4256-8cc1-e0a1b16740d5] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2147.241569] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.315s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2147.243871] env[62824]: INFO nova.compute.claims [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] [instance: bd7f2ed5-fa08-4a48-9755-60b61791ea0b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2147.263860] env[62824]: DEBUG nova.compute.manager [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 58243465-ba24-4b37-a640-04ab24963230] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2147.291628] env[62824]: DEBUG oslo_vmware.api [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52f083dd-4212-3e49-87e8-4c18e5dc9493, 'name': SearchDatastore_Task, 'duration_secs': 0.01034} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2147.291932] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2147.292266] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 7a2a51c4-e558-4f5f-b82c-718bc12c1df5/7a2a51c4-e558-4f5f-b82c-718bc12c1df5.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2147.292543] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0ff339a1-7a2b-4ec6-a6be-2005a1ae34bd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2147.301260] env[62824]: DEBUG oslo_vmware.api [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Waiting for the task: (returnval){ [ 2147.301260] env[62824]: value = "task-2146425" [ 2147.301260] env[62824]: _type = "Task" [ 2147.301260] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2147.316468] env[62824]: DEBUG oslo_vmware.api [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': task-2146425, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2147.342879] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0f6836d6-2e7d-41c5-ba6f-3d4c9dc474aa tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Acquiring lock "29140542-dc4b-411a-ac10-7d84086eabbb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2147.342879] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0f6836d6-2e7d-41c5-ba6f-3d4c9dc474aa tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Lock "29140542-dc4b-411a-ac10-7d84086eabbb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2147.342879] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0f6836d6-2e7d-41c5-ba6f-3d4c9dc474aa tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Acquiring lock "29140542-dc4b-411a-ac10-7d84086eabbb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2147.342879] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0f6836d6-2e7d-41c5-ba6f-3d4c9dc474aa tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Lock "29140542-dc4b-411a-ac10-7d84086eabbb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2147.342879] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0f6836d6-2e7d-41c5-ba6f-3d4c9dc474aa tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Lock "29140542-dc4b-411a-ac10-7d84086eabbb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2147.344477] env[62824]: INFO nova.compute.manager [None req-0f6836d6-2e7d-41c5-ba6f-3d4c9dc474aa tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 29140542-dc4b-411a-ac10-7d84086eabbb] Terminating instance [ 2147.347943] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d4d2155c-39c0-4e22-b591-16ede0093a90 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Lock "dadf590a-7288-4dd2-90de-125fe272f4ba" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 55.617s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2147.531426] env[62824]: DEBUG nova.network.neutron [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 20f93b46-5e7e-4256-8cc1-e0a1b16740d5] Updating instance_info_cache with network_info: [{"id": "32e4bbdc-8153-4bb8-9c2d-669a9c40b380", "address": "fa:16:3e:c5:54:00", "network": {"id": "91425d93-4190-4e61-ae47-94efd278247d", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1364948114-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "461117110de7467fbf8ada5ee9978069", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c894ab55-c869-4530-9702-cb46d173ce94", "external-id": "nsx-vlan-transportzone-792", "segmentation_id": 792, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap32e4bbdc-81", "ovs_interfaceid": "32e4bbdc-8153-4bb8-9c2d-669a9c40b380", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2147.618814] env[62824]: DEBUG nova.compute.manager [req-aad15606-a014-421b-a7b3-8a9f2a21137d req-3cf91935-6091-469e-979f-5bfeec354467 service nova] [instance: 20f93b46-5e7e-4256-8cc1-e0a1b16740d5] Received event network-vif-plugged-32e4bbdc-8153-4bb8-9c2d-669a9c40b380 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2147.618974] env[62824]: DEBUG oslo_concurrency.lockutils [req-aad15606-a014-421b-a7b3-8a9f2a21137d req-3cf91935-6091-469e-979f-5bfeec354467 service nova] Acquiring lock "20f93b46-5e7e-4256-8cc1-e0a1b16740d5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2147.619319] env[62824]: DEBUG oslo_concurrency.lockutils [req-aad15606-a014-421b-a7b3-8a9f2a21137d req-3cf91935-6091-469e-979f-5bfeec354467 service nova] Lock "20f93b46-5e7e-4256-8cc1-e0a1b16740d5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2147.619518] env[62824]: DEBUG oslo_concurrency.lockutils [req-aad15606-a014-421b-a7b3-8a9f2a21137d req-3cf91935-6091-469e-979f-5bfeec354467 service nova] Lock "20f93b46-5e7e-4256-8cc1-e0a1b16740d5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2147.619773] env[62824]: DEBUG nova.compute.manager [req-aad15606-a014-421b-a7b3-8a9f2a21137d req-3cf91935-6091-469e-979f-5bfeec354467 service nova] [instance: 20f93b46-5e7e-4256-8cc1-e0a1b16740d5] No waiting events found dispatching network-vif-plugged-32e4bbdc-8153-4bb8-9c2d-669a9c40b380 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2147.619839] env[62824]: WARNING nova.compute.manager [req-aad15606-a014-421b-a7b3-8a9f2a21137d req-3cf91935-6091-469e-979f-5bfeec354467 service nova] [instance: 20f93b46-5e7e-4256-8cc1-e0a1b16740d5] Received unexpected event network-vif-plugged-32e4bbdc-8153-4bb8-9c2d-669a9c40b380 for instance with vm_state building and task_state spawning. [ 2147.619994] env[62824]: DEBUG nova.compute.manager [req-aad15606-a014-421b-a7b3-8a9f2a21137d req-3cf91935-6091-469e-979f-5bfeec354467 service nova] [instance: 20f93b46-5e7e-4256-8cc1-e0a1b16740d5] Received event network-changed-32e4bbdc-8153-4bb8-9c2d-669a9c40b380 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2147.620166] env[62824]: DEBUG nova.compute.manager [req-aad15606-a014-421b-a7b3-8a9f2a21137d req-3cf91935-6091-469e-979f-5bfeec354467 service nova] [instance: 20f93b46-5e7e-4256-8cc1-e0a1b16740d5] Refreshing instance network info cache due to event network-changed-32e4bbdc-8153-4bb8-9c2d-669a9c40b380. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2147.620339] env[62824]: DEBUG oslo_concurrency.lockutils [req-aad15606-a014-421b-a7b3-8a9f2a21137d req-3cf91935-6091-469e-979f-5bfeec354467 service nova] Acquiring lock "refresh_cache-20f93b46-5e7e-4256-8cc1-e0a1b16740d5" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2147.754029] env[62824]: DEBUG nova.compute.utils [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2147.756295] env[62824]: DEBUG nova.compute.manager [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2147.756489] env[62824]: DEBUG nova.network.neutron [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2147.785349] env[62824]: DEBUG oslo_concurrency.lockutils [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2147.810908] env[62824]: DEBUG oslo_vmware.api [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': task-2146425, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.466117} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2147.812320] env[62824]: DEBUG nova.policy [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b989eb89498b45dab6604116870d2fa1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c987c5f844be42c99cc9d57a9ca84c31', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2147.813795] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 7a2a51c4-e558-4f5f-b82c-718bc12c1df5/7a2a51c4-e558-4f5f-b82c-718bc12c1df5.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2147.814031] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7a2a51c4-e558-4f5f-b82c-718bc12c1df5] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2147.814294] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8df1a866-b3b9-4411-8c69-2acd00b58ba1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2147.822239] env[62824]: DEBUG oslo_vmware.api [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Waiting for the task: (returnval){ [ 2147.822239] env[62824]: value = "task-2146426" [ 2147.822239] env[62824]: _type = "Task" [ 2147.822239] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2147.830304] env[62824]: DEBUG oslo_vmware.api [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': task-2146426, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2147.848866] env[62824]: DEBUG nova.compute.manager [None req-0f6836d6-2e7d-41c5-ba6f-3d4c9dc474aa tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 29140542-dc4b-411a-ac10-7d84086eabbb] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2147.849124] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-0f6836d6-2e7d-41c5-ba6f-3d4c9dc474aa tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 29140542-dc4b-411a-ac10-7d84086eabbb] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2147.850107] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7033860-65d6-45ba-b732-e3f759aac816 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2147.857772] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f6836d6-2e7d-41c5-ba6f-3d4c9dc474aa tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 29140542-dc4b-411a-ac10-7d84086eabbb] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2147.858020] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5d2291fc-9f66-4ddb-9b25-433d93eb84aa {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2147.864549] env[62824]: DEBUG oslo_vmware.api [None req-0f6836d6-2e7d-41c5-ba6f-3d4c9dc474aa tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Waiting for the task: (returnval){ [ 2147.864549] env[62824]: value = "task-2146427" [ 2147.864549] env[62824]: _type = "Task" [ 2147.864549] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2147.872430] env[62824]: DEBUG oslo_vmware.api [None req-0f6836d6-2e7d-41c5-ba6f-3d4c9dc474aa tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Task: {'id': task-2146427, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2148.034557] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Releasing lock "refresh_cache-20f93b46-5e7e-4256-8cc1-e0a1b16740d5" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2148.034887] env[62824]: DEBUG nova.compute.manager [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 20f93b46-5e7e-4256-8cc1-e0a1b16740d5] Instance network_info: |[{"id": "32e4bbdc-8153-4bb8-9c2d-669a9c40b380", "address": "fa:16:3e:c5:54:00", "network": {"id": "91425d93-4190-4e61-ae47-94efd278247d", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1364948114-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "461117110de7467fbf8ada5ee9978069", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c894ab55-c869-4530-9702-cb46d173ce94", "external-id": "nsx-vlan-transportzone-792", "segmentation_id": 792, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap32e4bbdc-81", "ovs_interfaceid": "32e4bbdc-8153-4bb8-9c2d-669a9c40b380", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2148.035219] env[62824]: DEBUG oslo_concurrency.lockutils [req-aad15606-a014-421b-a7b3-8a9f2a21137d req-3cf91935-6091-469e-979f-5bfeec354467 service nova] Acquired lock "refresh_cache-20f93b46-5e7e-4256-8cc1-e0a1b16740d5" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2148.035415] env[62824]: DEBUG nova.network.neutron [req-aad15606-a014-421b-a7b3-8a9f2a21137d req-3cf91935-6091-469e-979f-5bfeec354467 service nova] [instance: 20f93b46-5e7e-4256-8cc1-e0a1b16740d5] Refreshing network info cache for port 32e4bbdc-8153-4bb8-9c2d-669a9c40b380 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2148.036950] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 20f93b46-5e7e-4256-8cc1-e0a1b16740d5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c5:54:00', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c894ab55-c869-4530-9702-cb46d173ce94', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '32e4bbdc-8153-4bb8-9c2d-669a9c40b380', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2148.045268] env[62824]: DEBUG oslo.service.loopingcall [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2148.046052] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 20f93b46-5e7e-4256-8cc1-e0a1b16740d5] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2148.046052] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e3583a02-22ca-409d-bc22-285a9a7b6a93 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2148.071949] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2148.071949] env[62824]: value = "task-2146428" [ 2148.071949] env[62824]: _type = "Task" [ 2148.071949] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2148.081505] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146428, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2148.186387] env[62824]: DEBUG nova.network.neutron [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Successfully created port: e285a842-d54b-4614-84ea-239e34f5b41f {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2148.256945] env[62824]: DEBUG nova.compute.manager [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2148.335558] env[62824]: DEBUG oslo_vmware.api [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': task-2146426, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.148193} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2148.338458] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7a2a51c4-e558-4f5f-b82c-718bc12c1df5] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2148.339976] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a537bb5b-281b-4f2d-9c26-83b8bc6b1475 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2148.370128] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7a2a51c4-e558-4f5f-b82c-718bc12c1df5] Reconfiguring VM instance instance-0000004b to attach disk [datastore1] 7a2a51c4-e558-4f5f-b82c-718bc12c1df5/7a2a51c4-e558-4f5f-b82c-718bc12c1df5.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2148.374178] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-66ad792b-f470-4b4b-8a11-499ac0ee9493 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2148.400598] env[62824]: DEBUG oslo_vmware.api [None req-0f6836d6-2e7d-41c5-ba6f-3d4c9dc474aa tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Task: {'id': task-2146427, 'name': PowerOffVM_Task, 'duration_secs': 0.314069} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2148.402142] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f6836d6-2e7d-41c5-ba6f-3d4c9dc474aa tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 29140542-dc4b-411a-ac10-7d84086eabbb] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2148.402469] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-0f6836d6-2e7d-41c5-ba6f-3d4c9dc474aa tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 29140542-dc4b-411a-ac10-7d84086eabbb] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2148.402899] env[62824]: DEBUG oslo_vmware.api [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Waiting for the task: (returnval){ [ 2148.402899] env[62824]: value = "task-2146429" [ 2148.402899] env[62824]: _type = "Task" [ 2148.402899] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2148.403340] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-62ffd16f-9bb1-4843-afbc-d05fe65a7ed3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2148.418235] env[62824]: DEBUG oslo_vmware.api [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': task-2146429, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2148.489018] env[62824]: DEBUG nova.compute.manager [None req-ad975e62-21af-48fd-9c71-f29c163411b9 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] [instance: dadf590a-7288-4dd2-90de-125fe272f4ba] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2148.489018] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-415fa1f6-215f-4b41-83b3-f06bdfa422ad {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2148.571154] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ae8aad58-b430-4dd1-8bc0-b6bcc01e69fc tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Acquiring lock "dadf590a-7288-4dd2-90de-125fe272f4ba" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2148.571154] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ae8aad58-b430-4dd1-8bc0-b6bcc01e69fc tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Lock "dadf590a-7288-4dd2-90de-125fe272f4ba" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2148.571154] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ae8aad58-b430-4dd1-8bc0-b6bcc01e69fc tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Acquiring lock "dadf590a-7288-4dd2-90de-125fe272f4ba-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2148.571154] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ae8aad58-b430-4dd1-8bc0-b6bcc01e69fc tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Lock "dadf590a-7288-4dd2-90de-125fe272f4ba-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2148.571154] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ae8aad58-b430-4dd1-8bc0-b6bcc01e69fc tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Lock "dadf590a-7288-4dd2-90de-125fe272f4ba-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2148.576025] env[62824]: INFO nova.compute.manager [None req-ae8aad58-b430-4dd1-8bc0-b6bcc01e69fc tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] [instance: dadf590a-7288-4dd2-90de-125fe272f4ba] Terminating instance [ 2148.583532] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-0f6836d6-2e7d-41c5-ba6f-3d4c9dc474aa tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 29140542-dc4b-411a-ac10-7d84086eabbb] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2148.583895] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-0f6836d6-2e7d-41c5-ba6f-3d4c9dc474aa tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 29140542-dc4b-411a-ac10-7d84086eabbb] Deleting contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2148.584201] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-0f6836d6-2e7d-41c5-ba6f-3d4c9dc474aa tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Deleting the datastore file [datastore2] 29140542-dc4b-411a-ac10-7d84086eabbb {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2148.587747] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ffc54e61-a201-4930-b85d-d5886c9f655f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2148.595180] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146428, 'name': CreateVM_Task, 'duration_secs': 0.372761} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2148.599020] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 20f93b46-5e7e-4256-8cc1-e0a1b16740d5] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2148.599020] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2148.599020] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2148.599020] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2148.599020] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-80d93f46-5460-4b7f-9bad-5299d9739bc2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2148.604141] env[62824]: DEBUG oslo_vmware.api [None req-0f6836d6-2e7d-41c5-ba6f-3d4c9dc474aa tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Waiting for the task: (returnval){ [ 2148.604141] env[62824]: value = "task-2146431" [ 2148.604141] env[62824]: _type = "Task" [ 2148.604141] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2148.609735] env[62824]: DEBUG oslo_vmware.api [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Waiting for the task: (returnval){ [ 2148.609735] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52a5f88c-405c-1b38-9306-0106f5e5fd2c" [ 2148.609735] env[62824]: _type = "Task" [ 2148.609735] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2148.616630] env[62824]: DEBUG oslo_vmware.api [None req-0f6836d6-2e7d-41c5-ba6f-3d4c9dc474aa tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Task: {'id': task-2146431, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2148.628714] env[62824]: DEBUG oslo_vmware.api [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52a5f88c-405c-1b38-9306-0106f5e5fd2c, 'name': SearchDatastore_Task, 'duration_secs': 0.010469} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2148.628714] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2148.628714] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 20f93b46-5e7e-4256-8cc1-e0a1b16740d5] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2148.628714] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2148.628714] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2148.628714] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2148.628714] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c68792a8-8d1e-4627-bc26-a18ab0719472 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2148.639544] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2148.639932] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2148.641193] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3a066b4f-9170-4a9b-b8d7-8d7df930de93 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2148.651137] env[62824]: DEBUG oslo_vmware.api [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Waiting for the task: (returnval){ [ 2148.651137] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]529ba59f-89ff-5167-3c90-42400cb120c9" [ 2148.651137] env[62824]: _type = "Task" [ 2148.651137] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2148.661197] env[62824]: DEBUG oslo_vmware.api [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]529ba59f-89ff-5167-3c90-42400cb120c9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2148.814026] env[62824]: DEBUG nova.network.neutron [req-aad15606-a014-421b-a7b3-8a9f2a21137d req-3cf91935-6091-469e-979f-5bfeec354467 service nova] [instance: 20f93b46-5e7e-4256-8cc1-e0a1b16740d5] Updated VIF entry in instance network info cache for port 32e4bbdc-8153-4bb8-9c2d-669a9c40b380. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2148.814026] env[62824]: DEBUG nova.network.neutron [req-aad15606-a014-421b-a7b3-8a9f2a21137d req-3cf91935-6091-469e-979f-5bfeec354467 service nova] [instance: 20f93b46-5e7e-4256-8cc1-e0a1b16740d5] Updating instance_info_cache with network_info: [{"id": "32e4bbdc-8153-4bb8-9c2d-669a9c40b380", "address": "fa:16:3e:c5:54:00", "network": {"id": "91425d93-4190-4e61-ae47-94efd278247d", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1364948114-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "461117110de7467fbf8ada5ee9978069", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c894ab55-c869-4530-9702-cb46d173ce94", "external-id": "nsx-vlan-transportzone-792", "segmentation_id": 792, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap32e4bbdc-81", "ovs_interfaceid": "32e4bbdc-8153-4bb8-9c2d-669a9c40b380", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2148.855227] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61f95cc3-e583-4d6c-8859-be48782fad65 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2148.863381] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d569083-ce97-4874-b019-73dd1b6d6d98 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2148.895878] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5df7117-5818-4251-91bf-ac7988fa4f60 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2148.903891] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10253965-4d15-4aa7-8af5-9d8cc9d1e11b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2148.915847] env[62824]: DEBUG oslo_vmware.api [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': task-2146429, 'name': ReconfigVM_Task, 'duration_secs': 0.291524} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2148.923953] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7a2a51c4-e558-4f5f-b82c-718bc12c1df5] Reconfigured VM instance instance-0000004b to attach disk [datastore1] 7a2a51c4-e558-4f5f-b82c-718bc12c1df5/7a2a51c4-e558-4f5f-b82c-718bc12c1df5.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2148.924635] env[62824]: DEBUG nova.compute.provider_tree [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2148.925962] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-64f9ba7f-0d6e-4b81-aeb4-21ab0df7b5d4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2148.933532] env[62824]: DEBUG oslo_vmware.api [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Waiting for the task: (returnval){ [ 2148.933532] env[62824]: value = "task-2146432" [ 2148.933532] env[62824]: _type = "Task" [ 2148.933532] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2148.946787] env[62824]: DEBUG oslo_vmware.api [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': task-2146432, 'name': Rename_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2149.000539] env[62824]: INFO nova.compute.manager [None req-ad975e62-21af-48fd-9c71-f29c163411b9 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] [instance: dadf590a-7288-4dd2-90de-125fe272f4ba] instance snapshotting [ 2149.001385] env[62824]: DEBUG nova.objects.instance [None req-ad975e62-21af-48fd-9c71-f29c163411b9 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Lazy-loading 'flavor' on Instance uuid dadf590a-7288-4dd2-90de-125fe272f4ba {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2149.087679] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ae8aad58-b430-4dd1-8bc0-b6bcc01e69fc tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Acquiring lock "refresh_cache-dadf590a-7288-4dd2-90de-125fe272f4ba" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2149.087921] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ae8aad58-b430-4dd1-8bc0-b6bcc01e69fc tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Acquired lock "refresh_cache-dadf590a-7288-4dd2-90de-125fe272f4ba" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2149.088122] env[62824]: DEBUG nova.network.neutron [None req-ae8aad58-b430-4dd1-8bc0-b6bcc01e69fc tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] [instance: dadf590a-7288-4dd2-90de-125fe272f4ba] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2149.114669] env[62824]: DEBUG oslo_vmware.api [None req-0f6836d6-2e7d-41c5-ba6f-3d4c9dc474aa tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Task: {'id': task-2146431, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.162834} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2149.114917] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-0f6836d6-2e7d-41c5-ba6f-3d4c9dc474aa tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2149.115116] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-0f6836d6-2e7d-41c5-ba6f-3d4c9dc474aa tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 29140542-dc4b-411a-ac10-7d84086eabbb] Deleted contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2149.115328] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-0f6836d6-2e7d-41c5-ba6f-3d4c9dc474aa tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 29140542-dc4b-411a-ac10-7d84086eabbb] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2149.115508] env[62824]: INFO nova.compute.manager [None req-0f6836d6-2e7d-41c5-ba6f-3d4c9dc474aa tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 29140542-dc4b-411a-ac10-7d84086eabbb] Took 1.27 seconds to destroy the instance on the hypervisor. [ 2149.115745] env[62824]: DEBUG oslo.service.loopingcall [None req-0f6836d6-2e7d-41c5-ba6f-3d4c9dc474aa tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2149.115928] env[62824]: DEBUG nova.compute.manager [-] [instance: 29140542-dc4b-411a-ac10-7d84086eabbb] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2149.116037] env[62824]: DEBUG nova.network.neutron [-] [instance: 29140542-dc4b-411a-ac10-7d84086eabbb] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2149.161322] env[62824]: DEBUG oslo_vmware.api [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]529ba59f-89ff-5167-3c90-42400cb120c9, 'name': SearchDatastore_Task, 'duration_secs': 0.018108} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2149.162223] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-494e758f-80df-4443-854d-6c43c89962c5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2149.168978] env[62824]: DEBUG oslo_vmware.api [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Waiting for the task: (returnval){ [ 2149.168978] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52f6eab0-ef53-7ad4-7e71-c9f76f3d45ce" [ 2149.168978] env[62824]: _type = "Task" [ 2149.168978] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2149.177698] env[62824]: DEBUG oslo_vmware.api [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52f6eab0-ef53-7ad4-7e71-c9f76f3d45ce, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2149.273511] env[62824]: DEBUG nova.compute.manager [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2149.301112] env[62824]: DEBUG nova.virt.hardware [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2149.301582] env[62824]: DEBUG nova.virt.hardware [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2149.301894] env[62824]: DEBUG nova.virt.hardware [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2149.302222] env[62824]: DEBUG nova.virt.hardware [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2149.306017] env[62824]: DEBUG nova.virt.hardware [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2149.306017] env[62824]: DEBUG nova.virt.hardware [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2149.306017] env[62824]: DEBUG nova.virt.hardware [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2149.306017] env[62824]: DEBUG nova.virt.hardware [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2149.306017] env[62824]: DEBUG nova.virt.hardware [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2149.306017] env[62824]: DEBUG nova.virt.hardware [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2149.306017] env[62824]: DEBUG nova.virt.hardware [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2149.306017] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1f9db98-219c-45a3-8edf-ecd5f6c312d9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2149.315115] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-285d95ff-fa36-4c07-9688-9aa123092907 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2149.319495] env[62824]: DEBUG oslo_concurrency.lockutils [req-aad15606-a014-421b-a7b3-8a9f2a21137d req-3cf91935-6091-469e-979f-5bfeec354467 service nova] Releasing lock "refresh_cache-20f93b46-5e7e-4256-8cc1-e0a1b16740d5" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2149.402168] env[62824]: DEBUG nova.compute.manager [req-9eb646d4-4b34-4ca7-b63d-a007e5da591e req-30bf7be9-e0b4-4a13-96bc-c0012c0c9f88 service nova] [instance: 29140542-dc4b-411a-ac10-7d84086eabbb] Received event network-vif-deleted-8ad74d1d-5091-4557-bb57-ba65d9b98b45 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2149.402303] env[62824]: INFO nova.compute.manager [req-9eb646d4-4b34-4ca7-b63d-a007e5da591e req-30bf7be9-e0b4-4a13-96bc-c0012c0c9f88 service nova] [instance: 29140542-dc4b-411a-ac10-7d84086eabbb] Neutron deleted interface 8ad74d1d-5091-4557-bb57-ba65d9b98b45; detaching it from the instance and deleting it from the info cache [ 2149.402480] env[62824]: DEBUG nova.network.neutron [req-9eb646d4-4b34-4ca7-b63d-a007e5da591e req-30bf7be9-e0b4-4a13-96bc-c0012c0c9f88 service nova] [instance: 29140542-dc4b-411a-ac10-7d84086eabbb] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2149.429540] env[62824]: DEBUG nova.scheduler.client.report [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2149.443822] env[62824]: DEBUG oslo_vmware.api [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': task-2146432, 'name': Rename_Task, 'duration_secs': 0.133031} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2149.444132] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7a2a51c4-e558-4f5f-b82c-718bc12c1df5] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2149.444377] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c0bf5b14-8cec-4886-a73e-493724ad9359 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2149.451280] env[62824]: DEBUG oslo_vmware.api [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Waiting for the task: (returnval){ [ 2149.451280] env[62824]: value = "task-2146433" [ 2149.451280] env[62824]: _type = "Task" [ 2149.451280] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2149.458987] env[62824]: DEBUG oslo_vmware.api [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': task-2146433, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2149.507435] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e40c1199-d0e8-44b8-8355-f4c7583d420e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2149.524308] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a826f17-2b72-4b4a-897e-d82d7e54e6cd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2149.607785] env[62824]: DEBUG nova.network.neutron [None req-ae8aad58-b430-4dd1-8bc0-b6bcc01e69fc tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] [instance: dadf590a-7288-4dd2-90de-125fe272f4ba] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2149.658041] env[62824]: DEBUG nova.network.neutron [None req-ae8aad58-b430-4dd1-8bc0-b6bcc01e69fc tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] [instance: dadf590a-7288-4dd2-90de-125fe272f4ba] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2149.679453] env[62824]: DEBUG oslo_vmware.api [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52f6eab0-ef53-7ad4-7e71-c9f76f3d45ce, 'name': SearchDatastore_Task, 'duration_secs': 0.008946} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2149.679722] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2149.679979] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 20f93b46-5e7e-4256-8cc1-e0a1b16740d5/20f93b46-5e7e-4256-8cc1-e0a1b16740d5.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2149.680247] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-54bdbf0f-bda6-4fa5-bf73-f83121eeaa29 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2149.688088] env[62824]: DEBUG oslo_vmware.api [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Waiting for the task: (returnval){ [ 2149.688088] env[62824]: value = "task-2146434" [ 2149.688088] env[62824]: _type = "Task" [ 2149.688088] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2149.695677] env[62824]: DEBUG oslo_vmware.api [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': task-2146434, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2149.879917] env[62824]: DEBUG nova.network.neutron [-] [instance: 29140542-dc4b-411a-ac10-7d84086eabbb] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2149.904677] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0abd8834-d79f-434b-b7e8-1d8ceaa52510 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2149.915102] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17b263eb-a6b0-4270-b5c1-0493aba47d79 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2149.934343] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.693s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2149.934885] env[62824]: DEBUG nova.compute.manager [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] [instance: bd7f2ed5-fa08-4a48-9755-60b61791ea0b] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2149.937604] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7beba819-8d40-4b17-972a-05048027ddd5 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 28.135s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2149.939486] env[62824]: DEBUG nova.objects.instance [None req-7beba819-8d40-4b17-972a-05048027ddd5 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Lazy-loading 'resources' on Instance uuid 02f34565-6675-4c79-ac47-b131ceba9df8 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2149.959084] env[62824]: DEBUG nova.compute.manager [req-9eb646d4-4b34-4ca7-b63d-a007e5da591e req-30bf7be9-e0b4-4a13-96bc-c0012c0c9f88 service nova] [instance: 29140542-dc4b-411a-ac10-7d84086eabbb] Detach interface failed, port_id=8ad74d1d-5091-4557-bb57-ba65d9b98b45, reason: Instance 29140542-dc4b-411a-ac10-7d84086eabbb could not be found. {{(pid=62824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11482}} [ 2149.970300] env[62824]: DEBUG oslo_vmware.api [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': task-2146433, 'name': PowerOnVM_Task, 'duration_secs': 0.474499} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2149.970580] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7a2a51c4-e558-4f5f-b82c-718bc12c1df5] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2149.970580] env[62824]: INFO nova.compute.manager [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7a2a51c4-e558-4f5f-b82c-718bc12c1df5] Took 6.87 seconds to spawn the instance on the hypervisor. [ 2149.970799] env[62824]: DEBUG nova.compute.manager [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7a2a51c4-e558-4f5f-b82c-718bc12c1df5] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2149.971684] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3ec5406-7aab-4739-8b47-24107f004af6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2150.026912] env[62824]: DEBUG nova.compute.manager [req-032cca07-7059-409c-ac7b-87dcc7b2c3dc req-f21e791c-1b01-4028-bf1f-582467fbefaf service nova] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Received event network-vif-plugged-e285a842-d54b-4614-84ea-239e34f5b41f {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2150.027740] env[62824]: DEBUG oslo_concurrency.lockutils [req-032cca07-7059-409c-ac7b-87dcc7b2c3dc req-f21e791c-1b01-4028-bf1f-582467fbefaf service nova] Acquiring lock "fc77c44a-180c-46ff-9690-9072c6213c91-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2150.027916] env[62824]: DEBUG oslo_concurrency.lockutils [req-032cca07-7059-409c-ac7b-87dcc7b2c3dc req-f21e791c-1b01-4028-bf1f-582467fbefaf service nova] Lock "fc77c44a-180c-46ff-9690-9072c6213c91-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2150.028129] env[62824]: DEBUG oslo_concurrency.lockutils [req-032cca07-7059-409c-ac7b-87dcc7b2c3dc req-f21e791c-1b01-4028-bf1f-582467fbefaf service nova] Lock "fc77c44a-180c-46ff-9690-9072c6213c91-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2150.028359] env[62824]: DEBUG nova.compute.manager [req-032cca07-7059-409c-ac7b-87dcc7b2c3dc req-f21e791c-1b01-4028-bf1f-582467fbefaf service nova] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] No waiting events found dispatching network-vif-plugged-e285a842-d54b-4614-84ea-239e34f5b41f {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2150.028675] env[62824]: WARNING nova.compute.manager [req-032cca07-7059-409c-ac7b-87dcc7b2c3dc req-f21e791c-1b01-4028-bf1f-582467fbefaf service nova] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Received unexpected event network-vif-plugged-e285a842-d54b-4614-84ea-239e34f5b41f for instance with vm_state building and task_state spawning. [ 2150.033497] env[62824]: DEBUG nova.compute.manager [None req-ad975e62-21af-48fd-9c71-f29c163411b9 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] [instance: dadf590a-7288-4dd2-90de-125fe272f4ba] Instance disappeared during snapshot {{(pid=62824) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4595}} [ 2150.138041] env[62824]: DEBUG nova.network.neutron [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Successfully updated port: e285a842-d54b-4614-84ea-239e34f5b41f {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2150.161450] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ae8aad58-b430-4dd1-8bc0-b6bcc01e69fc tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Releasing lock "refresh_cache-dadf590a-7288-4dd2-90de-125fe272f4ba" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2150.161933] env[62824]: DEBUG nova.compute.manager [None req-ae8aad58-b430-4dd1-8bc0-b6bcc01e69fc tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] [instance: dadf590a-7288-4dd2-90de-125fe272f4ba] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2150.162166] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-ae8aad58-b430-4dd1-8bc0-b6bcc01e69fc tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] [instance: dadf590a-7288-4dd2-90de-125fe272f4ba] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2150.163202] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-648356b4-a061-47a0-a623-cddecea0390f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2150.172877] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae8aad58-b430-4dd1-8bc0-b6bcc01e69fc tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] [instance: dadf590a-7288-4dd2-90de-125fe272f4ba] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2150.173264] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-03a112ff-5433-48fa-8b77-21d7e0746212 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2150.180443] env[62824]: DEBUG oslo_vmware.api [None req-ae8aad58-b430-4dd1-8bc0-b6bcc01e69fc tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Waiting for the task: (returnval){ [ 2150.180443] env[62824]: value = "task-2146435" [ 2150.180443] env[62824]: _type = "Task" [ 2150.180443] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2150.185835] env[62824]: DEBUG nova.compute.manager [None req-ad975e62-21af-48fd-9c71-f29c163411b9 tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] [instance: dadf590a-7288-4dd2-90de-125fe272f4ba] Found 0 images (rotation: 2) {{(pid=62824) _rotate_backups /opt/stack/nova/nova/compute/manager.py:5017}} [ 2150.192667] env[62824]: DEBUG oslo_vmware.api [None req-ae8aad58-b430-4dd1-8bc0-b6bcc01e69fc tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Task: {'id': task-2146435, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2150.202583] env[62824]: DEBUG oslo_vmware.api [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': task-2146434, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2150.382968] env[62824]: INFO nova.compute.manager [-] [instance: 29140542-dc4b-411a-ac10-7d84086eabbb] Took 1.27 seconds to deallocate network for instance. [ 2150.441686] env[62824]: DEBUG nova.compute.utils [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2150.446041] env[62824]: DEBUG nova.compute.manager [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] [instance: bd7f2ed5-fa08-4a48-9755-60b61791ea0b] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2150.446041] env[62824]: DEBUG nova.network.neutron [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] [instance: bd7f2ed5-fa08-4a48-9755-60b61791ea0b] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2150.493047] env[62824]: INFO nova.compute.manager [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7a2a51c4-e558-4f5f-b82c-718bc12c1df5] Took 41.28 seconds to build instance. [ 2150.498973] env[62824]: DEBUG nova.policy [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '34ca3d531ecd4bd293977c255fcfcba8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '958e2482b37342959d733f14c9242618', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2150.641290] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquiring lock "refresh_cache-fc77c44a-180c-46ff-9690-9072c6213c91" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2150.641290] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquired lock "refresh_cache-fc77c44a-180c-46ff-9690-9072c6213c91" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2150.641421] env[62824]: DEBUG nova.network.neutron [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2150.692029] env[62824]: DEBUG oslo_vmware.api [None req-ae8aad58-b430-4dd1-8bc0-b6bcc01e69fc tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Task: {'id': task-2146435, 'name': PowerOffVM_Task, 'duration_secs': 0.199294} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2150.692445] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae8aad58-b430-4dd1-8bc0-b6bcc01e69fc tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] [instance: dadf590a-7288-4dd2-90de-125fe272f4ba] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2150.692527] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-ae8aad58-b430-4dd1-8bc0-b6bcc01e69fc tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] [instance: dadf590a-7288-4dd2-90de-125fe272f4ba] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2150.693180] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-37cbfe70-5fbb-4717-8106-768af5769476 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2150.705576] env[62824]: DEBUG oslo_vmware.api [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': task-2146434, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.55236} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2150.705872] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 20f93b46-5e7e-4256-8cc1-e0a1b16740d5/20f93b46-5e7e-4256-8cc1-e0a1b16740d5.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2150.706029] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 20f93b46-5e7e-4256-8cc1-e0a1b16740d5] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2150.706285] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c344aacf-ff7c-488d-a3b0-31db196eba9a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2150.711793] env[62824]: DEBUG oslo_vmware.api [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Waiting for the task: (returnval){ [ 2150.711793] env[62824]: value = "task-2146437" [ 2150.711793] env[62824]: _type = "Task" [ 2150.711793] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2150.721918] env[62824]: DEBUG oslo_vmware.api [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': task-2146437, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2150.724504] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-ae8aad58-b430-4dd1-8bc0-b6bcc01e69fc tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] [instance: dadf590a-7288-4dd2-90de-125fe272f4ba] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2150.724707] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-ae8aad58-b430-4dd1-8bc0-b6bcc01e69fc tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] [instance: dadf590a-7288-4dd2-90de-125fe272f4ba] Deleting contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2150.724886] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae8aad58-b430-4dd1-8bc0-b6bcc01e69fc tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Deleting the datastore file [datastore1] dadf590a-7288-4dd2-90de-125fe272f4ba {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2150.725175] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bfdea3b3-a51b-4c76-b368-e5438793237c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2150.730981] env[62824]: DEBUG oslo_vmware.api [None req-ae8aad58-b430-4dd1-8bc0-b6bcc01e69fc tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Waiting for the task: (returnval){ [ 2150.730981] env[62824]: value = "task-2146438" [ 2150.730981] env[62824]: _type = "Task" [ 2150.730981] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2150.742052] env[62824]: DEBUG oslo_vmware.api [None req-ae8aad58-b430-4dd1-8bc0-b6bcc01e69fc tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Task: {'id': task-2146438, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2150.884457] env[62824]: DEBUG nova.network.neutron [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] [instance: bd7f2ed5-fa08-4a48-9755-60b61791ea0b] Successfully created port: 01afe8c7-58f8-4fa9-97a6-4767b34ccb66 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2150.890024] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0f6836d6-2e7d-41c5-ba6f-3d4c9dc474aa tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2150.891634] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5af3ffc5-4c7a-4b68-9b58-ef0e9695a3f2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2150.899109] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf013c8c-899d-456a-aa1c-66b6bcfa6e05 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2150.931604] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2087097-f9f2-4127-9888-b192784747b8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2150.939323] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3373d1cd-3381-4f2b-a1d3-25d351a95bae {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2150.948777] env[62824]: DEBUG nova.compute.manager [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] [instance: bd7f2ed5-fa08-4a48-9755-60b61791ea0b] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2150.962433] env[62824]: DEBUG nova.compute.provider_tree [None req-7beba819-8d40-4b17-972a-05048027ddd5 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2150.995220] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Lock "7a2a51c4-e558-4f5f-b82c-718bc12c1df5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 57.987s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2151.124211] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2151.124843] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2151.183329] env[62824]: DEBUG nova.network.neutron [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2151.221389] env[62824]: DEBUG oslo_vmware.api [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': task-2146437, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.275599} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2151.223841] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 20f93b46-5e7e-4256-8cc1-e0a1b16740d5] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2151.224817] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c370f80e-d072-4cf3-a5f4-e3c12990987e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2151.248548] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 20f93b46-5e7e-4256-8cc1-e0a1b16740d5] Reconfiguring VM instance instance-0000004c to attach disk [datastore1] 20f93b46-5e7e-4256-8cc1-e0a1b16740d5/20f93b46-5e7e-4256-8cc1-e0a1b16740d5.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2151.251963] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9ab85897-db9a-43b2-95b0-63eae72a6ea0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2151.273856] env[62824]: DEBUG oslo_vmware.api [None req-ae8aad58-b430-4dd1-8bc0-b6bcc01e69fc tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Task: {'id': task-2146438, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.252842} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2151.275243] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae8aad58-b430-4dd1-8bc0-b6bcc01e69fc tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2151.275587] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-ae8aad58-b430-4dd1-8bc0-b6bcc01e69fc tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] [instance: dadf590a-7288-4dd2-90de-125fe272f4ba] Deleted contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2151.275784] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-ae8aad58-b430-4dd1-8bc0-b6bcc01e69fc tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] [instance: dadf590a-7288-4dd2-90de-125fe272f4ba] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2151.275988] env[62824]: INFO nova.compute.manager [None req-ae8aad58-b430-4dd1-8bc0-b6bcc01e69fc tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] [instance: dadf590a-7288-4dd2-90de-125fe272f4ba] Took 1.11 seconds to destroy the instance on the hypervisor. [ 2151.276258] env[62824]: DEBUG oslo.service.loopingcall [None req-ae8aad58-b430-4dd1-8bc0-b6bcc01e69fc tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2151.276518] env[62824]: DEBUG oslo_vmware.api [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Waiting for the task: (returnval){ [ 2151.276518] env[62824]: value = "task-2146439" [ 2151.276518] env[62824]: _type = "Task" [ 2151.276518] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2151.276702] env[62824]: DEBUG nova.compute.manager [-] [instance: dadf590a-7288-4dd2-90de-125fe272f4ba] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2151.276797] env[62824]: DEBUG nova.network.neutron [-] [instance: dadf590a-7288-4dd2-90de-125fe272f4ba] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2151.286415] env[62824]: DEBUG oslo_vmware.api [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': task-2146439, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2151.298964] env[62824]: DEBUG nova.network.neutron [-] [instance: dadf590a-7288-4dd2-90de-125fe272f4ba] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2151.358062] env[62824]: DEBUG nova.network.neutron [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Updating instance_info_cache with network_info: [{"id": "e285a842-d54b-4614-84ea-239e34f5b41f", "address": "fa:16:3e:9d:17:cd", "network": {"id": "aa1c09e5-8d72-43ad-a903-a0f5e711da80", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1643616572-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c987c5f844be42c99cc9d57a9ca84c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9bc2632-36f9-4912-8782-8bbb789f909d", "external-id": "nsx-vlan-transportzone-897", "segmentation_id": 897, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape285a842-d5", "ovs_interfaceid": "e285a842-d54b-4614-84ea-239e34f5b41f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2151.468550] env[62824]: DEBUG nova.scheduler.client.report [None req-7beba819-8d40-4b17-972a-05048027ddd5 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2151.634674] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2151.634884] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Starting heal instance info cache {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10445}} [ 2151.789615] env[62824]: DEBUG oslo_vmware.api [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': task-2146439, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2151.803131] env[62824]: DEBUG nova.network.neutron [-] [instance: dadf590a-7288-4dd2-90de-125fe272f4ba] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2151.864444] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Releasing lock "refresh_cache-fc77c44a-180c-46ff-9690-9072c6213c91" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2151.864773] env[62824]: DEBUG nova.compute.manager [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Instance network_info: |[{"id": "e285a842-d54b-4614-84ea-239e34f5b41f", "address": "fa:16:3e:9d:17:cd", "network": {"id": "aa1c09e5-8d72-43ad-a903-a0f5e711da80", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1643616572-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c987c5f844be42c99cc9d57a9ca84c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9bc2632-36f9-4912-8782-8bbb789f909d", "external-id": "nsx-vlan-transportzone-897", "segmentation_id": 897, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape285a842-d5", "ovs_interfaceid": "e285a842-d54b-4614-84ea-239e34f5b41f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2151.865282] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9d:17:cd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c9bc2632-36f9-4912-8782-8bbb789f909d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e285a842-d54b-4614-84ea-239e34f5b41f', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2151.873241] env[62824]: DEBUG oslo.service.loopingcall [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2151.873448] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2151.873672] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d81d2547-5a10-4330-8cc1-61354ea1aa80 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2151.894724] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2151.894724] env[62824]: value = "task-2146440" [ 2151.894724] env[62824]: _type = "Task" [ 2151.894724] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2151.902547] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146440, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2151.974375] env[62824]: DEBUG nova.compute.manager [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] [instance: bd7f2ed5-fa08-4a48-9755-60b61791ea0b] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2151.977800] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7beba819-8d40-4b17-972a-05048027ddd5 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.040s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2151.980897] env[62824]: DEBUG oslo_concurrency.lockutils [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.569s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2151.983252] env[62824]: INFO nova.compute.claims [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 2afd1f18-234c-40b5-9ede-7413ad30dafe] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2152.012254] env[62824]: DEBUG nova.virt.hardware [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2152.012523] env[62824]: DEBUG nova.virt.hardware [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2152.012684] env[62824]: DEBUG nova.virt.hardware [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2152.012859] env[62824]: DEBUG nova.virt.hardware [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2152.013014] env[62824]: DEBUG nova.virt.hardware [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2152.013217] env[62824]: DEBUG nova.virt.hardware [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2152.013497] env[62824]: DEBUG nova.virt.hardware [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2152.013669] env[62824]: DEBUG nova.virt.hardware [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2152.013839] env[62824]: DEBUG nova.virt.hardware [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2152.014036] env[62824]: DEBUG nova.virt.hardware [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2152.014229] env[62824]: DEBUG nova.virt.hardware [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2152.015331] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da67216b-4ddc-4b05-ad58-50f854a9052d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2152.020501] env[62824]: INFO nova.scheduler.client.report [None req-7beba819-8d40-4b17-972a-05048027ddd5 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Deleted allocations for instance 02f34565-6675-4c79-ac47-b131ceba9df8 [ 2152.028147] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15677148-bfb3-483d-b0f8-2541aaf138b8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2152.057732] env[62824]: DEBUG nova.compute.manager [req-e4df16e5-5ec8-4f3b-b4b9-77bf511e3c5c req-6080d2a6-6688-4063-a781-9f73f14ee3a7 service nova] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Received event network-changed-e285a842-d54b-4614-84ea-239e34f5b41f {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2152.058081] env[62824]: DEBUG nova.compute.manager [req-e4df16e5-5ec8-4f3b-b4b9-77bf511e3c5c req-6080d2a6-6688-4063-a781-9f73f14ee3a7 service nova] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Refreshing instance network info cache due to event network-changed-e285a842-d54b-4614-84ea-239e34f5b41f. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2152.058195] env[62824]: DEBUG oslo_concurrency.lockutils [req-e4df16e5-5ec8-4f3b-b4b9-77bf511e3c5c req-6080d2a6-6688-4063-a781-9f73f14ee3a7 service nova] Acquiring lock "refresh_cache-fc77c44a-180c-46ff-9690-9072c6213c91" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2152.058302] env[62824]: DEBUG oslo_concurrency.lockutils [req-e4df16e5-5ec8-4f3b-b4b9-77bf511e3c5c req-6080d2a6-6688-4063-a781-9f73f14ee3a7 service nova] Acquired lock "refresh_cache-fc77c44a-180c-46ff-9690-9072c6213c91" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2152.058458] env[62824]: DEBUG nova.network.neutron [req-e4df16e5-5ec8-4f3b-b4b9-77bf511e3c5c req-6080d2a6-6688-4063-a781-9f73f14ee3a7 service nova] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Refreshing network info cache for port e285a842-d54b-4614-84ea-239e34f5b41f {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2152.290314] env[62824]: DEBUG oslo_vmware.api [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': task-2146439, 'name': ReconfigVM_Task, 'duration_secs': 0.882411} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2152.290642] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 20f93b46-5e7e-4256-8cc1-e0a1b16740d5] Reconfigured VM instance instance-0000004c to attach disk [datastore1] 20f93b46-5e7e-4256-8cc1-e0a1b16740d5/20f93b46-5e7e-4256-8cc1-e0a1b16740d5.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2152.291326] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-08e74826-4ca5-4d9d-a97d-c2c453ef42a9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2152.297308] env[62824]: DEBUG oslo_vmware.api [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Waiting for the task: (returnval){ [ 2152.297308] env[62824]: value = "task-2146441" [ 2152.297308] env[62824]: _type = "Task" [ 2152.297308] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2152.304608] env[62824]: DEBUG oslo_vmware.api [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': task-2146441, 'name': Rename_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2152.306086] env[62824]: INFO nova.compute.manager [-] [instance: dadf590a-7288-4dd2-90de-125fe272f4ba] Took 1.03 seconds to deallocate network for instance. [ 2152.406826] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146440, 'name': CreateVM_Task, 'duration_secs': 0.382812} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2152.407130] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2152.407906] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2152.408083] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2152.408411] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2152.408675] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-222f1584-7934-4a06-bf78-481b3252f917 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2152.414127] env[62824]: DEBUG oslo_vmware.api [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Waiting for the task: (returnval){ [ 2152.414127] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52d6f262-77e3-5131-e55d-e3d02e1953c1" [ 2152.414127] env[62824]: _type = "Task" [ 2152.414127] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2152.424462] env[62824]: DEBUG oslo_vmware.api [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52d6f262-77e3-5131-e55d-e3d02e1953c1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2152.530904] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7beba819-8d40-4b17-972a-05048027ddd5 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Lock "02f34565-6675-4c79-ac47-b131ceba9df8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.487s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2152.764249] env[62824]: DEBUG oslo_concurrency.lockutils [None req-947b662c-414a-42b6-bd23-f5e60b455b8c tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Acquiring lock "02405b17-7022-4e2c-a357-415de88f63d0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2152.764249] env[62824]: DEBUG oslo_concurrency.lockutils [None req-947b662c-414a-42b6-bd23-f5e60b455b8c tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Lock "02405b17-7022-4e2c-a357-415de88f63d0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2152.764249] env[62824]: DEBUG oslo_concurrency.lockutils [None req-947b662c-414a-42b6-bd23-f5e60b455b8c tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Acquiring lock "02405b17-7022-4e2c-a357-415de88f63d0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2152.764249] env[62824]: DEBUG oslo_concurrency.lockutils [None req-947b662c-414a-42b6-bd23-f5e60b455b8c tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Lock "02405b17-7022-4e2c-a357-415de88f63d0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2152.764249] env[62824]: DEBUG oslo_concurrency.lockutils [None req-947b662c-414a-42b6-bd23-f5e60b455b8c tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Lock "02405b17-7022-4e2c-a357-415de88f63d0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2152.764249] env[62824]: INFO nova.compute.manager [None req-947b662c-414a-42b6-bd23-f5e60b455b8c tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02405b17-7022-4e2c-a357-415de88f63d0] Terminating instance [ 2152.807836] env[62824]: DEBUG oslo_vmware.api [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': task-2146441, 'name': Rename_Task, 'duration_secs': 0.128888} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2152.808134] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 20f93b46-5e7e-4256-8cc1-e0a1b16740d5] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2152.812325] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6e3330c9-9ab9-48da-8e25-e4079fbc7ee4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2152.812325] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ae8aad58-b430-4dd1-8bc0-b6bcc01e69fc tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2152.815421] env[62824]: DEBUG oslo_vmware.api [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Waiting for the task: (returnval){ [ 2152.815421] env[62824]: value = "task-2146442" [ 2152.815421] env[62824]: _type = "Task" [ 2152.815421] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2152.819388] env[62824]: DEBUG nova.network.neutron [req-e4df16e5-5ec8-4f3b-b4b9-77bf511e3c5c req-6080d2a6-6688-4063-a781-9f73f14ee3a7 service nova] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Updated VIF entry in instance network info cache for port e285a842-d54b-4614-84ea-239e34f5b41f. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2152.819755] env[62824]: DEBUG nova.network.neutron [req-e4df16e5-5ec8-4f3b-b4b9-77bf511e3c5c req-6080d2a6-6688-4063-a781-9f73f14ee3a7 service nova] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Updating instance_info_cache with network_info: [{"id": "e285a842-d54b-4614-84ea-239e34f5b41f", "address": "fa:16:3e:9d:17:cd", "network": {"id": "aa1c09e5-8d72-43ad-a903-a0f5e711da80", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1643616572-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c987c5f844be42c99cc9d57a9ca84c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9bc2632-36f9-4912-8782-8bbb789f909d", "external-id": "nsx-vlan-transportzone-897", "segmentation_id": 897, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape285a842-d5", "ovs_interfaceid": "e285a842-d54b-4614-84ea-239e34f5b41f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2152.831185] env[62824]: DEBUG oslo_vmware.api [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': task-2146442, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2152.926531] env[62824]: DEBUG oslo_vmware.api [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52d6f262-77e3-5131-e55d-e3d02e1953c1, 'name': SearchDatastore_Task, 'duration_secs': 0.018109} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2152.926832] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2152.927087] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2152.927349] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2152.927503] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2152.927702] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2152.927990] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4c3d7d16-76e5-45d5-8e6e-3308534f19a5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2152.936417] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2152.936704] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2152.937632] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-13d6fab5-11e9-4da9-b4b1-656f03ab467c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2152.943530] env[62824]: DEBUG oslo_vmware.api [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Waiting for the task: (returnval){ [ 2152.943530] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]525734d5-a2bc-8d94-512f-7aa3947909d2" [ 2152.943530] env[62824]: _type = "Task" [ 2152.943530] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2152.951544] env[62824]: DEBUG oslo_vmware.api [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]525734d5-a2bc-8d94-512f-7aa3947909d2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2153.001189] env[62824]: DEBUG nova.network.neutron [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] [instance: bd7f2ed5-fa08-4a48-9755-60b61791ea0b] Successfully updated port: 01afe8c7-58f8-4fa9-97a6-4767b34ccb66 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2153.192901] env[62824]: DEBUG nova.compute.manager [None req-947b662c-414a-42b6-bd23-f5e60b455b8c tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02405b17-7022-4e2c-a357-415de88f63d0] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2153.193215] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-947b662c-414a-42b6-bd23-f5e60b455b8c tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02405b17-7022-4e2c-a357-415de88f63d0] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2153.198153] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19121dd9-a42a-4d2a-8078-175a5b0e80bb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2153.206190] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-947b662c-414a-42b6-bd23-f5e60b455b8c tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02405b17-7022-4e2c-a357-415de88f63d0] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2153.206390] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-826f34ec-bb3c-4d27-94b2-248d331f1298 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2153.215156] env[62824]: DEBUG oslo_vmware.api [None req-947b662c-414a-42b6-bd23-f5e60b455b8c tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for the task: (returnval){ [ 2153.215156] env[62824]: value = "task-2146443" [ 2153.215156] env[62824]: _type = "Task" [ 2153.215156] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2153.222915] env[62824]: DEBUG oslo_vmware.api [None req-947b662c-414a-42b6-bd23-f5e60b455b8c tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146443, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2153.327402] env[62824]: DEBUG oslo_vmware.api [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': task-2146442, 'name': PowerOnVM_Task} progress is 78%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2153.333093] env[62824]: DEBUG oslo_concurrency.lockutils [req-e4df16e5-5ec8-4f3b-b4b9-77bf511e3c5c req-6080d2a6-6688-4063-a781-9f73f14ee3a7 service nova] Releasing lock "refresh_cache-fc77c44a-180c-46ff-9690-9072c6213c91" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2153.437462] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65164c8c-7f12-493a-ac7e-e3c6e7588ecb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2153.449490] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9918012a-a71f-4a7b-a55f-592d18f40877 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2153.460067] env[62824]: DEBUG oslo_vmware.api [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]525734d5-a2bc-8d94-512f-7aa3947909d2, 'name': SearchDatastore_Task, 'duration_secs': 0.008771} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2153.487785] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-58305e91-a302-4054-b4c2-9e467edff7bf {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2153.491330] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1478032-eb6c-4dc3-aa81-56eebb603623 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2153.497997] env[62824]: DEBUG oslo_vmware.api [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Waiting for the task: (returnval){ [ 2153.497997] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52915336-00db-6128-7840-647153c154e4" [ 2153.497997] env[62824]: _type = "Task" [ 2153.497997] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2153.504913] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b8f7e97-b79b-4ab0-bc82-e289252894c0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2153.509710] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Acquiring lock "refresh_cache-bd7f2ed5-fa08-4a48-9755-60b61791ea0b" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2153.509710] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Acquired lock "refresh_cache-bd7f2ed5-fa08-4a48-9755-60b61791ea0b" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2153.509710] env[62824]: DEBUG nova.network.neutron [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] [instance: bd7f2ed5-fa08-4a48-9755-60b61791ea0b] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2153.516671] env[62824]: DEBUG oslo_vmware.api [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52915336-00db-6128-7840-647153c154e4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2153.525266] env[62824]: DEBUG nova.compute.provider_tree [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2153.726246] env[62824]: DEBUG oslo_vmware.api [None req-947b662c-414a-42b6-bd23-f5e60b455b8c tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146443, 'name': PowerOffVM_Task, 'duration_secs': 0.270778} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2153.726542] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-947b662c-414a-42b6-bd23-f5e60b455b8c tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02405b17-7022-4e2c-a357-415de88f63d0] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2153.726723] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-947b662c-414a-42b6-bd23-f5e60b455b8c tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02405b17-7022-4e2c-a357-415de88f63d0] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2153.726959] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0dec315a-40e9-448a-b527-07564e1d171e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2153.825663] env[62824]: DEBUG oslo_vmware.api [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': task-2146442, 'name': PowerOnVM_Task, 'duration_secs': 0.746441} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2153.825875] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 20f93b46-5e7e-4256-8cc1-e0a1b16740d5] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2153.826093] env[62824]: INFO nova.compute.manager [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 20f93b46-5e7e-4256-8cc1-e0a1b16740d5] Took 8.13 seconds to spawn the instance on the hypervisor. [ 2153.826275] env[62824]: DEBUG nova.compute.manager [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 20f93b46-5e7e-4256-8cc1-e0a1b16740d5] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2153.827137] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea73b25a-0828-4464-b135-5b0efacd9b97 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2153.963065] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-947b662c-414a-42b6-bd23-f5e60b455b8c tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02405b17-7022-4e2c-a357-415de88f63d0] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2153.963348] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-947b662c-414a-42b6-bd23-f5e60b455b8c tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02405b17-7022-4e2c-a357-415de88f63d0] Deleting contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2153.963486] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-947b662c-414a-42b6-bd23-f5e60b455b8c tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Deleting the datastore file [datastore1] 02405b17-7022-4e2c-a357-415de88f63d0 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2153.963757] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-13b8a784-1ef8-45ae-b862-7c977d6b99a0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2153.970026] env[62824]: DEBUG oslo_vmware.api [None req-947b662c-414a-42b6-bd23-f5e60b455b8c tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for the task: (returnval){ [ 2153.970026] env[62824]: value = "task-2146445" [ 2153.970026] env[62824]: _type = "Task" [ 2153.970026] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2153.977794] env[62824]: DEBUG oslo_vmware.api [None req-947b662c-414a-42b6-bd23-f5e60b455b8c tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146445, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2154.010608] env[62824]: DEBUG oslo_vmware.api [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52915336-00db-6128-7840-647153c154e4, 'name': SearchDatastore_Task, 'duration_secs': 0.018978} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2154.010608] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2154.010870] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] fc77c44a-180c-46ff-9690-9072c6213c91/fc77c44a-180c-46ff-9690-9072c6213c91.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2154.011134] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a832e501-839e-4e9c-a1cf-4643f3817c63 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2154.019662] env[62824]: DEBUG oslo_vmware.api [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Waiting for the task: (returnval){ [ 2154.019662] env[62824]: value = "task-2146446" [ 2154.019662] env[62824]: _type = "Task" [ 2154.019662] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2154.028093] env[62824]: DEBUG nova.scheduler.client.report [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2154.031842] env[62824]: DEBUG oslo_vmware.api [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146446, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2154.048426] env[62824]: DEBUG nova.network.neutron [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] [instance: bd7f2ed5-fa08-4a48-9755-60b61791ea0b] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2154.088441] env[62824]: DEBUG nova.compute.manager [req-253af903-ba20-42d8-83f8-a1c995934aa1 req-402dda4a-df0c-4a97-8374-4f50d35bd520 service nova] [instance: bd7f2ed5-fa08-4a48-9755-60b61791ea0b] Received event network-vif-plugged-01afe8c7-58f8-4fa9-97a6-4767b34ccb66 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2154.089055] env[62824]: DEBUG oslo_concurrency.lockutils [req-253af903-ba20-42d8-83f8-a1c995934aa1 req-402dda4a-df0c-4a97-8374-4f50d35bd520 service nova] Acquiring lock "bd7f2ed5-fa08-4a48-9755-60b61791ea0b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2154.089055] env[62824]: DEBUG oslo_concurrency.lockutils [req-253af903-ba20-42d8-83f8-a1c995934aa1 req-402dda4a-df0c-4a97-8374-4f50d35bd520 service nova] Lock "bd7f2ed5-fa08-4a48-9755-60b61791ea0b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2154.089227] env[62824]: DEBUG oslo_concurrency.lockutils [req-253af903-ba20-42d8-83f8-a1c995934aa1 req-402dda4a-df0c-4a97-8374-4f50d35bd520 service nova] Lock "bd7f2ed5-fa08-4a48-9755-60b61791ea0b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2154.089313] env[62824]: DEBUG nova.compute.manager [req-253af903-ba20-42d8-83f8-a1c995934aa1 req-402dda4a-df0c-4a97-8374-4f50d35bd520 service nova] [instance: bd7f2ed5-fa08-4a48-9755-60b61791ea0b] No waiting events found dispatching network-vif-plugged-01afe8c7-58f8-4fa9-97a6-4767b34ccb66 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2154.089509] env[62824]: WARNING nova.compute.manager [req-253af903-ba20-42d8-83f8-a1c995934aa1 req-402dda4a-df0c-4a97-8374-4f50d35bd520 service nova] [instance: bd7f2ed5-fa08-4a48-9755-60b61791ea0b] Received unexpected event network-vif-plugged-01afe8c7-58f8-4fa9-97a6-4767b34ccb66 for instance with vm_state building and task_state spawning. [ 2154.089701] env[62824]: DEBUG nova.compute.manager [req-253af903-ba20-42d8-83f8-a1c995934aa1 req-402dda4a-df0c-4a97-8374-4f50d35bd520 service nova] [instance: bd7f2ed5-fa08-4a48-9755-60b61791ea0b] Received event network-changed-01afe8c7-58f8-4fa9-97a6-4767b34ccb66 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2154.089782] env[62824]: DEBUG nova.compute.manager [req-253af903-ba20-42d8-83f8-a1c995934aa1 req-402dda4a-df0c-4a97-8374-4f50d35bd520 service nova] [instance: bd7f2ed5-fa08-4a48-9755-60b61791ea0b] Refreshing instance network info cache due to event network-changed-01afe8c7-58f8-4fa9-97a6-4767b34ccb66. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2154.089949] env[62824]: DEBUG oslo_concurrency.lockutils [req-253af903-ba20-42d8-83f8-a1c995934aa1 req-402dda4a-df0c-4a97-8374-4f50d35bd520 service nova] Acquiring lock "refresh_cache-bd7f2ed5-fa08-4a48-9755-60b61791ea0b" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2154.270209] env[62824]: DEBUG nova.network.neutron [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] [instance: bd7f2ed5-fa08-4a48-9755-60b61791ea0b] Updating instance_info_cache with network_info: [{"id": "01afe8c7-58f8-4fa9-97a6-4767b34ccb66", "address": "fa:16:3e:e6:db:30", "network": {"id": "8e0a3000-b2ed-456c-adf4-de3ab93890f7", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-2078976590-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "958e2482b37342959d733f14c9242618", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73eeba7c-29e1-4fdf-82b3-d62e63e86051", "external-id": "cl2-zone-659", "segmentation_id": 659, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01afe8c7-58", "ovs_interfaceid": "01afe8c7-58f8-4fa9-97a6-4767b34ccb66", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2154.346356] env[62824]: INFO nova.compute.manager [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 20f93b46-5e7e-4256-8cc1-e0a1b16740d5] Took 38.75 seconds to build instance. [ 2154.481158] env[62824]: DEBUG oslo_vmware.api [None req-947b662c-414a-42b6-bd23-f5e60b455b8c tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146445, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.224481} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2154.481433] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-947b662c-414a-42b6-bd23-f5e60b455b8c tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2154.481624] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-947b662c-414a-42b6-bd23-f5e60b455b8c tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02405b17-7022-4e2c-a357-415de88f63d0] Deleted contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2154.481807] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-947b662c-414a-42b6-bd23-f5e60b455b8c tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02405b17-7022-4e2c-a357-415de88f63d0] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2154.481984] env[62824]: INFO nova.compute.manager [None req-947b662c-414a-42b6-bd23-f5e60b455b8c tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 02405b17-7022-4e2c-a357-415de88f63d0] Took 1.29 seconds to destroy the instance on the hypervisor. [ 2154.482243] env[62824]: DEBUG oslo.service.loopingcall [None req-947b662c-414a-42b6-bd23-f5e60b455b8c tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2154.482448] env[62824]: DEBUG nova.compute.manager [-] [instance: 02405b17-7022-4e2c-a357-415de88f63d0] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2154.482550] env[62824]: DEBUG nova.network.neutron [-] [instance: 02405b17-7022-4e2c-a357-415de88f63d0] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2154.529527] env[62824]: DEBUG oslo_vmware.api [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146446, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2154.533536] env[62824]: DEBUG oslo_concurrency.lockutils [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.553s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2154.534070] env[62824]: DEBUG nova.compute.manager [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 2afd1f18-234c-40b5-9ede-7413ad30dafe] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2154.536741] env[62824]: DEBUG oslo_concurrency.lockutils [None req-66e16106-e2f9-4347-968f-3ba76efdbbd5 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.017s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2154.536964] env[62824]: DEBUG nova.objects.instance [None req-66e16106-e2f9-4347-968f-3ba76efdbbd5 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Lazy-loading 'resources' on Instance uuid 363177c0-dc40-429a-a74b-e690da133edb {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2154.773026] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Releasing lock "refresh_cache-bd7f2ed5-fa08-4a48-9755-60b61791ea0b" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2154.773393] env[62824]: DEBUG nova.compute.manager [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] [instance: bd7f2ed5-fa08-4a48-9755-60b61791ea0b] Instance network_info: |[{"id": "01afe8c7-58f8-4fa9-97a6-4767b34ccb66", "address": "fa:16:3e:e6:db:30", "network": {"id": "8e0a3000-b2ed-456c-adf4-de3ab93890f7", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-2078976590-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "958e2482b37342959d733f14c9242618", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73eeba7c-29e1-4fdf-82b3-d62e63e86051", "external-id": "cl2-zone-659", "segmentation_id": 659, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01afe8c7-58", "ovs_interfaceid": "01afe8c7-58f8-4fa9-97a6-4767b34ccb66", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2154.773704] env[62824]: DEBUG oslo_concurrency.lockutils [req-253af903-ba20-42d8-83f8-a1c995934aa1 req-402dda4a-df0c-4a97-8374-4f50d35bd520 service nova] Acquired lock "refresh_cache-bd7f2ed5-fa08-4a48-9755-60b61791ea0b" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2154.773883] env[62824]: DEBUG nova.network.neutron [req-253af903-ba20-42d8-83f8-a1c995934aa1 req-402dda4a-df0c-4a97-8374-4f50d35bd520 service nova] [instance: bd7f2ed5-fa08-4a48-9755-60b61791ea0b] Refreshing network info cache for port 01afe8c7-58f8-4fa9-97a6-4767b34ccb66 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2154.775075] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] [instance: bd7f2ed5-fa08-4a48-9755-60b61791ea0b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e6:db:30', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '73eeba7c-29e1-4fdf-82b3-d62e63e86051', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '01afe8c7-58f8-4fa9-97a6-4767b34ccb66', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2154.783633] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Creating folder: Project (958e2482b37342959d733f14c9242618). Parent ref: group-v438503. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 2154.785212] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-69f8aa98-ab44-4531-a255-fe32efd7b395 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2154.798640] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Created folder: Project (958e2482b37342959d733f14c9242618) in parent group-v438503. [ 2154.798864] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Creating folder: Instances. Parent ref: group-v438700. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 2154.799144] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d70b5b66-29b0-44e4-a641-8b25f5cc123e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2154.808181] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Created folder: Instances in parent group-v438700. [ 2154.809890] env[62824]: DEBUG oslo.service.loopingcall [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2154.809890] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bd7f2ed5-fa08-4a48-9755-60b61791ea0b] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2154.809890] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-41465a95-29b0-4135-889b-700aa2fdcce0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2154.831844] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2154.831844] env[62824]: value = "task-2146449" [ 2154.831844] env[62824]: _type = "Task" [ 2154.831844] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2154.842666] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146449, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2154.848213] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bccc42f8-2e3c-4a65-bd8a-e0e1e35b44fe tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Lock "20f93b46-5e7e-4256-8cc1-e0a1b16740d5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 61.794s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2155.030373] env[62824]: DEBUG oslo_vmware.api [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146446, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.828069} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2155.031664] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] fc77c44a-180c-46ff-9690-9072c6213c91/fc77c44a-180c-46ff-9690-9072c6213c91.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2155.031841] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2155.034396] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-03ab3d42-aab3-40f4-94dc-5b5c12ebb908 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2155.042491] env[62824]: DEBUG nova.compute.utils [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2155.046538] env[62824]: DEBUG nova.compute.manager [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 2afd1f18-234c-40b5-9ede-7413ad30dafe] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2155.046774] env[62824]: DEBUG nova.network.neutron [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 2afd1f18-234c-40b5-9ede-7413ad30dafe] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2155.051273] env[62824]: DEBUG oslo_vmware.api [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Waiting for the task: (returnval){ [ 2155.051273] env[62824]: value = "task-2146450" [ 2155.051273] env[62824]: _type = "Task" [ 2155.051273] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2155.062824] env[62824]: DEBUG oslo_vmware.api [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146450, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2155.125217] env[62824]: DEBUG nova.policy [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd64a0b7abab54b15a91ec809b0cf87f3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2e59203d676d48c8bd8402af4eb58f65', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2155.350480] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146449, 'name': CreateVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2155.441456] env[62824]: DEBUG nova.network.neutron [-] [instance: 02405b17-7022-4e2c-a357-415de88f63d0] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2155.529174] env[62824]: DEBUG nova.network.neutron [req-253af903-ba20-42d8-83f8-a1c995934aa1 req-402dda4a-df0c-4a97-8374-4f50d35bd520 service nova] [instance: bd7f2ed5-fa08-4a48-9755-60b61791ea0b] Updated VIF entry in instance network info cache for port 01afe8c7-58f8-4fa9-97a6-4767b34ccb66. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2155.529544] env[62824]: DEBUG nova.network.neutron [req-253af903-ba20-42d8-83f8-a1c995934aa1 req-402dda4a-df0c-4a97-8374-4f50d35bd520 service nova] [instance: bd7f2ed5-fa08-4a48-9755-60b61791ea0b] Updating instance_info_cache with network_info: [{"id": "01afe8c7-58f8-4fa9-97a6-4767b34ccb66", "address": "fa:16:3e:e6:db:30", "network": {"id": "8e0a3000-b2ed-456c-adf4-de3ab93890f7", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-2078976590-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "958e2482b37342959d733f14c9242618", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73eeba7c-29e1-4fdf-82b3-d62e63e86051", "external-id": "cl2-zone-659", "segmentation_id": 659, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01afe8c7-58", "ovs_interfaceid": "01afe8c7-58f8-4fa9-97a6-4767b34ccb66", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2155.534225] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e4531cd-8f6b-4632-a450-8b045412a194 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2155.537043] env[62824]: DEBUG oslo_concurrency.lockutils [None req-aefa2097-549e-43b5-9152-0e8ce3b5da7a tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Acquiring lock "7a2a51c4-e558-4f5f-b82c-718bc12c1df5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2155.537346] env[62824]: DEBUG oslo_concurrency.lockutils [None req-aefa2097-549e-43b5-9152-0e8ce3b5da7a tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Lock "7a2a51c4-e558-4f5f-b82c-718bc12c1df5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2155.537680] env[62824]: DEBUG oslo_concurrency.lockutils [None req-aefa2097-549e-43b5-9152-0e8ce3b5da7a tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Acquiring lock "7a2a51c4-e558-4f5f-b82c-718bc12c1df5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2155.537891] env[62824]: DEBUG oslo_concurrency.lockutils [None req-aefa2097-549e-43b5-9152-0e8ce3b5da7a tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Lock "7a2a51c4-e558-4f5f-b82c-718bc12c1df5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2155.538087] env[62824]: DEBUG oslo_concurrency.lockutils [None req-aefa2097-549e-43b5-9152-0e8ce3b5da7a tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Lock "7a2a51c4-e558-4f5f-b82c-718bc12c1df5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2155.545865] env[62824]: INFO nova.compute.manager [None req-aefa2097-549e-43b5-9152-0e8ce3b5da7a tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7a2a51c4-e558-4f5f-b82c-718bc12c1df5] Terminating instance [ 2155.548040] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-650e86b5-12d7-49c8-b4d0-edfbe9304668 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2155.553193] env[62824]: DEBUG nova.compute.manager [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 2afd1f18-234c-40b5-9ede-7413ad30dafe] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2155.568677] env[62824]: DEBUG oslo_vmware.api [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146450, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.177256} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2155.595612] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2155.597511] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b80a3b91-70f1-49bf-8486-a3b3bc794133 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2155.600535] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d39300f-21b0-4e42-82ce-fcd18bffa195 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2155.619513] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11f28401-a6a8-437b-a322-c6551414c005 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2155.633244] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Reconfiguring VM instance instance-0000004d to attach disk [datastore2] fc77c44a-180c-46ff-9690-9072c6213c91/fc77c44a-180c-46ff-9690-9072c6213c91.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2155.634212] env[62824]: DEBUG nova.network.neutron [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 2afd1f18-234c-40b5-9ede-7413ad30dafe] Successfully created port: 6d0cd52e-ace8-4459-860b-f270fa266268 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2155.636472] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7f94c0cf-c8c7-4d17-95f7-6b8a2234ba0b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2155.666131] env[62824]: DEBUG nova.compute.provider_tree [None req-66e16106-e2f9-4347-968f-3ba76efdbbd5 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2155.668598] env[62824]: DEBUG oslo_vmware.api [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Waiting for the task: (returnval){ [ 2155.668598] env[62824]: value = "task-2146451" [ 2155.668598] env[62824]: _type = "Task" [ 2155.668598] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2155.669519] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5fb295c4-ec07-4626-81a5-775a8bd81d25 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Acquiring lock "20f93b46-5e7e-4256-8cc1-e0a1b16740d5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2155.669745] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5fb295c4-ec07-4626-81a5-775a8bd81d25 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Lock "20f93b46-5e7e-4256-8cc1-e0a1b16740d5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2155.669945] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5fb295c4-ec07-4626-81a5-775a8bd81d25 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Acquiring lock "20f93b46-5e7e-4256-8cc1-e0a1b16740d5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2155.670145] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5fb295c4-ec07-4626-81a5-775a8bd81d25 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Lock "20f93b46-5e7e-4256-8cc1-e0a1b16740d5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2155.670310] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5fb295c4-ec07-4626-81a5-775a8bd81d25 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Lock "20f93b46-5e7e-4256-8cc1-e0a1b16740d5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2155.675184] env[62824]: INFO nova.compute.manager [None req-5fb295c4-ec07-4626-81a5-775a8bd81d25 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 20f93b46-5e7e-4256-8cc1-e0a1b16740d5] Terminating instance [ 2155.683809] env[62824]: DEBUG oslo_vmware.api [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146451, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2155.699957] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "refresh_cache-07ce931f-45ef-409b-b714-9f1cd47a3a88" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2155.700113] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquired lock "refresh_cache-07ce931f-45ef-409b-b714-9f1cd47a3a88" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2155.700266] env[62824]: DEBUG nova.network.neutron [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Forcefully refreshing network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 2155.843451] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146449, 'name': CreateVM_Task, 'duration_secs': 0.583797} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2155.844990] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bd7f2ed5-fa08-4a48-9755-60b61791ea0b] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2155.844990] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2155.844990] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2155.844990] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2155.845312] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2feabdde-0e4d-41d6-84df-3c0b0a2fcd5e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2155.851580] env[62824]: DEBUG oslo_vmware.api [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Waiting for the task: (returnval){ [ 2155.851580] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52007d8d-74de-22d6-2519-31cc60ff4a19" [ 2155.851580] env[62824]: _type = "Task" [ 2155.851580] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2155.863406] env[62824]: DEBUG oslo_vmware.api [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52007d8d-74de-22d6-2519-31cc60ff4a19, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2155.944181] env[62824]: INFO nova.compute.manager [-] [instance: 02405b17-7022-4e2c-a357-415de88f63d0] Took 1.46 seconds to deallocate network for instance. [ 2156.032794] env[62824]: DEBUG oslo_concurrency.lockutils [req-253af903-ba20-42d8-83f8-a1c995934aa1 req-402dda4a-df0c-4a97-8374-4f50d35bd520 service nova] Releasing lock "refresh_cache-bd7f2ed5-fa08-4a48-9755-60b61791ea0b" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2156.057315] env[62824]: DEBUG nova.compute.manager [None req-aefa2097-549e-43b5-9152-0e8ce3b5da7a tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7a2a51c4-e558-4f5f-b82c-718bc12c1df5] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2156.057556] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-aefa2097-549e-43b5-9152-0e8ce3b5da7a tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7a2a51c4-e558-4f5f-b82c-718bc12c1df5] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2156.061699] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab4a0d3a-f7cf-4585-81c6-ee606f4dcd89 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2156.071623] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-aefa2097-549e-43b5-9152-0e8ce3b5da7a tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7a2a51c4-e558-4f5f-b82c-718bc12c1df5] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2156.071623] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9d64f2c5-2280-4622-b6d8-c5d660766c01 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2156.079378] env[62824]: DEBUG oslo_vmware.api [None req-aefa2097-549e-43b5-9152-0e8ce3b5da7a tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Waiting for the task: (returnval){ [ 2156.079378] env[62824]: value = "task-2146452" [ 2156.079378] env[62824]: _type = "Task" [ 2156.079378] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2156.088073] env[62824]: DEBUG oslo_vmware.api [None req-aefa2097-549e-43b5-9152-0e8ce3b5da7a tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': task-2146452, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2156.120646] env[62824]: DEBUG nova.compute.manager [req-3893d770-f720-4a32-83db-d4daabc0c32a req-d3abfb0b-88e7-40c4-b307-b29524065d25 service nova] [instance: 02405b17-7022-4e2c-a357-415de88f63d0] Received event network-vif-deleted-cab3935d-90df-47ef-9198-10691fe660b4 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2156.170691] env[62824]: DEBUG nova.scheduler.client.report [None req-66e16106-e2f9-4347-968f-3ba76efdbbd5 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2156.178694] env[62824]: DEBUG nova.compute.manager [None req-5fb295c4-ec07-4626-81a5-775a8bd81d25 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 20f93b46-5e7e-4256-8cc1-e0a1b16740d5] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2156.178907] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5fb295c4-ec07-4626-81a5-775a8bd81d25 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 20f93b46-5e7e-4256-8cc1-e0a1b16740d5] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2156.179712] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee096117-ecd1-4b0c-9dbf-301ae7fbc0ad {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2156.188405] env[62824]: DEBUG oslo_vmware.api [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146451, 'name': ReconfigVM_Task, 'duration_secs': 0.294315} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2156.189657] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Reconfigured VM instance instance-0000004d to attach disk [datastore2] fc77c44a-180c-46ff-9690-9072c6213c91/fc77c44a-180c-46ff-9690-9072c6213c91.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2156.190505] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6c46d1ee-bb16-4853-a3cb-229ad1b51ce0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2156.194399] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fb295c4-ec07-4626-81a5-775a8bd81d25 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 20f93b46-5e7e-4256-8cc1-e0a1b16740d5] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2156.194993] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c9adae69-2c9b-42f9-bebb-5c7625fee5e1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2156.201569] env[62824]: DEBUG oslo_vmware.api [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Waiting for the task: (returnval){ [ 2156.201569] env[62824]: value = "task-2146453" [ 2156.201569] env[62824]: _type = "Task" [ 2156.201569] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2156.204859] env[62824]: DEBUG oslo_vmware.api [None req-5fb295c4-ec07-4626-81a5-775a8bd81d25 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Waiting for the task: (returnval){ [ 2156.204859] env[62824]: value = "task-2146454" [ 2156.204859] env[62824]: _type = "Task" [ 2156.204859] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2156.217486] env[62824]: DEBUG oslo_vmware.api [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146453, 'name': Rename_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2156.220740] env[62824]: DEBUG oslo_vmware.api [None req-5fb295c4-ec07-4626-81a5-775a8bd81d25 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': task-2146454, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2156.364906] env[62824]: DEBUG oslo_vmware.api [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52007d8d-74de-22d6-2519-31cc60ff4a19, 'name': SearchDatastore_Task, 'duration_secs': 0.017285} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2156.365336] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2156.365651] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] [instance: bd7f2ed5-fa08-4a48-9755-60b61791ea0b] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2156.365863] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2156.366035] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2156.366258] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2156.366558] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-77376e7c-4d9d-464a-9e2e-cef7d6f7be88 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2156.383687] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2156.383687] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2156.385062] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-39ac42e9-659b-4b0b-a94c-58bba6e2ba8d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2156.393898] env[62824]: DEBUG oslo_vmware.api [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Waiting for the task: (returnval){ [ 2156.393898] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]522c4cde-eb87-be52-5bb1-113490ec9a46" [ 2156.393898] env[62824]: _type = "Task" [ 2156.393898] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2156.402805] env[62824]: DEBUG oslo_vmware.api [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]522c4cde-eb87-be52-5bb1-113490ec9a46, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2156.452124] env[62824]: DEBUG oslo_concurrency.lockutils [None req-947b662c-414a-42b6-bd23-f5e60b455b8c tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2156.569021] env[62824]: DEBUG nova.compute.manager [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 2afd1f18-234c-40b5-9ede-7413ad30dafe] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2156.592786] env[62824]: DEBUG oslo_vmware.api [None req-aefa2097-549e-43b5-9152-0e8ce3b5da7a tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': task-2146452, 'name': PowerOffVM_Task, 'duration_secs': 0.361388} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2156.593283] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-aefa2097-549e-43b5-9152-0e8ce3b5da7a tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7a2a51c4-e558-4f5f-b82c-718bc12c1df5] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2156.593283] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-aefa2097-549e-43b5-9152-0e8ce3b5da7a tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7a2a51c4-e558-4f5f-b82c-718bc12c1df5] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2156.593511] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6cd41674-11cf-458f-b2a2-b32063875645 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2156.598722] env[62824]: DEBUG nova.virt.hardware [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2156.598949] env[62824]: DEBUG nova.virt.hardware [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2156.599123] env[62824]: DEBUG nova.virt.hardware [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2156.599494] env[62824]: DEBUG nova.virt.hardware [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2156.599494] env[62824]: DEBUG nova.virt.hardware [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2156.599734] env[62824]: DEBUG nova.virt.hardware [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2156.599803] env[62824]: DEBUG nova.virt.hardware [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2156.599951] env[62824]: DEBUG nova.virt.hardware [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2156.600137] env[62824]: DEBUG nova.virt.hardware [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2156.600305] env[62824]: DEBUG nova.virt.hardware [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2156.600479] env[62824]: DEBUG nova.virt.hardware [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2156.601993] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d32c3b9-d6e6-4487-999d-ffc2df79470d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2156.613133] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad600f8c-57c0-4ed6-bd7e-0ebbfc115d3a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2156.678985] env[62824]: DEBUG oslo_concurrency.lockutils [None req-66e16106-e2f9-4347-968f-3ba76efdbbd5 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.142s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2156.681395] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.145s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2156.682828] env[62824]: INFO nova.compute.claims [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] [instance: ab37b1a2-0012-40fb-9341-b613525e89cf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2156.709707] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-aefa2097-549e-43b5-9152-0e8ce3b5da7a tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7a2a51c4-e558-4f5f-b82c-718bc12c1df5] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2156.709938] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-aefa2097-549e-43b5-9152-0e8ce3b5da7a tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7a2a51c4-e558-4f5f-b82c-718bc12c1df5] Deleting contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2156.710169] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-aefa2097-549e-43b5-9152-0e8ce3b5da7a tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Deleting the datastore file [datastore1] 7a2a51c4-e558-4f5f-b82c-718bc12c1df5 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2156.711305] env[62824]: INFO nova.scheduler.client.report [None req-66e16106-e2f9-4347-968f-3ba76efdbbd5 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Deleted allocations for instance 363177c0-dc40-429a-a74b-e690da133edb [ 2156.716385] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-caad5770-f426-4fa4-8d77-96e9d57733bc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2156.723924] env[62824]: DEBUG oslo_vmware.api [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146453, 'name': Rename_Task, 'duration_secs': 0.16665} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2156.725165] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2156.725494] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8541ad8e-91a1-492a-9914-e023f5601f04 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2156.731257] env[62824]: DEBUG oslo_vmware.api [None req-5fb295c4-ec07-4626-81a5-775a8bd81d25 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': task-2146454, 'name': PowerOffVM_Task, 'duration_secs': 0.227984} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2156.732010] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fb295c4-ec07-4626-81a5-775a8bd81d25 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 20f93b46-5e7e-4256-8cc1-e0a1b16740d5] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2156.732200] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5fb295c4-ec07-4626-81a5-775a8bd81d25 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 20f93b46-5e7e-4256-8cc1-e0a1b16740d5] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2156.733983] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bf137866-e84d-4bd1-bd56-0830af445aec {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2156.735556] env[62824]: DEBUG oslo_vmware.api [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Waiting for the task: (returnval){ [ 2156.735556] env[62824]: value = "task-2146457" [ 2156.735556] env[62824]: _type = "Task" [ 2156.735556] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2156.735770] env[62824]: DEBUG oslo_vmware.api [None req-aefa2097-549e-43b5-9152-0e8ce3b5da7a tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Waiting for the task: (returnval){ [ 2156.735770] env[62824]: value = "task-2146456" [ 2156.735770] env[62824]: _type = "Task" [ 2156.735770] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2156.753455] env[62824]: DEBUG oslo_vmware.api [None req-aefa2097-549e-43b5-9152-0e8ce3b5da7a tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': task-2146456, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2156.755888] env[62824]: DEBUG oslo_vmware.api [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146457, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2156.905244] env[62824]: DEBUG oslo_vmware.api [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]522c4cde-eb87-be52-5bb1-113490ec9a46, 'name': SearchDatastore_Task, 'duration_secs': 0.01542} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2156.906072] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0e2606a3-f729-4fe2-9e26-fbc64e2ef30d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2156.912017] env[62824]: DEBUG oslo_vmware.api [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Waiting for the task: (returnval){ [ 2156.912017] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52bd51b3-093b-88cf-8cbf-c03d650e5c83" [ 2156.912017] env[62824]: _type = "Task" [ 2156.912017] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2156.920649] env[62824]: DEBUG oslo_vmware.api [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52bd51b3-093b-88cf-8cbf-c03d650e5c83, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2156.930320] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5fb295c4-ec07-4626-81a5-775a8bd81d25 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 20f93b46-5e7e-4256-8cc1-e0a1b16740d5] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2156.930502] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5fb295c4-ec07-4626-81a5-775a8bd81d25 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 20f93b46-5e7e-4256-8cc1-e0a1b16740d5] Deleting contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2156.930636] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-5fb295c4-ec07-4626-81a5-775a8bd81d25 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Deleting the datastore file [datastore1] 20f93b46-5e7e-4256-8cc1-e0a1b16740d5 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2156.930900] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-585321dd-e10f-4135-af2b-5b4c54d7e049 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2156.937567] env[62824]: DEBUG oslo_vmware.api [None req-5fb295c4-ec07-4626-81a5-775a8bd81d25 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Waiting for the task: (returnval){ [ 2156.937567] env[62824]: value = "task-2146459" [ 2156.937567] env[62824]: _type = "Task" [ 2156.937567] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2156.941878] env[62824]: DEBUG nova.network.neutron [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Updating instance_info_cache with network_info: [{"id": "309b4346-1935-41f1-a374-d322c4a8c016", "address": "fa:16:3e:07:e9:a0", "network": {"id": "fc381ca5-198f-4afc-9ebf-bc6e3de68c35", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-425455451-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "248af2bc3e7b41d2a74b0c6074b359e1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4c9a12d2-469f-4199-bfaa-f791d765deac", "external-id": "nsx-vlan-transportzone-96", "segmentation_id": 96, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap309b4346-19", "ovs_interfaceid": "309b4346-1935-41f1-a374-d322c4a8c016", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2156.947457] env[62824]: DEBUG oslo_vmware.api [None req-5fb295c4-ec07-4626-81a5-775a8bd81d25 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': task-2146459, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2157.228642] env[62824]: DEBUG oslo_concurrency.lockutils [None req-66e16106-e2f9-4347-968f-3ba76efdbbd5 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Lock "363177c0-dc40-429a-a74b-e690da133edb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.232s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2157.256699] env[62824]: DEBUG oslo_vmware.api [None req-aefa2097-549e-43b5-9152-0e8ce3b5da7a tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': task-2146456, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.148509} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2157.262195] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-aefa2097-549e-43b5-9152-0e8ce3b5da7a tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2157.262563] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-aefa2097-549e-43b5-9152-0e8ce3b5da7a tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7a2a51c4-e558-4f5f-b82c-718bc12c1df5] Deleted contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2157.262874] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-aefa2097-549e-43b5-9152-0e8ce3b5da7a tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7a2a51c4-e558-4f5f-b82c-718bc12c1df5] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2157.263194] env[62824]: INFO nova.compute.manager [None req-aefa2097-549e-43b5-9152-0e8ce3b5da7a tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7a2a51c4-e558-4f5f-b82c-718bc12c1df5] Took 1.21 seconds to destroy the instance on the hypervisor. [ 2157.263580] env[62824]: DEBUG oslo.service.loopingcall [None req-aefa2097-549e-43b5-9152-0e8ce3b5da7a tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2157.264650] env[62824]: DEBUG oslo_vmware.api [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146457, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2157.264727] env[62824]: DEBUG nova.compute.manager [-] [instance: 7a2a51c4-e558-4f5f-b82c-718bc12c1df5] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2157.265865] env[62824]: DEBUG nova.network.neutron [-] [instance: 7a2a51c4-e558-4f5f-b82c-718bc12c1df5] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2157.323961] env[62824]: DEBUG nova.network.neutron [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 2afd1f18-234c-40b5-9ede-7413ad30dafe] Successfully updated port: 6d0cd52e-ace8-4459-860b-f270fa266268 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2157.423794] env[62824]: DEBUG oslo_vmware.api [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52bd51b3-093b-88cf-8cbf-c03d650e5c83, 'name': SearchDatastore_Task, 'duration_secs': 0.015845} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2157.424217] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2157.424559] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] bd7f2ed5-fa08-4a48-9755-60b61791ea0b/bd7f2ed5-fa08-4a48-9755-60b61791ea0b.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2157.424793] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ec74600d-e389-489c-84d2-821e183989a1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2157.433685] env[62824]: DEBUG oslo_vmware.api [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Waiting for the task: (returnval){ [ 2157.433685] env[62824]: value = "task-2146460" [ 2157.433685] env[62824]: _type = "Task" [ 2157.433685] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2157.445713] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Releasing lock "refresh_cache-07ce931f-45ef-409b-b714-9f1cd47a3a88" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2157.445965] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Updated the network info_cache for instance {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10516}} [ 2157.446222] env[62824]: DEBUG oslo_vmware.api [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Task: {'id': task-2146460, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2157.446921] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2157.447157] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2157.447396] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2157.447638] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2157.447860] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2157.448015] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2157.448159] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62824) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11064}} [ 2157.448381] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager.update_available_resource {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2157.453514] env[62824]: DEBUG oslo_vmware.api [None req-5fb295c4-ec07-4626-81a5-775a8bd81d25 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': task-2146459, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.139948} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2157.453822] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-5fb295c4-ec07-4626-81a5-775a8bd81d25 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2157.454140] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5fb295c4-ec07-4626-81a5-775a8bd81d25 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 20f93b46-5e7e-4256-8cc1-e0a1b16740d5] Deleted contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2157.454271] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5fb295c4-ec07-4626-81a5-775a8bd81d25 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 20f93b46-5e7e-4256-8cc1-e0a1b16740d5] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2157.454574] env[62824]: INFO nova.compute.manager [None req-5fb295c4-ec07-4626-81a5-775a8bd81d25 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 20f93b46-5e7e-4256-8cc1-e0a1b16740d5] Took 1.28 seconds to destroy the instance on the hypervisor. [ 2157.455068] env[62824]: DEBUG oslo.service.loopingcall [None req-5fb295c4-ec07-4626-81a5-775a8bd81d25 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2157.455797] env[62824]: DEBUG nova.compute.manager [-] [instance: 20f93b46-5e7e-4256-8cc1-e0a1b16740d5] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2157.455797] env[62824]: DEBUG nova.network.neutron [-] [instance: 20f93b46-5e7e-4256-8cc1-e0a1b16740d5] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2157.549075] env[62824]: DEBUG nova.compute.manager [req-a63409e9-c472-4f31-96b2-1dff5e24bdeb req-74cbb388-a2bc-4622-aa71-3412e96a74bf service nova] [instance: 7a2a51c4-e558-4f5f-b82c-718bc12c1df5] Received event network-vif-deleted-9ad6111d-68a7-4700-9b92-1a7482806e37 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2157.549270] env[62824]: INFO nova.compute.manager [req-a63409e9-c472-4f31-96b2-1dff5e24bdeb req-74cbb388-a2bc-4622-aa71-3412e96a74bf service nova] [instance: 7a2a51c4-e558-4f5f-b82c-718bc12c1df5] Neutron deleted interface 9ad6111d-68a7-4700-9b92-1a7482806e37; detaching it from the instance and deleting it from the info cache [ 2157.549480] env[62824]: DEBUG nova.network.neutron [req-a63409e9-c472-4f31-96b2-1dff5e24bdeb req-74cbb388-a2bc-4622-aa71-3412e96a74bf service nova] [instance: 7a2a51c4-e558-4f5f-b82c-718bc12c1df5] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2157.749226] env[62824]: DEBUG oslo_vmware.api [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146457, 'name': PowerOnVM_Task, 'duration_secs': 0.529888} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2157.752203] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2157.752424] env[62824]: INFO nova.compute.manager [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Took 8.48 seconds to spawn the instance on the hypervisor. [ 2157.752605] env[62824]: DEBUG nova.compute.manager [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2157.754034] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bc45864-df45-441c-8ee8-a72c42e3a170 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2157.829240] env[62824]: DEBUG oslo_concurrency.lockutils [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Acquiring lock "refresh_cache-2afd1f18-234c-40b5-9ede-7413ad30dafe" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2157.829376] env[62824]: DEBUG oslo_concurrency.lockutils [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Acquired lock "refresh_cache-2afd1f18-234c-40b5-9ede-7413ad30dafe" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2157.829533] env[62824]: DEBUG nova.network.neutron [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 2afd1f18-234c-40b5-9ede-7413ad30dafe] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2157.944716] env[62824]: DEBUG oslo_vmware.api [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Task: {'id': task-2146460, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2157.951983] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2158.017671] env[62824]: DEBUG nova.network.neutron [-] [instance: 7a2a51c4-e558-4f5f-b82c-718bc12c1df5] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2158.054325] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0c2f6a44-919d-45cc-9dfb-4a0e521d50e5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2158.066487] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38189c79-56b7-485d-be45-22ceb66f44b7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2158.110555] env[62824]: DEBUG nova.compute.manager [req-a63409e9-c472-4f31-96b2-1dff5e24bdeb req-74cbb388-a2bc-4622-aa71-3412e96a74bf service nova] [instance: 7a2a51c4-e558-4f5f-b82c-718bc12c1df5] Detach interface failed, port_id=9ad6111d-68a7-4700-9b92-1a7482806e37, reason: Instance 7a2a51c4-e558-4f5f-b82c-718bc12c1df5 could not be found. {{(pid=62824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11482}} [ 2158.160269] env[62824]: DEBUG nova.compute.manager [req-e2c4170c-48ef-44bd-869f-b0a4e5a2df0b req-179c5b83-72c3-441a-8ecb-8549c27ca0f3 service nova] [instance: 2afd1f18-234c-40b5-9ede-7413ad30dafe] Received event network-vif-plugged-6d0cd52e-ace8-4459-860b-f270fa266268 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2158.160627] env[62824]: DEBUG oslo_concurrency.lockutils [req-e2c4170c-48ef-44bd-869f-b0a4e5a2df0b req-179c5b83-72c3-441a-8ecb-8549c27ca0f3 service nova] Acquiring lock "2afd1f18-234c-40b5-9ede-7413ad30dafe-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2158.160967] env[62824]: DEBUG oslo_concurrency.lockutils [req-e2c4170c-48ef-44bd-869f-b0a4e5a2df0b req-179c5b83-72c3-441a-8ecb-8549c27ca0f3 service nova] Lock "2afd1f18-234c-40b5-9ede-7413ad30dafe-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2158.161265] env[62824]: DEBUG oslo_concurrency.lockutils [req-e2c4170c-48ef-44bd-869f-b0a4e5a2df0b req-179c5b83-72c3-441a-8ecb-8549c27ca0f3 service nova] Lock "2afd1f18-234c-40b5-9ede-7413ad30dafe-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2158.161543] env[62824]: DEBUG nova.compute.manager [req-e2c4170c-48ef-44bd-869f-b0a4e5a2df0b req-179c5b83-72c3-441a-8ecb-8549c27ca0f3 service nova] [instance: 2afd1f18-234c-40b5-9ede-7413ad30dafe] No waiting events found dispatching network-vif-plugged-6d0cd52e-ace8-4459-860b-f270fa266268 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2158.161820] env[62824]: WARNING nova.compute.manager [req-e2c4170c-48ef-44bd-869f-b0a4e5a2df0b req-179c5b83-72c3-441a-8ecb-8549c27ca0f3 service nova] [instance: 2afd1f18-234c-40b5-9ede-7413ad30dafe] Received unexpected event network-vif-plugged-6d0cd52e-ace8-4459-860b-f270fa266268 for instance with vm_state building and task_state spawning. [ 2158.162115] env[62824]: DEBUG nova.compute.manager [req-e2c4170c-48ef-44bd-869f-b0a4e5a2df0b req-179c5b83-72c3-441a-8ecb-8549c27ca0f3 service nova] [instance: 2afd1f18-234c-40b5-9ede-7413ad30dafe] Received event network-changed-6d0cd52e-ace8-4459-860b-f270fa266268 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2158.162115] env[62824]: DEBUG nova.compute.manager [req-e2c4170c-48ef-44bd-869f-b0a4e5a2df0b req-179c5b83-72c3-441a-8ecb-8549c27ca0f3 service nova] [instance: 2afd1f18-234c-40b5-9ede-7413ad30dafe] Refreshing instance network info cache due to event network-changed-6d0cd52e-ace8-4459-860b-f270fa266268. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2158.163741] env[62824]: DEBUG oslo_concurrency.lockutils [req-e2c4170c-48ef-44bd-869f-b0a4e5a2df0b req-179c5b83-72c3-441a-8ecb-8549c27ca0f3 service nova] Acquiring lock "refresh_cache-2afd1f18-234c-40b5-9ede-7413ad30dafe" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2158.192267] env[62824]: DEBUG oslo_concurrency.lockutils [None req-07015f60-1e88-49e8-8469-e207c9f08bc2 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Acquiring lock "b0456b8a-348a-4503-a92c-58e5ab455d1c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2158.192794] env[62824]: DEBUG oslo_concurrency.lockutils [None req-07015f60-1e88-49e8-8469-e207c9f08bc2 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Lock "b0456b8a-348a-4503-a92c-58e5ab455d1c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2158.193042] env[62824]: DEBUG oslo_concurrency.lockutils [None req-07015f60-1e88-49e8-8469-e207c9f08bc2 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Acquiring lock "b0456b8a-348a-4503-a92c-58e5ab455d1c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2158.193285] env[62824]: DEBUG oslo_concurrency.lockutils [None req-07015f60-1e88-49e8-8469-e207c9f08bc2 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Lock "b0456b8a-348a-4503-a92c-58e5ab455d1c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2158.193470] env[62824]: DEBUG oslo_concurrency.lockutils [None req-07015f60-1e88-49e8-8469-e207c9f08bc2 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Lock "b0456b8a-348a-4503-a92c-58e5ab455d1c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2158.196196] env[62824]: INFO nova.compute.manager [None req-07015f60-1e88-49e8-8469-e207c9f08bc2 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: b0456b8a-348a-4503-a92c-58e5ab455d1c] Terminating instance [ 2158.252025] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc89149b-a453-4886-a010-a88b764f2b25 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2158.263110] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbf27e6c-f153-4166-81ed-09f47f0acd38 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2158.308713] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b21d04b-723c-4f1d-bfdd-88c24a64987e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2158.311636] env[62824]: INFO nova.compute.manager [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Took 41.46 seconds to build instance. [ 2158.322286] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2220d5a3-2d15-4f29-aa00-82c99a7f43f0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2158.348156] env[62824]: DEBUG nova.compute.provider_tree [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2158.388090] env[62824]: DEBUG nova.network.neutron [-] [instance: 20f93b46-5e7e-4256-8cc1-e0a1b16740d5] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2158.404662] env[62824]: DEBUG nova.network.neutron [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 2afd1f18-234c-40b5-9ede-7413ad30dafe] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2158.446085] env[62824]: DEBUG oslo_vmware.api [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Task: {'id': task-2146460, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.926021} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2158.446331] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] bd7f2ed5-fa08-4a48-9755-60b61791ea0b/bd7f2ed5-fa08-4a48-9755-60b61791ea0b.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2158.446549] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] [instance: bd7f2ed5-fa08-4a48-9755-60b61791ea0b] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2158.446800] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8c0e055a-449a-4bf2-9c75-86cb075a70e5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2158.454969] env[62824]: DEBUG oslo_vmware.api [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Waiting for the task: (returnval){ [ 2158.454969] env[62824]: value = "task-2146461" [ 2158.454969] env[62824]: _type = "Task" [ 2158.454969] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2158.464234] env[62824]: DEBUG oslo_vmware.api [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Task: {'id': task-2146461, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2158.520603] env[62824]: INFO nova.compute.manager [-] [instance: 7a2a51c4-e558-4f5f-b82c-718bc12c1df5] Took 1.26 seconds to deallocate network for instance. [ 2158.703526] env[62824]: DEBUG nova.compute.manager [None req-07015f60-1e88-49e8-8469-e207c9f08bc2 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: b0456b8a-348a-4503-a92c-58e5ab455d1c] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2158.703752] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-07015f60-1e88-49e8-8469-e207c9f08bc2 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: b0456b8a-348a-4503-a92c-58e5ab455d1c] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2158.704648] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8763fa5-c5c3-4446-97d8-4cd188e7a066 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2158.716259] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-07015f60-1e88-49e8-8469-e207c9f08bc2 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: b0456b8a-348a-4503-a92c-58e5ab455d1c] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2158.716259] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8387849a-978a-4315-a755-7e73562039c5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2158.724031] env[62824]: DEBUG oslo_vmware.api [None req-07015f60-1e88-49e8-8469-e207c9f08bc2 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Waiting for the task: (returnval){ [ 2158.724031] env[62824]: value = "task-2146462" [ 2158.724031] env[62824]: _type = "Task" [ 2158.724031] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2158.739912] env[62824]: DEBUG oslo_vmware.api [None req-07015f60-1e88-49e8-8469-e207c9f08bc2 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146462, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2158.791192] env[62824]: DEBUG nova.network.neutron [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 2afd1f18-234c-40b5-9ede-7413ad30dafe] Updating instance_info_cache with network_info: [{"id": "6d0cd52e-ace8-4459-860b-f270fa266268", "address": "fa:16:3e:5b:95:b5", "network": {"id": "60037885-38d9-4586-9fd7-4907321e9c7e", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-370573912-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2e59203d676d48c8bd8402af4eb58f65", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f856fca-9fb5-41ea-a057-ac4193bd323d", "external-id": "nsx-vlan-transportzone-148", "segmentation_id": 148, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6d0cd52e-ac", "ovs_interfaceid": "6d0cd52e-ace8-4459-860b-f270fa266268", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2158.814326] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ba1f335d-67af-4599-b3c2-1dec114873ff tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lock "fc77c44a-180c-46ff-9690-9072c6213c91" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 55.091s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2158.853647] env[62824]: DEBUG nova.scheduler.client.report [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2158.891936] env[62824]: INFO nova.compute.manager [-] [instance: 20f93b46-5e7e-4256-8cc1-e0a1b16740d5] Took 1.44 seconds to deallocate network for instance. [ 2158.968263] env[62824]: DEBUG oslo_vmware.api [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Task: {'id': task-2146461, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.303398} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2158.968263] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] [instance: bd7f2ed5-fa08-4a48-9755-60b61791ea0b] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2158.968468] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d45811a-9b63-4bcb-a2e3-c6dd4cae61b2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2158.994268] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] [instance: bd7f2ed5-fa08-4a48-9755-60b61791ea0b] Reconfiguring VM instance instance-0000004e to attach disk [datastore2] bd7f2ed5-fa08-4a48-9755-60b61791ea0b/bd7f2ed5-fa08-4a48-9755-60b61791ea0b.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2158.994979] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2d7d5a32-a626-4732-ba93-8b30a00ff2f8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2159.024210] env[62824]: DEBUG oslo_vmware.api [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Waiting for the task: (returnval){ [ 2159.024210] env[62824]: value = "task-2146463" [ 2159.024210] env[62824]: _type = "Task" [ 2159.024210] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2159.028993] env[62824]: DEBUG oslo_concurrency.lockutils [None req-aefa2097-549e-43b5-9152-0e8ce3b5da7a tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2159.036459] env[62824]: DEBUG oslo_vmware.api [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Task: {'id': task-2146463, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2159.234151] env[62824]: DEBUG oslo_vmware.api [None req-07015f60-1e88-49e8-8469-e207c9f08bc2 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146462, 'name': PowerOffVM_Task, 'duration_secs': 0.318026} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2159.234388] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-07015f60-1e88-49e8-8469-e207c9f08bc2 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: b0456b8a-348a-4503-a92c-58e5ab455d1c] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2159.234562] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-07015f60-1e88-49e8-8469-e207c9f08bc2 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: b0456b8a-348a-4503-a92c-58e5ab455d1c] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2159.234825] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cbd9dfe1-2438-4c53-a3aa-9bd8ddf8f41b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2159.294504] env[62824]: DEBUG oslo_concurrency.lockutils [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Releasing lock "refresh_cache-2afd1f18-234c-40b5-9ede-7413ad30dafe" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2159.294887] env[62824]: DEBUG nova.compute.manager [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 2afd1f18-234c-40b5-9ede-7413ad30dafe] Instance network_info: |[{"id": "6d0cd52e-ace8-4459-860b-f270fa266268", "address": "fa:16:3e:5b:95:b5", "network": {"id": "60037885-38d9-4586-9fd7-4907321e9c7e", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-370573912-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2e59203d676d48c8bd8402af4eb58f65", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f856fca-9fb5-41ea-a057-ac4193bd323d", "external-id": "nsx-vlan-transportzone-148", "segmentation_id": 148, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6d0cd52e-ac", "ovs_interfaceid": "6d0cd52e-ace8-4459-860b-f270fa266268", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2159.295174] env[62824]: DEBUG oslo_concurrency.lockutils [req-e2c4170c-48ef-44bd-869f-b0a4e5a2df0b req-179c5b83-72c3-441a-8ecb-8549c27ca0f3 service nova] Acquired lock "refresh_cache-2afd1f18-234c-40b5-9ede-7413ad30dafe" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2159.295510] env[62824]: DEBUG nova.network.neutron [req-e2c4170c-48ef-44bd-869f-b0a4e5a2df0b req-179c5b83-72c3-441a-8ecb-8549c27ca0f3 service nova] [instance: 2afd1f18-234c-40b5-9ede-7413ad30dafe] Refreshing network info cache for port 6d0cd52e-ace8-4459-860b-f270fa266268 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2159.296772] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 2afd1f18-234c-40b5-9ede-7413ad30dafe] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5b:95:b5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9f856fca-9fb5-41ea-a057-ac4193bd323d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6d0cd52e-ace8-4459-860b-f270fa266268', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2159.305052] env[62824]: DEBUG oslo.service.loopingcall [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2159.305700] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2afd1f18-234c-40b5-9ede-7413ad30dafe] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2159.306576] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-69266e52-c219-4fb1-a4dc-0d061c542835 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2159.329674] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2159.329674] env[62824]: value = "task-2146465" [ 2159.329674] env[62824]: _type = "Task" [ 2159.329674] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2159.339576] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146465, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2159.358179] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.677s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2159.358730] env[62824]: DEBUG nova.compute.manager [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] [instance: ab37b1a2-0012-40fb-9341-b613525e89cf] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2159.365023] env[62824]: DEBUG oslo_concurrency.lockutils [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.791s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2159.365023] env[62824]: INFO nova.compute.claims [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: da1ad739-b252-4e29-a22a-ab3bcab173ec] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2159.400657] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5fb295c4-ec07-4626-81a5-775a8bd81d25 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2159.409462] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-07015f60-1e88-49e8-8469-e207c9f08bc2 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: b0456b8a-348a-4503-a92c-58e5ab455d1c] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2159.409681] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-07015f60-1e88-49e8-8469-e207c9f08bc2 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: b0456b8a-348a-4503-a92c-58e5ab455d1c] Deleting contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2159.409862] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-07015f60-1e88-49e8-8469-e207c9f08bc2 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Deleting the datastore file [datastore1] b0456b8a-348a-4503-a92c-58e5ab455d1c {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2159.410352] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d856d964-6d68-40ca-81b1-f30edcf7e729 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2159.420104] env[62824]: DEBUG oslo_vmware.api [None req-07015f60-1e88-49e8-8469-e207c9f08bc2 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Waiting for the task: (returnval){ [ 2159.420104] env[62824]: value = "task-2146466" [ 2159.420104] env[62824]: _type = "Task" [ 2159.420104] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2159.428997] env[62824]: DEBUG oslo_vmware.api [None req-07015f60-1e88-49e8-8469-e207c9f08bc2 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146466, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2159.536309] env[62824]: DEBUG oslo_vmware.api [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Task: {'id': task-2146463, 'name': ReconfigVM_Task, 'duration_secs': 0.316182} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2159.536663] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] [instance: bd7f2ed5-fa08-4a48-9755-60b61791ea0b] Reconfigured VM instance instance-0000004e to attach disk [datastore2] bd7f2ed5-fa08-4a48-9755-60b61791ea0b/bd7f2ed5-fa08-4a48-9755-60b61791ea0b.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2159.537315] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7cddc643-7358-4187-9ade-8a579699094e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2159.548245] env[62824]: DEBUG oslo_vmware.api [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Waiting for the task: (returnval){ [ 2159.548245] env[62824]: value = "task-2146467" [ 2159.548245] env[62824]: _type = "Task" [ 2159.548245] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2159.562018] env[62824]: DEBUG oslo_vmware.api [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Task: {'id': task-2146467, 'name': Rename_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2159.843039] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146465, 'name': CreateVM_Task, 'duration_secs': 0.346633} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2159.843222] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2afd1f18-234c-40b5-9ede-7413ad30dafe] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2159.843968] env[62824]: DEBUG oslo_concurrency.lockutils [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2159.844168] env[62824]: DEBUG oslo_concurrency.lockutils [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2159.844495] env[62824]: DEBUG oslo_concurrency.lockutils [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2159.844750] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e6b33c66-25d2-4201-becb-c73e02297835 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2159.853095] env[62824]: DEBUG oslo_vmware.api [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Waiting for the task: (returnval){ [ 2159.853095] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52c20dda-99fe-ad5e-27bc-7c9bfbaf2725" [ 2159.853095] env[62824]: _type = "Task" [ 2159.853095] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2159.862129] env[62824]: DEBUG oslo_vmware.api [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52c20dda-99fe-ad5e-27bc-7c9bfbaf2725, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2159.867487] env[62824]: DEBUG nova.compute.utils [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2159.871136] env[62824]: DEBUG nova.compute.manager [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] [instance: ab37b1a2-0012-40fb-9341-b613525e89cf] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2159.871136] env[62824]: DEBUG nova.network.neutron [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] [instance: ab37b1a2-0012-40fb-9341-b613525e89cf] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2159.921827] env[62824]: DEBUG nova.policy [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9d7a3ba04f65461da9b9c8627b17ef8d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '330a1532c0a04bcb8e8712b08883deb1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2159.932951] env[62824]: DEBUG oslo_vmware.api [None req-07015f60-1e88-49e8-8469-e207c9f08bc2 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Task: {'id': task-2146466, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.197581} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2159.933233] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-07015f60-1e88-49e8-8469-e207c9f08bc2 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2159.933417] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-07015f60-1e88-49e8-8469-e207c9f08bc2 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: b0456b8a-348a-4503-a92c-58e5ab455d1c] Deleted contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2159.933633] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-07015f60-1e88-49e8-8469-e207c9f08bc2 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: b0456b8a-348a-4503-a92c-58e5ab455d1c] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2159.934156] env[62824]: INFO nova.compute.manager [None req-07015f60-1e88-49e8-8469-e207c9f08bc2 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] [instance: b0456b8a-348a-4503-a92c-58e5ab455d1c] Took 1.23 seconds to destroy the instance on the hypervisor. [ 2159.934156] env[62824]: DEBUG oslo.service.loopingcall [None req-07015f60-1e88-49e8-8469-e207c9f08bc2 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2159.934264] env[62824]: DEBUG nova.compute.manager [-] [instance: b0456b8a-348a-4503-a92c-58e5ab455d1c] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2159.934340] env[62824]: DEBUG nova.network.neutron [-] [instance: b0456b8a-348a-4503-a92c-58e5ab455d1c] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2160.059951] env[62824]: DEBUG oslo_vmware.api [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Task: {'id': task-2146467, 'name': Rename_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2160.085171] env[62824]: DEBUG nova.network.neutron [req-e2c4170c-48ef-44bd-869f-b0a4e5a2df0b req-179c5b83-72c3-441a-8ecb-8549c27ca0f3 service nova] [instance: 2afd1f18-234c-40b5-9ede-7413ad30dafe] Updated VIF entry in instance network info cache for port 6d0cd52e-ace8-4459-860b-f270fa266268. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2160.085626] env[62824]: DEBUG nova.network.neutron [req-e2c4170c-48ef-44bd-869f-b0a4e5a2df0b req-179c5b83-72c3-441a-8ecb-8549c27ca0f3 service nova] [instance: 2afd1f18-234c-40b5-9ede-7413ad30dafe] Updating instance_info_cache with network_info: [{"id": "6d0cd52e-ace8-4459-860b-f270fa266268", "address": "fa:16:3e:5b:95:b5", "network": {"id": "60037885-38d9-4586-9fd7-4907321e9c7e", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-370573912-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2e59203d676d48c8bd8402af4eb58f65", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f856fca-9fb5-41ea-a057-ac4193bd323d", "external-id": "nsx-vlan-transportzone-148", "segmentation_id": 148, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6d0cd52e-ac", "ovs_interfaceid": "6d0cd52e-ace8-4459-860b-f270fa266268", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2160.192612] env[62824]: DEBUG nova.compute.manager [req-ab0a1c8a-22dd-4ea4-afc4-1da6b4c4e7d5 req-c1cd146e-6e86-4757-93bb-52fa100ada59 service nova] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Received event network-changed-e285a842-d54b-4614-84ea-239e34f5b41f {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2160.192910] env[62824]: DEBUG nova.compute.manager [req-ab0a1c8a-22dd-4ea4-afc4-1da6b4c4e7d5 req-c1cd146e-6e86-4757-93bb-52fa100ada59 service nova] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Refreshing instance network info cache due to event network-changed-e285a842-d54b-4614-84ea-239e34f5b41f. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2160.193440] env[62824]: DEBUG oslo_concurrency.lockutils [req-ab0a1c8a-22dd-4ea4-afc4-1da6b4c4e7d5 req-c1cd146e-6e86-4757-93bb-52fa100ada59 service nova] Acquiring lock "refresh_cache-fc77c44a-180c-46ff-9690-9072c6213c91" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2160.193675] env[62824]: DEBUG oslo_concurrency.lockutils [req-ab0a1c8a-22dd-4ea4-afc4-1da6b4c4e7d5 req-c1cd146e-6e86-4757-93bb-52fa100ada59 service nova] Acquired lock "refresh_cache-fc77c44a-180c-46ff-9690-9072c6213c91" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2160.193959] env[62824]: DEBUG nova.network.neutron [req-ab0a1c8a-22dd-4ea4-afc4-1da6b4c4e7d5 req-c1cd146e-6e86-4757-93bb-52fa100ada59 service nova] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Refreshing network info cache for port e285a842-d54b-4614-84ea-239e34f5b41f {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2160.366302] env[62824]: DEBUG oslo_vmware.api [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52c20dda-99fe-ad5e-27bc-7c9bfbaf2725, 'name': SearchDatastore_Task, 'duration_secs': 0.010666} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2160.367101] env[62824]: DEBUG nova.network.neutron [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] [instance: ab37b1a2-0012-40fb-9341-b613525e89cf] Successfully created port: a7172f1f-6473-48fe-bec7-5c84411c8fd2 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2160.369248] env[62824]: DEBUG oslo_concurrency.lockutils [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2160.369505] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 2afd1f18-234c-40b5-9ede-7413ad30dafe] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2160.369769] env[62824]: DEBUG oslo_concurrency.lockutils [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2160.369861] env[62824]: DEBUG oslo_concurrency.lockutils [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2160.370047] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2160.370582] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e70e11f2-3a6c-4c03-a981-7a35ddf9200a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2160.373105] env[62824]: DEBUG nova.compute.manager [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] [instance: ab37b1a2-0012-40fb-9341-b613525e89cf] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2160.388422] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2160.388422] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2160.388707] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-277a566e-9d26-40e7-b70b-7fb0ad4db3cd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2160.395302] env[62824]: DEBUG oslo_vmware.api [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Waiting for the task: (returnval){ [ 2160.395302] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5286859b-9b90-187f-977d-e5e125e89178" [ 2160.395302] env[62824]: _type = "Task" [ 2160.395302] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2160.408088] env[62824]: DEBUG oslo_vmware.api [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5286859b-9b90-187f-977d-e5e125e89178, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2160.426140] env[62824]: DEBUG nova.compute.manager [req-d457a89f-6d36-4acb-9b44-18ac1dea094d req-86d7d7b1-feaf-4014-af94-817c3184d375 service nova] [instance: b0456b8a-348a-4503-a92c-58e5ab455d1c] Received event network-vif-deleted-0534c02e-d202-458d-9289-7d0ad2c5596f {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2160.426140] env[62824]: INFO nova.compute.manager [req-d457a89f-6d36-4acb-9b44-18ac1dea094d req-86d7d7b1-feaf-4014-af94-817c3184d375 service nova] [instance: b0456b8a-348a-4503-a92c-58e5ab455d1c] Neutron deleted interface 0534c02e-d202-458d-9289-7d0ad2c5596f; detaching it from the instance and deleting it from the info cache [ 2160.426140] env[62824]: DEBUG nova.network.neutron [req-d457a89f-6d36-4acb-9b44-18ac1dea094d req-86d7d7b1-feaf-4014-af94-817c3184d375 service nova] [instance: b0456b8a-348a-4503-a92c-58e5ab455d1c] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2160.564087] env[62824]: DEBUG oslo_vmware.api [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Task: {'id': task-2146467, 'name': Rename_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2160.591191] env[62824]: DEBUG oslo_concurrency.lockutils [req-e2c4170c-48ef-44bd-869f-b0a4e5a2df0b req-179c5b83-72c3-441a-8ecb-8549c27ca0f3 service nova] Releasing lock "refresh_cache-2afd1f18-234c-40b5-9ede-7413ad30dafe" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2160.591971] env[62824]: DEBUG nova.compute.manager [req-e2c4170c-48ef-44bd-869f-b0a4e5a2df0b req-179c5b83-72c3-441a-8ecb-8549c27ca0f3 service nova] [instance: 20f93b46-5e7e-4256-8cc1-e0a1b16740d5] Received event network-vif-deleted-32e4bbdc-8153-4bb8-9c2d-669a9c40b380 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2160.591971] env[62824]: INFO nova.compute.manager [req-e2c4170c-48ef-44bd-869f-b0a4e5a2df0b req-179c5b83-72c3-441a-8ecb-8549c27ca0f3 service nova] [instance: 20f93b46-5e7e-4256-8cc1-e0a1b16740d5] Neutron deleted interface 32e4bbdc-8153-4bb8-9c2d-669a9c40b380; detaching it from the instance and deleting it from the info cache [ 2160.591971] env[62824]: DEBUG nova.network.neutron [req-e2c4170c-48ef-44bd-869f-b0a4e5a2df0b req-179c5b83-72c3-441a-8ecb-8549c27ca0f3 service nova] [instance: 20f93b46-5e7e-4256-8cc1-e0a1b16740d5] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2160.835538] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f032014b-7c7b-471e-a5f3-ec0491b1bb47 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2160.844257] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cfdef6e-b4e5-4c02-8d60-347ac9907fe2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2160.848028] env[62824]: DEBUG nova.network.neutron [-] [instance: b0456b8a-348a-4503-a92c-58e5ab455d1c] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2160.885669] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8826140d-96ba-4ce3-999f-3f724fdf4905 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2160.895401] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0341f70-af9d-4da8-ae70-9ad7b000dcd3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2160.913801] env[62824]: DEBUG nova.compute.provider_tree [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2160.918129] env[62824]: DEBUG oslo_vmware.api [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5286859b-9b90-187f-977d-e5e125e89178, 'name': SearchDatastore_Task, 'duration_secs': 0.009872} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2160.919611] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bf63e2c9-f2b6-4844-8a94-19b7ba8fd9f2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2160.928353] env[62824]: DEBUG oslo_vmware.api [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Waiting for the task: (returnval){ [ 2160.928353] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52ae70c5-af30-f5ec-65d4-ac1e6375474b" [ 2160.928353] env[62824]: _type = "Task" [ 2160.928353] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2160.929233] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-504bd4c4-a557-42f0-8e49-628d563e744d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2160.944717] env[62824]: DEBUG oslo_vmware.api [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52ae70c5-af30-f5ec-65d4-ac1e6375474b, 'name': SearchDatastore_Task, 'duration_secs': 0.011622} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2160.945967] env[62824]: DEBUG oslo_concurrency.lockutils [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2160.946239] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 2afd1f18-234c-40b5-9ede-7413ad30dafe/2afd1f18-234c-40b5-9ede-7413ad30dafe.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2160.946775] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b38d1b5d-6cda-461e-8707-788b2dbb3190 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2160.951171] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a32397f-a0cc-4878-a632-b7c1185b7bed {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2160.961681] env[62824]: DEBUG nova.network.neutron [req-ab0a1c8a-22dd-4ea4-afc4-1da6b4c4e7d5 req-c1cd146e-6e86-4757-93bb-52fa100ada59 service nova] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Updated VIF entry in instance network info cache for port e285a842-d54b-4614-84ea-239e34f5b41f. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2160.961992] env[62824]: DEBUG nova.network.neutron [req-ab0a1c8a-22dd-4ea4-afc4-1da6b4c4e7d5 req-c1cd146e-6e86-4757-93bb-52fa100ada59 service nova] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Updating instance_info_cache with network_info: [{"id": "e285a842-d54b-4614-84ea-239e34f5b41f", "address": "fa:16:3e:9d:17:cd", "network": {"id": "aa1c09e5-8d72-43ad-a903-a0f5e711da80", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1643616572-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c987c5f844be42c99cc9d57a9ca84c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9bc2632-36f9-4912-8782-8bbb789f909d", "external-id": "nsx-vlan-transportzone-897", "segmentation_id": 897, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape285a842-d5", "ovs_interfaceid": "e285a842-d54b-4614-84ea-239e34f5b41f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2160.971915] env[62824]: DEBUG oslo_vmware.api [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Waiting for the task: (returnval){ [ 2160.971915] env[62824]: value = "task-2146468" [ 2160.971915] env[62824]: _type = "Task" [ 2160.971915] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2160.992357] env[62824]: DEBUG nova.compute.manager [req-d457a89f-6d36-4acb-9b44-18ac1dea094d req-86d7d7b1-feaf-4014-af94-817c3184d375 service nova] [instance: b0456b8a-348a-4503-a92c-58e5ab455d1c] Detach interface failed, port_id=0534c02e-d202-458d-9289-7d0ad2c5596f, reason: Instance b0456b8a-348a-4503-a92c-58e5ab455d1c could not be found. {{(pid=62824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11482}} [ 2160.996252] env[62824]: DEBUG oslo_vmware.api [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Task: {'id': task-2146468, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2161.060714] env[62824]: DEBUG oslo_vmware.api [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Task: {'id': task-2146467, 'name': Rename_Task, 'duration_secs': 1.196364} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2161.061009] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] [instance: bd7f2ed5-fa08-4a48-9755-60b61791ea0b] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2161.061376] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-17600d63-f2c5-4124-8331-e88e64655fa1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2161.068432] env[62824]: DEBUG oslo_vmware.api [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Waiting for the task: (returnval){ [ 2161.068432] env[62824]: value = "task-2146469" [ 2161.068432] env[62824]: _type = "Task" [ 2161.068432] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2161.076683] env[62824]: DEBUG oslo_vmware.api [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Task: {'id': task-2146469, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2161.096054] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-24adf11a-c9ad-4574-8ef9-8e148219a3e3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2161.105882] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf8c6f4c-e3ed-45bb-9e41-da4ffc38c841 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2161.150905] env[62824]: DEBUG nova.compute.manager [req-e2c4170c-48ef-44bd-869f-b0a4e5a2df0b req-179c5b83-72c3-441a-8ecb-8549c27ca0f3 service nova] [instance: 20f93b46-5e7e-4256-8cc1-e0a1b16740d5] Detach interface failed, port_id=32e4bbdc-8153-4bb8-9c2d-669a9c40b380, reason: Instance 20f93b46-5e7e-4256-8cc1-e0a1b16740d5 could not be found. {{(pid=62824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11482}} [ 2161.350858] env[62824]: INFO nova.compute.manager [-] [instance: b0456b8a-348a-4503-a92c-58e5ab455d1c] Took 1.42 seconds to deallocate network for instance. [ 2161.390174] env[62824]: DEBUG nova.compute.manager [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] [instance: ab37b1a2-0012-40fb-9341-b613525e89cf] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2161.418619] env[62824]: DEBUG nova.virt.hardware [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2161.418878] env[62824]: DEBUG nova.virt.hardware [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2161.419046] env[62824]: DEBUG nova.virt.hardware [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2161.419230] env[62824]: DEBUG nova.virt.hardware [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2161.419375] env[62824]: DEBUG nova.virt.hardware [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2161.419518] env[62824]: DEBUG nova.virt.hardware [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2161.419735] env[62824]: DEBUG nova.virt.hardware [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2161.419887] env[62824]: DEBUG nova.virt.hardware [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2161.420066] env[62824]: DEBUG nova.virt.hardware [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2161.420235] env[62824]: DEBUG nova.virt.hardware [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2161.420408] env[62824]: DEBUG nova.virt.hardware [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2161.421270] env[62824]: DEBUG nova.scheduler.client.report [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2161.424907] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d484a407-2a37-4d56-a7e3-8f302c050d74 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2161.434343] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29f177e1-0a7e-44f0-b7ca-ac2b0164bf4c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2161.465268] env[62824]: DEBUG oslo_concurrency.lockutils [req-ab0a1c8a-22dd-4ea4-afc4-1da6b4c4e7d5 req-c1cd146e-6e86-4757-93bb-52fa100ada59 service nova] Releasing lock "refresh_cache-fc77c44a-180c-46ff-9690-9072c6213c91" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2161.482479] env[62824]: DEBUG oslo_vmware.api [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Task: {'id': task-2146468, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.440524} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2161.482761] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 2afd1f18-234c-40b5-9ede-7413ad30dafe/2afd1f18-234c-40b5-9ede-7413ad30dafe.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2161.482997] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 2afd1f18-234c-40b5-9ede-7413ad30dafe] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2161.483309] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d171e5e8-4113-4597-aeed-ac2186bedeee {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2161.490657] env[62824]: DEBUG oslo_vmware.api [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Waiting for the task: (returnval){ [ 2161.490657] env[62824]: value = "task-2146470" [ 2161.490657] env[62824]: _type = "Task" [ 2161.490657] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2161.500824] env[62824]: DEBUG oslo_vmware.api [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Task: {'id': task-2146470, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2161.578979] env[62824]: DEBUG oslo_vmware.api [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Task: {'id': task-2146469, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2161.859290] env[62824]: DEBUG oslo_concurrency.lockutils [None req-07015f60-1e88-49e8-8469-e207c9f08bc2 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2161.905509] env[62824]: DEBUG nova.network.neutron [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] [instance: ab37b1a2-0012-40fb-9341-b613525e89cf] Successfully updated port: a7172f1f-6473-48fe-bec7-5c84411c8fd2 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2161.931089] env[62824]: DEBUG oslo_concurrency.lockutils [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.569s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2161.931344] env[62824]: DEBUG nova.compute.manager [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: da1ad739-b252-4e29-a22a-ab3bcab173ec] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2161.934541] env[62824]: DEBUG oslo_concurrency.lockutils [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.315s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2161.935965] env[62824]: INFO nova.compute.claims [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2162.002571] env[62824]: DEBUG oslo_vmware.api [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Task: {'id': task-2146470, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061586} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2162.002571] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 2afd1f18-234c-40b5-9ede-7413ad30dafe] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2162.003081] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ea72d1b-230a-4d80-86b1-c7823383b370 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2162.039789] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 2afd1f18-234c-40b5-9ede-7413ad30dafe] Reconfiguring VM instance instance-0000004f to attach disk [datastore2] 2afd1f18-234c-40b5-9ede-7413ad30dafe/2afd1f18-234c-40b5-9ede-7413ad30dafe.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2162.040124] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ab5f91cd-44cf-45f6-8155-35343e593b9f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2162.061016] env[62824]: DEBUG oslo_vmware.api [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Waiting for the task: (returnval){ [ 2162.061016] env[62824]: value = "task-2146471" [ 2162.061016] env[62824]: _type = "Task" [ 2162.061016] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2162.070058] env[62824]: DEBUG oslo_vmware.api [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Task: {'id': task-2146471, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2162.077811] env[62824]: DEBUG oslo_vmware.api [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Task: {'id': task-2146469, 'name': PowerOnVM_Task, 'duration_secs': 0.592863} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2162.078016] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] [instance: bd7f2ed5-fa08-4a48-9755-60b61791ea0b] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2162.078223] env[62824]: INFO nova.compute.manager [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] [instance: bd7f2ed5-fa08-4a48-9755-60b61791ea0b] Took 10.10 seconds to spawn the instance on the hypervisor. [ 2162.078420] env[62824]: DEBUG nova.compute.manager [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] [instance: bd7f2ed5-fa08-4a48-9755-60b61791ea0b] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2162.079206] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ceb64e19-53e2-4f25-b176-4cea46e65a21 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2162.408424] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Acquiring lock "refresh_cache-ab37b1a2-0012-40fb-9341-b613525e89cf" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2162.408708] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Acquired lock "refresh_cache-ab37b1a2-0012-40fb-9341-b613525e89cf" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2162.408776] env[62824]: DEBUG nova.network.neutron [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] [instance: ab37b1a2-0012-40fb-9341-b613525e89cf] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2162.436428] env[62824]: DEBUG nova.compute.utils [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2162.437821] env[62824]: DEBUG nova.compute.manager [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: da1ad739-b252-4e29-a22a-ab3bcab173ec] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2162.437992] env[62824]: DEBUG nova.network.neutron [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: da1ad739-b252-4e29-a22a-ab3bcab173ec] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2162.454247] env[62824]: DEBUG nova.compute.manager [req-fbe6c4b9-1808-4136-b74f-341eb6f71526 req-e7c4ea70-ca1b-43e2-94e2-dbddec404d07 service nova] [instance: ab37b1a2-0012-40fb-9341-b613525e89cf] Received event network-vif-plugged-a7172f1f-6473-48fe-bec7-5c84411c8fd2 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2162.454247] env[62824]: DEBUG oslo_concurrency.lockutils [req-fbe6c4b9-1808-4136-b74f-341eb6f71526 req-e7c4ea70-ca1b-43e2-94e2-dbddec404d07 service nova] Acquiring lock "ab37b1a2-0012-40fb-9341-b613525e89cf-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2162.454247] env[62824]: DEBUG oslo_concurrency.lockutils [req-fbe6c4b9-1808-4136-b74f-341eb6f71526 req-e7c4ea70-ca1b-43e2-94e2-dbddec404d07 service nova] Lock "ab37b1a2-0012-40fb-9341-b613525e89cf-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2162.454247] env[62824]: DEBUG oslo_concurrency.lockutils [req-fbe6c4b9-1808-4136-b74f-341eb6f71526 req-e7c4ea70-ca1b-43e2-94e2-dbddec404d07 service nova] Lock "ab37b1a2-0012-40fb-9341-b613525e89cf-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2162.454445] env[62824]: DEBUG nova.compute.manager [req-fbe6c4b9-1808-4136-b74f-341eb6f71526 req-e7c4ea70-ca1b-43e2-94e2-dbddec404d07 service nova] [instance: ab37b1a2-0012-40fb-9341-b613525e89cf] No waiting events found dispatching network-vif-plugged-a7172f1f-6473-48fe-bec7-5c84411c8fd2 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2162.454569] env[62824]: WARNING nova.compute.manager [req-fbe6c4b9-1808-4136-b74f-341eb6f71526 req-e7c4ea70-ca1b-43e2-94e2-dbddec404d07 service nova] [instance: ab37b1a2-0012-40fb-9341-b613525e89cf] Received unexpected event network-vif-plugged-a7172f1f-6473-48fe-bec7-5c84411c8fd2 for instance with vm_state building and task_state spawning. [ 2162.454698] env[62824]: DEBUG nova.compute.manager [req-fbe6c4b9-1808-4136-b74f-341eb6f71526 req-e7c4ea70-ca1b-43e2-94e2-dbddec404d07 service nova] [instance: ab37b1a2-0012-40fb-9341-b613525e89cf] Received event network-changed-a7172f1f-6473-48fe-bec7-5c84411c8fd2 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2162.454994] env[62824]: DEBUG nova.compute.manager [req-fbe6c4b9-1808-4136-b74f-341eb6f71526 req-e7c4ea70-ca1b-43e2-94e2-dbddec404d07 service nova] [instance: ab37b1a2-0012-40fb-9341-b613525e89cf] Refreshing instance network info cache due to event network-changed-a7172f1f-6473-48fe-bec7-5c84411c8fd2. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2162.455207] env[62824]: DEBUG oslo_concurrency.lockutils [req-fbe6c4b9-1808-4136-b74f-341eb6f71526 req-e7c4ea70-ca1b-43e2-94e2-dbddec404d07 service nova] Acquiring lock "refresh_cache-ab37b1a2-0012-40fb-9341-b613525e89cf" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2162.487385] env[62824]: DEBUG nova.policy [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4c914343569a462d9b8564f3376c2f54', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '10ee5fe93d584725ace90e76605a1b2e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2162.572954] env[62824]: DEBUG oslo_vmware.api [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Task: {'id': task-2146471, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2162.599457] env[62824]: INFO nova.compute.manager [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] [instance: bd7f2ed5-fa08-4a48-9755-60b61791ea0b] Took 45.69 seconds to build instance. [ 2162.741313] env[62824]: DEBUG nova.network.neutron [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: da1ad739-b252-4e29-a22a-ab3bcab173ec] Successfully created port: 734a463a-d647-4c05-b8a9-29a3dd47e376 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2162.941847] env[62824]: DEBUG nova.compute.manager [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: da1ad739-b252-4e29-a22a-ab3bcab173ec] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2162.953578] env[62824]: DEBUG nova.network.neutron [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] [instance: ab37b1a2-0012-40fb-9341-b613525e89cf] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2163.075907] env[62824]: DEBUG oslo_vmware.api [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Task: {'id': task-2146471, 'name': ReconfigVM_Task, 'duration_secs': 0.774196} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2163.075907] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 2afd1f18-234c-40b5-9ede-7413ad30dafe] Reconfigured VM instance instance-0000004f to attach disk [datastore2] 2afd1f18-234c-40b5-9ede-7413ad30dafe/2afd1f18-234c-40b5-9ede-7413ad30dafe.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2163.076399] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0972f5a4-2679-417e-ac16-81bb5b0856c8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2163.085884] env[62824]: DEBUG oslo_vmware.api [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Waiting for the task: (returnval){ [ 2163.085884] env[62824]: value = "task-2146472" [ 2163.085884] env[62824]: _type = "Task" [ 2163.085884] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2163.097906] env[62824]: DEBUG oslo_vmware.api [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Task: {'id': task-2146472, 'name': Rename_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2163.103452] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2007c676-c6e4-419a-8ddc-a508749362d4 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Lock "bd7f2ed5-fa08-4a48-9755-60b61791ea0b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 57.111s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2163.158566] env[62824]: DEBUG nova.network.neutron [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] [instance: ab37b1a2-0012-40fb-9341-b613525e89cf] Updating instance_info_cache with network_info: [{"id": "a7172f1f-6473-48fe-bec7-5c84411c8fd2", "address": "fa:16:3e:40:67:d8", "network": {"id": "a7b19303-00d1-4e23-8595-8aac1cce4ce0", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-1700104384-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "330a1532c0a04bcb8e8712b08883deb1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa7172f1f-64", "ovs_interfaceid": "a7172f1f-6473-48fe-bec7-5c84411c8fd2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2163.395578] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b05b36b-a613-4667-8289-23dbfa4f2c45 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2163.404151] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e9c6475-cfeb-46e5-9442-455179257a79 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2163.439641] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-424c2b5a-4fc9-4b29-b0fa-6a9ec65c1f7b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2163.455209] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a432fb8a-3f60-4a65-af63-21922c3e6cd8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2163.470939] env[62824]: DEBUG nova.compute.provider_tree [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2163.596632] env[62824]: DEBUG oslo_vmware.api [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Task: {'id': task-2146472, 'name': Rename_Task, 'duration_secs': 0.149474} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2163.596930] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 2afd1f18-234c-40b5-9ede-7413ad30dafe] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2163.597189] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-46b81ef2-722f-4acf-a5a8-9e5ad9b1fecd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2163.604146] env[62824]: DEBUG oslo_vmware.api [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Waiting for the task: (returnval){ [ 2163.604146] env[62824]: value = "task-2146473" [ 2163.604146] env[62824]: _type = "Task" [ 2163.604146] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2163.612382] env[62824]: DEBUG oslo_vmware.api [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Task: {'id': task-2146473, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2163.660647] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Releasing lock "refresh_cache-ab37b1a2-0012-40fb-9341-b613525e89cf" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2163.661091] env[62824]: DEBUG nova.compute.manager [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] [instance: ab37b1a2-0012-40fb-9341-b613525e89cf] Instance network_info: |[{"id": "a7172f1f-6473-48fe-bec7-5c84411c8fd2", "address": "fa:16:3e:40:67:d8", "network": {"id": "a7b19303-00d1-4e23-8595-8aac1cce4ce0", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-1700104384-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "330a1532c0a04bcb8e8712b08883deb1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa7172f1f-64", "ovs_interfaceid": "a7172f1f-6473-48fe-bec7-5c84411c8fd2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2163.661466] env[62824]: DEBUG oslo_concurrency.lockutils [req-fbe6c4b9-1808-4136-b74f-341eb6f71526 req-e7c4ea70-ca1b-43e2-94e2-dbddec404d07 service nova] Acquired lock "refresh_cache-ab37b1a2-0012-40fb-9341-b613525e89cf" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2163.661757] env[62824]: DEBUG nova.network.neutron [req-fbe6c4b9-1808-4136-b74f-341eb6f71526 req-e7c4ea70-ca1b-43e2-94e2-dbddec404d07 service nova] [instance: ab37b1a2-0012-40fb-9341-b613525e89cf] Refreshing network info cache for port a7172f1f-6473-48fe-bec7-5c84411c8fd2 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2163.663138] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] [instance: ab37b1a2-0012-40fb-9341-b613525e89cf] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:40:67:d8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '92cdccfd-4b10-4024-b724-5f22792dd4de', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a7172f1f-6473-48fe-bec7-5c84411c8fd2', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2163.671028] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Creating folder: Project (330a1532c0a04bcb8e8712b08883deb1). Parent ref: group-v438503. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 2163.671165] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0600eb3d-c2a5-4e7b-a193-9a2ee57c73e2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2163.684633] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Created folder: Project (330a1532c0a04bcb8e8712b08883deb1) in parent group-v438503. [ 2163.684847] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Creating folder: Instances. Parent ref: group-v438704. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 2163.685169] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-96865ca8-1603-4c56-8815-1a2d6eefe3ea {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2163.698902] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Created folder: Instances in parent group-v438704. [ 2163.699238] env[62824]: DEBUG oslo.service.loopingcall [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2163.699468] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ab37b1a2-0012-40fb-9341-b613525e89cf] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2163.699719] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a9904dd4-31a3-40fa-94d5-c14c215063cf {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2163.723052] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2163.723052] env[62824]: value = "task-2146476" [ 2163.723052] env[62824]: _type = "Task" [ 2163.723052] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2163.731551] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146476, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2163.947167] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3a1b2b23-923e-4532-b132-fcd23342ce08 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Acquiring lock "bd7f2ed5-fa08-4a48-9755-60b61791ea0b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2163.947445] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3a1b2b23-923e-4532-b132-fcd23342ce08 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Lock "bd7f2ed5-fa08-4a48-9755-60b61791ea0b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2163.947665] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3a1b2b23-923e-4532-b132-fcd23342ce08 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Acquiring lock "bd7f2ed5-fa08-4a48-9755-60b61791ea0b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2163.947857] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3a1b2b23-923e-4532-b132-fcd23342ce08 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Lock "bd7f2ed5-fa08-4a48-9755-60b61791ea0b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2163.948038] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3a1b2b23-923e-4532-b132-fcd23342ce08 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Lock "bd7f2ed5-fa08-4a48-9755-60b61791ea0b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2163.950232] env[62824]: INFO nova.compute.manager [None req-3a1b2b23-923e-4532-b132-fcd23342ce08 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] [instance: bd7f2ed5-fa08-4a48-9755-60b61791ea0b] Terminating instance [ 2163.961717] env[62824]: DEBUG nova.compute.manager [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: da1ad739-b252-4e29-a22a-ab3bcab173ec] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2163.974265] env[62824]: DEBUG nova.scheduler.client.report [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2163.996058] env[62824]: DEBUG nova.virt.hardware [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='68920724353aacbe0794fa387f73ec8e',container_format='bare',created_at=2025-01-16T14:55:03Z,direct_url=,disk_format='vmdk',id=758d9d98-91b8-46a1-9c95-5705c074be1a,min_disk=1,min_ram=0,name='tempest-test-snap-624807388',owner='10ee5fe93d584725ace90e76605a1b2e',properties=ImageMetaProps,protected=,size=21334016,status='active',tags=,updated_at=2025-01-16T14:55:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2163.996297] env[62824]: DEBUG nova.virt.hardware [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2163.996422] env[62824]: DEBUG nova.virt.hardware [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2163.996638] env[62824]: DEBUG nova.virt.hardware [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2163.996813] env[62824]: DEBUG nova.virt.hardware [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2163.996973] env[62824]: DEBUG nova.virt.hardware [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2163.997219] env[62824]: DEBUG nova.virt.hardware [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2163.997524] env[62824]: DEBUG nova.virt.hardware [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2163.997712] env[62824]: DEBUG nova.virt.hardware [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2163.997894] env[62824]: DEBUG nova.virt.hardware [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2163.998102] env[62824]: DEBUG nova.virt.hardware [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2163.999930] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bef2263-2842-4b52-80d5-24e9c088832e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2164.010681] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2120d22-3e68-4a4e-b379-4d79bbd68dca {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2164.117310] env[62824]: DEBUG oslo_vmware.api [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Task: {'id': task-2146473, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2164.233747] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146476, 'name': CreateVM_Task, 'duration_secs': 0.429847} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2164.234173] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ab37b1a2-0012-40fb-9341-b613525e89cf] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2164.234970] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2164.235163] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2164.235530] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2164.236146] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-01433bdc-5898-49de-a3e9-910db296923d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2164.241560] env[62824]: DEBUG oslo_vmware.api [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Waiting for the task: (returnval){ [ 2164.241560] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52d1f597-c936-ee7d-9776-a10b9e2b39f6" [ 2164.241560] env[62824]: _type = "Task" [ 2164.241560] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2164.256368] env[62824]: DEBUG oslo_vmware.api [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52d1f597-c936-ee7d-9776-a10b9e2b39f6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2164.454826] env[62824]: DEBUG nova.compute.manager [None req-3a1b2b23-923e-4532-b132-fcd23342ce08 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] [instance: bd7f2ed5-fa08-4a48-9755-60b61791ea0b] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2164.455142] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-3a1b2b23-923e-4532-b132-fcd23342ce08 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] [instance: bd7f2ed5-fa08-4a48-9755-60b61791ea0b] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2164.456080] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6852ec15-7c00-4ed5-8848-bbfdb6b293f3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2164.465821] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a1b2b23-923e-4532-b132-fcd23342ce08 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] [instance: bd7f2ed5-fa08-4a48-9755-60b61791ea0b] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2164.466063] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4a3eb1b4-2f7e-4619-92df-17b4a7c69a46 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2164.475963] env[62824]: DEBUG oslo_vmware.api [None req-3a1b2b23-923e-4532-b132-fcd23342ce08 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Waiting for the task: (returnval){ [ 2164.475963] env[62824]: value = "task-2146477" [ 2164.475963] env[62824]: _type = "Task" [ 2164.475963] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2164.482183] env[62824]: DEBUG nova.compute.manager [req-dff3db2e-baf0-47db-b66e-af1fdf1e0434 req-34b2678c-03ad-48a1-b4af-18aa18ae4590 service nova] [instance: da1ad739-b252-4e29-a22a-ab3bcab173ec] Received event network-vif-plugged-734a463a-d647-4c05-b8a9-29a3dd47e376 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2164.482499] env[62824]: DEBUG oslo_concurrency.lockutils [req-dff3db2e-baf0-47db-b66e-af1fdf1e0434 req-34b2678c-03ad-48a1-b4af-18aa18ae4590 service nova] Acquiring lock "da1ad739-b252-4e29-a22a-ab3bcab173ec-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2164.482913] env[62824]: DEBUG oslo_concurrency.lockutils [req-dff3db2e-baf0-47db-b66e-af1fdf1e0434 req-34b2678c-03ad-48a1-b4af-18aa18ae4590 service nova] Lock "da1ad739-b252-4e29-a22a-ab3bcab173ec-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2164.483146] env[62824]: DEBUG oslo_concurrency.lockutils [req-dff3db2e-baf0-47db-b66e-af1fdf1e0434 req-34b2678c-03ad-48a1-b4af-18aa18ae4590 service nova] Lock "da1ad739-b252-4e29-a22a-ab3bcab173ec-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2164.483324] env[62824]: DEBUG nova.compute.manager [req-dff3db2e-baf0-47db-b66e-af1fdf1e0434 req-34b2678c-03ad-48a1-b4af-18aa18ae4590 service nova] [instance: da1ad739-b252-4e29-a22a-ab3bcab173ec] No waiting events found dispatching network-vif-plugged-734a463a-d647-4c05-b8a9-29a3dd47e376 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2164.483527] env[62824]: WARNING nova.compute.manager [req-dff3db2e-baf0-47db-b66e-af1fdf1e0434 req-34b2678c-03ad-48a1-b4af-18aa18ae4590 service nova] [instance: da1ad739-b252-4e29-a22a-ab3bcab173ec] Received unexpected event network-vif-plugged-734a463a-d647-4c05-b8a9-29a3dd47e376 for instance with vm_state building and task_state spawning. [ 2164.484702] env[62824]: DEBUG oslo_concurrency.lockutils [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.550s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2164.485194] env[62824]: DEBUG nova.compute.manager [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2164.488120] env[62824]: DEBUG oslo_concurrency.lockutils [None req-61bb4e80-84d9-4e19-b9c0-195ff35ecb41 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.541s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2164.488120] env[62824]: DEBUG nova.objects.instance [None req-61bb4e80-84d9-4e19-b9c0-195ff35ecb41 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lazy-loading 'resources' on Instance uuid bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2164.495230] env[62824]: DEBUG oslo_vmware.api [None req-3a1b2b23-923e-4532-b132-fcd23342ce08 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Task: {'id': task-2146477, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2164.543296] env[62824]: DEBUG nova.network.neutron [req-fbe6c4b9-1808-4136-b74f-341eb6f71526 req-e7c4ea70-ca1b-43e2-94e2-dbddec404d07 service nova] [instance: ab37b1a2-0012-40fb-9341-b613525e89cf] Updated VIF entry in instance network info cache for port a7172f1f-6473-48fe-bec7-5c84411c8fd2. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2164.543296] env[62824]: DEBUG nova.network.neutron [req-fbe6c4b9-1808-4136-b74f-341eb6f71526 req-e7c4ea70-ca1b-43e2-94e2-dbddec404d07 service nova] [instance: ab37b1a2-0012-40fb-9341-b613525e89cf] Updating instance_info_cache with network_info: [{"id": "a7172f1f-6473-48fe-bec7-5c84411c8fd2", "address": "fa:16:3e:40:67:d8", "network": {"id": "a7b19303-00d1-4e23-8595-8aac1cce4ce0", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-1700104384-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "330a1532c0a04bcb8e8712b08883deb1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92cdccfd-4b10-4024-b724-5f22792dd4de", "external-id": "nsx-vlan-transportzone-902", "segmentation_id": 902, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa7172f1f-64", "ovs_interfaceid": "a7172f1f-6473-48fe-bec7-5c84411c8fd2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2164.562230] env[62824]: DEBUG nova.network.neutron [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: da1ad739-b252-4e29-a22a-ab3bcab173ec] Successfully updated port: 734a463a-d647-4c05-b8a9-29a3dd47e376 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2164.616760] env[62824]: DEBUG oslo_vmware.api [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Task: {'id': task-2146473, 'name': PowerOnVM_Task, 'duration_secs': 0.636763} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2164.617223] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 2afd1f18-234c-40b5-9ede-7413ad30dafe] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2164.617583] env[62824]: INFO nova.compute.manager [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 2afd1f18-234c-40b5-9ede-7413ad30dafe] Took 8.05 seconds to spawn the instance on the hypervisor. [ 2164.617924] env[62824]: DEBUG nova.compute.manager [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 2afd1f18-234c-40b5-9ede-7413ad30dafe] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2164.618819] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-516c2cfe-7a6d-45ff-9392-fc49dd73a46e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2164.753636] env[62824]: DEBUG oslo_vmware.api [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52d1f597-c936-ee7d-9776-a10b9e2b39f6, 'name': SearchDatastore_Task, 'duration_secs': 0.012446} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2164.753965] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2164.754237] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] [instance: ab37b1a2-0012-40fb-9341-b613525e89cf] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2164.754471] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2164.754852] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2164.754852] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2164.755081] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a806595a-95a9-4c93-9571-bda13892ab5e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2164.766067] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2164.766277] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2164.767008] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0ccb1d22-37f7-4fc7-a1b3-0bede855a43f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2164.773282] env[62824]: DEBUG oslo_vmware.api [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Waiting for the task: (returnval){ [ 2164.773282] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52f5d465-e622-efff-13a9-7c8beb889114" [ 2164.773282] env[62824]: _type = "Task" [ 2164.773282] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2164.782269] env[62824]: DEBUG oslo_vmware.api [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52f5d465-e622-efff-13a9-7c8beb889114, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2164.989175] env[62824]: DEBUG oslo_vmware.api [None req-3a1b2b23-923e-4532-b132-fcd23342ce08 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Task: {'id': task-2146477, 'name': PowerOffVM_Task, 'duration_secs': 0.382316} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2164.989468] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a1b2b23-923e-4532-b132-fcd23342ce08 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] [instance: bd7f2ed5-fa08-4a48-9755-60b61791ea0b] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2164.989643] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-3a1b2b23-923e-4532-b132-fcd23342ce08 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] [instance: bd7f2ed5-fa08-4a48-9755-60b61791ea0b] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2164.989957] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4ea1c922-3e65-4760-a05b-fb0253dd7ba7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2164.992719] env[62824]: DEBUG nova.compute.utils [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2164.997602] env[62824]: DEBUG nova.compute.manager [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2164.997824] env[62824]: DEBUG nova.network.neutron [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2165.041143] env[62824]: DEBUG nova.policy [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '06cbf37cc90244589bc8b645cb572bfd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c8133feb4bde41e9898b3f7019c24d30', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2165.047031] env[62824]: DEBUG oslo_concurrency.lockutils [req-fbe6c4b9-1808-4136-b74f-341eb6f71526 req-e7c4ea70-ca1b-43e2-94e2-dbddec404d07 service nova] Releasing lock "refresh_cache-ab37b1a2-0012-40fb-9341-b613525e89cf" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2165.068024] env[62824]: DEBUG oslo_concurrency.lockutils [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Acquiring lock "refresh_cache-da1ad739-b252-4e29-a22a-ab3bcab173ec" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2165.068024] env[62824]: DEBUG oslo_concurrency.lockutils [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Acquired lock "refresh_cache-da1ad739-b252-4e29-a22a-ab3bcab173ec" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2165.068024] env[62824]: DEBUG nova.network.neutron [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: da1ad739-b252-4e29-a22a-ab3bcab173ec] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2165.107834] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-3a1b2b23-923e-4532-b132-fcd23342ce08 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] [instance: bd7f2ed5-fa08-4a48-9755-60b61791ea0b] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2165.108084] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-3a1b2b23-923e-4532-b132-fcd23342ce08 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] [instance: bd7f2ed5-fa08-4a48-9755-60b61791ea0b] Deleting contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2165.108281] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-3a1b2b23-923e-4532-b132-fcd23342ce08 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Deleting the datastore file [datastore2] bd7f2ed5-fa08-4a48-9755-60b61791ea0b {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2165.108566] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-35f585e4-2fad-430b-8a19-7b6c80e4e557 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2165.120030] env[62824]: DEBUG oslo_vmware.api [None req-3a1b2b23-923e-4532-b132-fcd23342ce08 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Waiting for the task: (returnval){ [ 2165.120030] env[62824]: value = "task-2146479" [ 2165.120030] env[62824]: _type = "Task" [ 2165.120030] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2165.129738] env[62824]: DEBUG oslo_vmware.api [None req-3a1b2b23-923e-4532-b132-fcd23342ce08 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Task: {'id': task-2146479, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2165.141077] env[62824]: INFO nova.compute.manager [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 2afd1f18-234c-40b5-9ede-7413ad30dafe] Took 38.75 seconds to build instance. [ 2165.293987] env[62824]: DEBUG oslo_vmware.api [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52f5d465-e622-efff-13a9-7c8beb889114, 'name': SearchDatastore_Task, 'duration_secs': 0.010131} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2165.298097] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a1bb9bbd-7151-4017-a340-49bef748eeaf {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2165.304724] env[62824]: DEBUG oslo_vmware.api [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Waiting for the task: (returnval){ [ 2165.304724] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]522dc0be-2255-4279-3f07-ec10b7b6d229" [ 2165.304724] env[62824]: _type = "Task" [ 2165.304724] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2165.315965] env[62824]: DEBUG oslo_vmware.api [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]522dc0be-2255-4279-3f07-ec10b7b6d229, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2165.317797] env[62824]: DEBUG nova.network.neutron [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Successfully created port: 0cefe7bc-5dd6-4f4c-872b-ef1b658a4d60 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2165.464443] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19872975-3cab-47ba-864d-b9121ba6c0e2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2165.472510] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab35c800-0b31-44bf-ba71-b5b9e9e8be72 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2165.505134] env[62824]: DEBUG nova.compute.manager [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2165.508460] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b3902e6-9103-4684-9d02-11ecca7b54de {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2165.518817] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-232190c0-2ad0-421b-8891-4968931cbd8b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2165.536974] env[62824]: DEBUG nova.compute.provider_tree [None req-61bb4e80-84d9-4e19-b9c0-195ff35ecb41 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2165.597502] env[62824]: DEBUG nova.network.neutron [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: da1ad739-b252-4e29-a22a-ab3bcab173ec] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2165.630272] env[62824]: DEBUG oslo_vmware.api [None req-3a1b2b23-923e-4532-b132-fcd23342ce08 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Task: {'id': task-2146479, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.212928} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2165.630444] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-3a1b2b23-923e-4532-b132-fcd23342ce08 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2165.630629] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-3a1b2b23-923e-4532-b132-fcd23342ce08 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] [instance: bd7f2ed5-fa08-4a48-9755-60b61791ea0b] Deleted contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2165.630802] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-3a1b2b23-923e-4532-b132-fcd23342ce08 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] [instance: bd7f2ed5-fa08-4a48-9755-60b61791ea0b] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2165.630970] env[62824]: INFO nova.compute.manager [None req-3a1b2b23-923e-4532-b132-fcd23342ce08 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] [instance: bd7f2ed5-fa08-4a48-9755-60b61791ea0b] Took 1.18 seconds to destroy the instance on the hypervisor. [ 2165.631226] env[62824]: DEBUG oslo.service.loopingcall [None req-3a1b2b23-923e-4532-b132-fcd23342ce08 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2165.631416] env[62824]: DEBUG nova.compute.manager [-] [instance: bd7f2ed5-fa08-4a48-9755-60b61791ea0b] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2165.631523] env[62824]: DEBUG nova.network.neutron [-] [instance: bd7f2ed5-fa08-4a48-9755-60b61791ea0b] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2165.647312] env[62824]: DEBUG oslo_concurrency.lockutils [None req-35195f52-4424-4cd1-bddb-8f535bc441cf tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Lock "2afd1f18-234c-40b5-9ede-7413ad30dafe" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 55.169s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2165.732096] env[62824]: DEBUG nova.network.neutron [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: da1ad739-b252-4e29-a22a-ab3bcab173ec] Updating instance_info_cache with network_info: [{"id": "734a463a-d647-4c05-b8a9-29a3dd47e376", "address": "fa:16:3e:91:c4:ef", "network": {"id": "aeacc588-ae73-49ec-9aa1-68f7d92055be", "bridge": "br-int", "label": "tempest-ImagesTestJSON-314094063-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10ee5fe93d584725ace90e76605a1b2e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1e7a4976-597e-4636-990e-6062b5faadee", "external-id": "nsx-vlan-transportzone-847", "segmentation_id": 847, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap734a463a-d6", "ovs_interfaceid": "734a463a-d647-4c05-b8a9-29a3dd47e376", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2165.816900] env[62824]: DEBUG oslo_vmware.api [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]522dc0be-2255-4279-3f07-ec10b7b6d229, 'name': SearchDatastore_Task, 'duration_secs': 0.011125} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2165.817222] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2165.817527] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] ab37b1a2-0012-40fb-9341-b613525e89cf/ab37b1a2-0012-40fb-9341-b613525e89cf.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2165.817770] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-406a5b61-2bf8-4e7a-9aff-0366dd50155e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2165.825759] env[62824]: DEBUG oslo_vmware.api [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Waiting for the task: (returnval){ [ 2165.825759] env[62824]: value = "task-2146480" [ 2165.825759] env[62824]: _type = "Task" [ 2165.825759] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2165.833870] env[62824]: DEBUG oslo_vmware.api [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Task: {'id': task-2146480, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2166.040391] env[62824]: DEBUG nova.scheduler.client.report [None req-61bb4e80-84d9-4e19-b9c0-195ff35ecb41 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2166.238038] env[62824]: DEBUG oslo_concurrency.lockutils [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Releasing lock "refresh_cache-da1ad739-b252-4e29-a22a-ab3bcab173ec" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2166.238038] env[62824]: DEBUG nova.compute.manager [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: da1ad739-b252-4e29-a22a-ab3bcab173ec] Instance network_info: |[{"id": "734a463a-d647-4c05-b8a9-29a3dd47e376", "address": "fa:16:3e:91:c4:ef", "network": {"id": "aeacc588-ae73-49ec-9aa1-68f7d92055be", "bridge": "br-int", "label": "tempest-ImagesTestJSON-314094063-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10ee5fe93d584725ace90e76605a1b2e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1e7a4976-597e-4636-990e-6062b5faadee", "external-id": "nsx-vlan-transportzone-847", "segmentation_id": 847, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap734a463a-d6", "ovs_interfaceid": "734a463a-d647-4c05-b8a9-29a3dd47e376", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2166.238038] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: da1ad739-b252-4e29-a22a-ab3bcab173ec] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:91:c4:ef', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1e7a4976-597e-4636-990e-6062b5faadee', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '734a463a-d647-4c05-b8a9-29a3dd47e376', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2166.246698] env[62824]: DEBUG oslo.service.loopingcall [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2166.247784] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: da1ad739-b252-4e29-a22a-ab3bcab173ec] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2166.248257] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e8199521-bc8c-4f3a-b817-3e277a634f83 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2166.271671] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2166.271671] env[62824]: value = "task-2146481" [ 2166.271671] env[62824]: _type = "Task" [ 2166.271671] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2166.284970] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146481, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2166.308574] env[62824]: DEBUG nova.compute.manager [req-ad1bfc78-e0fc-4c8c-b4fc-e5b4fe4c604a req-eb864471-875e-4928-8ca0-947d71971bf1 service nova] [instance: bd7f2ed5-fa08-4a48-9755-60b61791ea0b] Received event network-vif-deleted-01afe8c7-58f8-4fa9-97a6-4767b34ccb66 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2166.308700] env[62824]: INFO nova.compute.manager [req-ad1bfc78-e0fc-4c8c-b4fc-e5b4fe4c604a req-eb864471-875e-4928-8ca0-947d71971bf1 service nova] [instance: bd7f2ed5-fa08-4a48-9755-60b61791ea0b] Neutron deleted interface 01afe8c7-58f8-4fa9-97a6-4767b34ccb66; detaching it from the instance and deleting it from the info cache [ 2166.308884] env[62824]: DEBUG nova.network.neutron [req-ad1bfc78-e0fc-4c8c-b4fc-e5b4fe4c604a req-eb864471-875e-4928-8ca0-947d71971bf1 service nova] [instance: bd7f2ed5-fa08-4a48-9755-60b61791ea0b] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2166.336643] env[62824]: DEBUG oslo_vmware.api [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Task: {'id': task-2146480, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.476583} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2166.337180] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] ab37b1a2-0012-40fb-9341-b613525e89cf/ab37b1a2-0012-40fb-9341-b613525e89cf.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2166.337402] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] [instance: ab37b1a2-0012-40fb-9341-b613525e89cf] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2166.337748] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e4c134df-ab2d-4c66-965a-5bb3dad74926 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2166.347353] env[62824]: DEBUG oslo_vmware.api [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Waiting for the task: (returnval){ [ 2166.347353] env[62824]: value = "task-2146482" [ 2166.347353] env[62824]: _type = "Task" [ 2166.347353] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2166.358224] env[62824]: DEBUG oslo_vmware.api [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Task: {'id': task-2146482, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2166.520297] env[62824]: DEBUG nova.compute.manager [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2166.548226] env[62824]: DEBUG oslo_concurrency.lockutils [None req-61bb4e80-84d9-4e19-b9c0-195ff35ecb41 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.060s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2166.550517] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b30d3e55-ba09-4502-8b58-e6c2d9aaba20 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 28.508s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2166.551181] env[62824]: DEBUG nova.objects.instance [None req-b30d3e55-ba09-4502-8b58-e6c2d9aaba20 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Lazy-loading 'resources' on Instance uuid 49818910-61de-4fb8-bbab-d5d61d2f1ada {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2166.553940] env[62824]: DEBUG nova.virt.hardware [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2166.554700] env[62824]: DEBUG nova.virt.hardware [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2166.554700] env[62824]: DEBUG nova.virt.hardware [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2166.554700] env[62824]: DEBUG nova.virt.hardware [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2166.554700] env[62824]: DEBUG nova.virt.hardware [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2166.554851] env[62824]: DEBUG nova.virt.hardware [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2166.554959] env[62824]: DEBUG nova.virt.hardware [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2166.555135] env[62824]: DEBUG nova.virt.hardware [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2166.555419] env[62824]: DEBUG nova.virt.hardware [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2166.555563] env[62824]: DEBUG nova.virt.hardware [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2166.556169] env[62824]: DEBUG nova.virt.hardware [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2166.556169] env[62824]: DEBUG nova.network.neutron [-] [instance: bd7f2ed5-fa08-4a48-9755-60b61791ea0b] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2166.557685] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01eaa725-cc33-4cab-85e2-cd2dd10d5cca {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2166.567759] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c790791c-d1d3-425f-a942-d9023cd537be {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2166.585655] env[62824]: INFO nova.scheduler.client.report [None req-61bb4e80-84d9-4e19-b9c0-195ff35ecb41 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Deleted allocations for instance bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e [ 2166.686042] env[62824]: DEBUG nova.compute.manager [req-c750d96f-2f05-4ef1-aa25-c1c64a1c5858 req-8f336664-a2a2-4132-9e81-1fe333e8e9c7 service nova] [instance: da1ad739-b252-4e29-a22a-ab3bcab173ec] Received event network-changed-734a463a-d647-4c05-b8a9-29a3dd47e376 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2166.687683] env[62824]: DEBUG nova.compute.manager [req-c750d96f-2f05-4ef1-aa25-c1c64a1c5858 req-8f336664-a2a2-4132-9e81-1fe333e8e9c7 service nova] [instance: da1ad739-b252-4e29-a22a-ab3bcab173ec] Refreshing instance network info cache due to event network-changed-734a463a-d647-4c05-b8a9-29a3dd47e376. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2166.688077] env[62824]: DEBUG oslo_concurrency.lockutils [req-c750d96f-2f05-4ef1-aa25-c1c64a1c5858 req-8f336664-a2a2-4132-9e81-1fe333e8e9c7 service nova] Acquiring lock "refresh_cache-da1ad739-b252-4e29-a22a-ab3bcab173ec" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2166.688334] env[62824]: DEBUG oslo_concurrency.lockutils [req-c750d96f-2f05-4ef1-aa25-c1c64a1c5858 req-8f336664-a2a2-4132-9e81-1fe333e8e9c7 service nova] Acquired lock "refresh_cache-da1ad739-b252-4e29-a22a-ab3bcab173ec" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2166.688541] env[62824]: DEBUG nova.network.neutron [req-c750d96f-2f05-4ef1-aa25-c1c64a1c5858 req-8f336664-a2a2-4132-9e81-1fe333e8e9c7 service nova] [instance: da1ad739-b252-4e29-a22a-ab3bcab173ec] Refreshing network info cache for port 734a463a-d647-4c05-b8a9-29a3dd47e376 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2166.782049] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146481, 'name': CreateVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2166.811621] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1e016330-c74d-490a-aab8-ce7b9d5bc055 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2166.825827] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4feeb99-8c5c-4d3b-866b-499eb00215b6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2166.853364] env[62824]: DEBUG nova.network.neutron [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Successfully updated port: 0cefe7bc-5dd6-4f4c-872b-ef1b658a4d60 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2166.872827] env[62824]: DEBUG nova.compute.manager [req-ad1bfc78-e0fc-4c8c-b4fc-e5b4fe4c604a req-eb864471-875e-4928-8ca0-947d71971bf1 service nova] [instance: bd7f2ed5-fa08-4a48-9755-60b61791ea0b] Detach interface failed, port_id=01afe8c7-58f8-4fa9-97a6-4767b34ccb66, reason: Instance bd7f2ed5-fa08-4a48-9755-60b61791ea0b could not be found. {{(pid=62824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11482}} [ 2166.873692] env[62824]: DEBUG oslo_concurrency.lockutils [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Acquiring lock "refresh_cache-da46d1d5-f75f-4ef1-b571-fbebab89c2a6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2166.873835] env[62824]: DEBUG oslo_concurrency.lockutils [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Acquired lock "refresh_cache-da46d1d5-f75f-4ef1-b571-fbebab89c2a6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2166.873982] env[62824]: DEBUG nova.network.neutron [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2166.880660] env[62824]: DEBUG oslo_vmware.api [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Task: {'id': task-2146482, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.131871} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2166.880905] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] [instance: ab37b1a2-0012-40fb-9341-b613525e89cf] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2166.881759] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86594fde-c060-4f66-aba1-7a3933e646ab {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2166.905226] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] [instance: ab37b1a2-0012-40fb-9341-b613525e89cf] Reconfiguring VM instance instance-00000050 to attach disk [datastore1] ab37b1a2-0012-40fb-9341-b613525e89cf/ab37b1a2-0012-40fb-9341-b613525e89cf.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2166.905970] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-06e5764b-fca3-46b3-a243-f27a1eb45552 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2166.927276] env[62824]: DEBUG oslo_vmware.api [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Waiting for the task: (returnval){ [ 2166.927276] env[62824]: value = "task-2146483" [ 2166.927276] env[62824]: _type = "Task" [ 2166.927276] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2166.935695] env[62824]: DEBUG oslo_vmware.api [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Task: {'id': task-2146483, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2167.064366] env[62824]: INFO nova.compute.manager [-] [instance: bd7f2ed5-fa08-4a48-9755-60b61791ea0b] Took 1.43 seconds to deallocate network for instance. [ 2167.095536] env[62824]: DEBUG oslo_concurrency.lockutils [None req-61bb4e80-84d9-4e19-b9c0-195ff35ecb41 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.062s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2167.286445] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146481, 'name': CreateVM_Task, 'duration_secs': 0.600341} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2167.286776] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: da1ad739-b252-4e29-a22a-ab3bcab173ec] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2167.287328] env[62824]: DEBUG oslo_concurrency.lockutils [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/758d9d98-91b8-46a1-9c95-5705c074be1a" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2167.287532] env[62824]: DEBUG oslo_concurrency.lockutils [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Acquired lock "[datastore1] devstack-image-cache_base/758d9d98-91b8-46a1-9c95-5705c074be1a" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2167.288016] env[62824]: DEBUG oslo_concurrency.lockutils [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/758d9d98-91b8-46a1-9c95-5705c074be1a" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2167.288703] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a14ea770-6a71-4bc7-8b48-ead8d82618f5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2167.296550] env[62824]: DEBUG oslo_vmware.api [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the task: (returnval){ [ 2167.296550] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5202fe84-ab8c-0c87-dfb0-392da0c074f4" [ 2167.296550] env[62824]: _type = "Task" [ 2167.296550] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2167.304639] env[62824]: DEBUG oslo_vmware.api [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5202fe84-ab8c-0c87-dfb0-392da0c074f4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2167.307718] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ce4bd0ad-2f42-4d62-b47f-581ab3fbf6be tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Acquiring lock "2afd1f18-234c-40b5-9ede-7413ad30dafe" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2167.307951] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ce4bd0ad-2f42-4d62-b47f-581ab3fbf6be tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Lock "2afd1f18-234c-40b5-9ede-7413ad30dafe" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2167.308775] env[62824]: INFO nova.compute.manager [None req-ce4bd0ad-2f42-4d62-b47f-581ab3fbf6be tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 2afd1f18-234c-40b5-9ede-7413ad30dafe] Rebooting instance [ 2167.426248] env[62824]: DEBUG nova.network.neutron [req-c750d96f-2f05-4ef1-aa25-c1c64a1c5858 req-8f336664-a2a2-4132-9e81-1fe333e8e9c7 service nova] [instance: da1ad739-b252-4e29-a22a-ab3bcab173ec] Updated VIF entry in instance network info cache for port 734a463a-d647-4c05-b8a9-29a3dd47e376. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2167.427272] env[62824]: DEBUG nova.network.neutron [req-c750d96f-2f05-4ef1-aa25-c1c64a1c5858 req-8f336664-a2a2-4132-9e81-1fe333e8e9c7 service nova] [instance: da1ad739-b252-4e29-a22a-ab3bcab173ec] Updating instance_info_cache with network_info: [{"id": "734a463a-d647-4c05-b8a9-29a3dd47e376", "address": "fa:16:3e:91:c4:ef", "network": {"id": "aeacc588-ae73-49ec-9aa1-68f7d92055be", "bridge": "br-int", "label": "tempest-ImagesTestJSON-314094063-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10ee5fe93d584725ace90e76605a1b2e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1e7a4976-597e-4636-990e-6062b5faadee", "external-id": "nsx-vlan-transportzone-847", "segmentation_id": 847, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap734a463a-d6", "ovs_interfaceid": "734a463a-d647-4c05-b8a9-29a3dd47e376", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2167.436032] env[62824]: DEBUG nova.network.neutron [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2167.442651] env[62824]: DEBUG oslo_vmware.api [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Task: {'id': task-2146483, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2167.494268] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c24427c-c0c2-4f38-a80d-02182c585447 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2167.504572] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5ed130a-c522-4201-8705-9ffef92c85d7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2167.543685] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09eb3d11-d4b5-4922-abb6-b0358553fb87 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2167.552805] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bea57c27-d8ee-4ec1-9eed-af86f9cbd392 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2167.577788] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3a1b2b23-923e-4532-b132-fcd23342ce08 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2167.578625] env[62824]: DEBUG nova.compute.provider_tree [None req-b30d3e55-ba09-4502-8b58-e6c2d9aaba20 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2167.704019] env[62824]: DEBUG nova.network.neutron [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Updating instance_info_cache with network_info: [{"id": "0cefe7bc-5dd6-4f4c-872b-ef1b658a4d60", "address": "fa:16:3e:e0:51:b5", "network": {"id": "3210f2eb-2afb-4bce-bb59-f66e85888333", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-216881410-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "c8133feb4bde41e9898b3f7019c24d30", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d2bf584a-b4a3-4e7a-b0b7-eb8a2bc5a11d", "external-id": "nsx-vlan-transportzone-286", "segmentation_id": 286, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0cefe7bc-5d", "ovs_interfaceid": "0cefe7bc-5dd6-4f4c-872b-ef1b658a4d60", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2167.808113] env[62824]: DEBUG oslo_concurrency.lockutils [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Releasing lock "[datastore1] devstack-image-cache_base/758d9d98-91b8-46a1-9c95-5705c074be1a" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2167.808387] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: da1ad739-b252-4e29-a22a-ab3bcab173ec] Processing image 758d9d98-91b8-46a1-9c95-5705c074be1a {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2167.808728] env[62824]: DEBUG oslo_concurrency.lockutils [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/758d9d98-91b8-46a1-9c95-5705c074be1a/758d9d98-91b8-46a1-9c95-5705c074be1a.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2167.808944] env[62824]: DEBUG oslo_concurrency.lockutils [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Acquired lock "[datastore1] devstack-image-cache_base/758d9d98-91b8-46a1-9c95-5705c074be1a/758d9d98-91b8-46a1-9c95-5705c074be1a.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2167.809230] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2167.809608] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3482f5b0-f6ca-462a-8128-1b596b075d3f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2167.822550] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2167.822550] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2167.822550] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c6c9ff12-af60-499e-9eac-712935698c47 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2167.828038] env[62824]: DEBUG oslo_vmware.api [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the task: (returnval){ [ 2167.828038] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5214af57-a146-e23a-e037-15aca6fba46d" [ 2167.828038] env[62824]: _type = "Task" [ 2167.828038] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2167.839081] env[62824]: DEBUG oslo_vmware.api [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5214af57-a146-e23a-e037-15aca6fba46d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2167.840217] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ce4bd0ad-2f42-4d62-b47f-581ab3fbf6be tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Acquiring lock "refresh_cache-2afd1f18-234c-40b5-9ede-7413ad30dafe" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2167.840374] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ce4bd0ad-2f42-4d62-b47f-581ab3fbf6be tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Acquired lock "refresh_cache-2afd1f18-234c-40b5-9ede-7413ad30dafe" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2167.840549] env[62824]: DEBUG nova.network.neutron [None req-ce4bd0ad-2f42-4d62-b47f-581ab3fbf6be tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 2afd1f18-234c-40b5-9ede-7413ad30dafe] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2167.929430] env[62824]: DEBUG oslo_concurrency.lockutils [req-c750d96f-2f05-4ef1-aa25-c1c64a1c5858 req-8f336664-a2a2-4132-9e81-1fe333e8e9c7 service nova] Releasing lock "refresh_cache-da1ad739-b252-4e29-a22a-ab3bcab173ec" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2167.939445] env[62824]: DEBUG oslo_vmware.api [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Task: {'id': task-2146483, 'name': ReconfigVM_Task, 'duration_secs': 0.843145} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2167.939757] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] [instance: ab37b1a2-0012-40fb-9341-b613525e89cf] Reconfigured VM instance instance-00000050 to attach disk [datastore1] ab37b1a2-0012-40fb-9341-b613525e89cf/ab37b1a2-0012-40fb-9341-b613525e89cf.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2167.940477] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4ef97a1f-a626-4a8c-9e94-d1853cece614 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2167.949374] env[62824]: DEBUG oslo_vmware.api [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Waiting for the task: (returnval){ [ 2167.949374] env[62824]: value = "task-2146484" [ 2167.949374] env[62824]: _type = "Task" [ 2167.949374] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2167.959586] env[62824]: DEBUG oslo_vmware.api [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Task: {'id': task-2146484, 'name': Rename_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2168.042966] env[62824]: DEBUG oslo_concurrency.lockutils [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquiring lock "8c2ac679-6915-4149-aa51-99170b0df685" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2168.043190] env[62824]: DEBUG oslo_concurrency.lockutils [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "8c2ac679-6915-4149-aa51-99170b0df685" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2168.081300] env[62824]: DEBUG nova.scheduler.client.report [None req-b30d3e55-ba09-4502-8b58-e6c2d9aaba20 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2168.206841] env[62824]: DEBUG oslo_concurrency.lockutils [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Releasing lock "refresh_cache-da46d1d5-f75f-4ef1-b571-fbebab89c2a6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2168.207210] env[62824]: DEBUG nova.compute.manager [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Instance network_info: |[{"id": "0cefe7bc-5dd6-4f4c-872b-ef1b658a4d60", "address": "fa:16:3e:e0:51:b5", "network": {"id": "3210f2eb-2afb-4bce-bb59-f66e85888333", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-216881410-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "c8133feb4bde41e9898b3f7019c24d30", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d2bf584a-b4a3-4e7a-b0b7-eb8a2bc5a11d", "external-id": "nsx-vlan-transportzone-286", "segmentation_id": 286, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0cefe7bc-5d", "ovs_interfaceid": "0cefe7bc-5dd6-4f4c-872b-ef1b658a4d60", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2168.207751] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e0:51:b5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd2bf584a-b4a3-4e7a-b0b7-eb8a2bc5a11d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0cefe7bc-5dd6-4f4c-872b-ef1b658a4d60', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2168.215598] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Creating folder: Project (c8133feb4bde41e9898b3f7019c24d30). Parent ref: group-v438503. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 2168.215869] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-24755c3b-807a-4133-bda3-b3b6ec1843cd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2168.227720] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Created folder: Project (c8133feb4bde41e9898b3f7019c24d30) in parent group-v438503. [ 2168.227910] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Creating folder: Instances. Parent ref: group-v438708. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 2168.228537] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-862f9d41-1fe0-4716-8d33-773cf5cf6af9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2168.238787] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Created folder: Instances in parent group-v438708. [ 2168.239017] env[62824]: DEBUG oslo.service.loopingcall [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2168.239214] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2168.239429] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5d4a3753-c090-4f7a-a254-7f596051339f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2168.258843] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2168.258843] env[62824]: value = "task-2146487" [ 2168.258843] env[62824]: _type = "Task" [ 2168.258843] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2168.266490] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146487, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2168.337698] env[62824]: DEBUG nova.compute.manager [req-ca04d89a-90ed-454e-bac6-dfc6f03c62fd req-4410380a-6bdb-4ef0-85c1-959af4316f52 service nova] [instance: 2afd1f18-234c-40b5-9ede-7413ad30dafe] Received event network-changed-6d0cd52e-ace8-4459-860b-f270fa266268 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2168.337903] env[62824]: DEBUG nova.compute.manager [req-ca04d89a-90ed-454e-bac6-dfc6f03c62fd req-4410380a-6bdb-4ef0-85c1-959af4316f52 service nova] [instance: 2afd1f18-234c-40b5-9ede-7413ad30dafe] Refreshing instance network info cache due to event network-changed-6d0cd52e-ace8-4459-860b-f270fa266268. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2168.338116] env[62824]: DEBUG oslo_concurrency.lockutils [req-ca04d89a-90ed-454e-bac6-dfc6f03c62fd req-4410380a-6bdb-4ef0-85c1-959af4316f52 service nova] Acquiring lock "refresh_cache-2afd1f18-234c-40b5-9ede-7413ad30dafe" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2168.341993] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: da1ad739-b252-4e29-a22a-ab3bcab173ec] Preparing fetch location {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 2168.342249] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: da1ad739-b252-4e29-a22a-ab3bcab173ec] Fetch image to [datastore1] OSTACK_IMG_cbb54347-fd26-4d25-aa17-fd598a754d99/OSTACK_IMG_cbb54347-fd26-4d25-aa17-fd598a754d99.vmdk {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 2168.342430] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: da1ad739-b252-4e29-a22a-ab3bcab173ec] Downloading stream optimized image 758d9d98-91b8-46a1-9c95-5705c074be1a to [datastore1] OSTACK_IMG_cbb54347-fd26-4d25-aa17-fd598a754d99/OSTACK_IMG_cbb54347-fd26-4d25-aa17-fd598a754d99.vmdk on the data store datastore1 as vApp {{(pid=62824) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 2168.342598] env[62824]: DEBUG nova.virt.vmwareapi.images [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: da1ad739-b252-4e29-a22a-ab3bcab173ec] Downloading image file data 758d9d98-91b8-46a1-9c95-5705c074be1a to the ESX as VM named 'OSTACK_IMG_cbb54347-fd26-4d25-aa17-fd598a754d99' {{(pid=62824) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 2168.426565] env[62824]: DEBUG oslo_vmware.rw_handles [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 2168.426565] env[62824]: value = "resgroup-9" [ 2168.426565] env[62824]: _type = "ResourcePool" [ 2168.426565] env[62824]: }. {{(pid=62824) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 2168.427157] env[62824]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-87930d39-abed-4f68-bf87-0448eaaff2b2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2168.455246] env[62824]: DEBUG oslo_vmware.rw_handles [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Lease: (returnval){ [ 2168.455246] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52b57880-aae0-73d3-f814-415d53c6c43e" [ 2168.455246] env[62824]: _type = "HttpNfcLease" [ 2168.455246] env[62824]: } obtained for vApp import into resource pool (val){ [ 2168.455246] env[62824]: value = "resgroup-9" [ 2168.455246] env[62824]: _type = "ResourcePool" [ 2168.455246] env[62824]: }. {{(pid=62824) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 2168.455833] env[62824]: DEBUG oslo_vmware.api [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the lease: (returnval){ [ 2168.455833] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52b57880-aae0-73d3-f814-415d53c6c43e" [ 2168.455833] env[62824]: _type = "HttpNfcLease" [ 2168.455833] env[62824]: } to be ready. {{(pid=62824) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 2168.467567] env[62824]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 2168.467567] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52b57880-aae0-73d3-f814-415d53c6c43e" [ 2168.467567] env[62824]: _type = "HttpNfcLease" [ 2168.467567] env[62824]: } is initializing. {{(pid=62824) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 2168.470684] env[62824]: DEBUG oslo_vmware.api [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Task: {'id': task-2146484, 'name': Rename_Task, 'duration_secs': 0.146069} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2168.473265] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] [instance: ab37b1a2-0012-40fb-9341-b613525e89cf] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2168.473543] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-530e0b56-9cce-4ef1-9f8c-cf68a340cf30 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2168.481936] env[62824]: DEBUG oslo_vmware.api [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Waiting for the task: (returnval){ [ 2168.481936] env[62824]: value = "task-2146489" [ 2168.481936] env[62824]: _type = "Task" [ 2168.481936] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2168.492957] env[62824]: DEBUG oslo_vmware.api [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Task: {'id': task-2146489, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2168.548028] env[62824]: DEBUG nova.compute.manager [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 8c2ac679-6915-4149-aa51-99170b0df685] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2168.587907] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b30d3e55-ba09-4502-8b58-e6c2d9aaba20 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.037s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2168.592326] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d39ad21e-d65d-4fc4-9829-53a13e0b4169 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 28.241s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2168.592326] env[62824]: DEBUG nova.objects.instance [None req-d39ad21e-d65d-4fc4-9829-53a13e0b4169 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Lazy-loading 'resources' on Instance uuid 59b5b883-4188-471c-8862-444f3ce08cb0 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2168.614264] env[62824]: INFO nova.scheduler.client.report [None req-b30d3e55-ba09-4502-8b58-e6c2d9aaba20 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Deleted allocations for instance 49818910-61de-4fb8-bbab-d5d61d2f1ada [ 2168.666844] env[62824]: DEBUG nova.network.neutron [None req-ce4bd0ad-2f42-4d62-b47f-581ab3fbf6be tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 2afd1f18-234c-40b5-9ede-7413ad30dafe] Updating instance_info_cache with network_info: [{"id": "6d0cd52e-ace8-4459-860b-f270fa266268", "address": "fa:16:3e:5b:95:b5", "network": {"id": "60037885-38d9-4586-9fd7-4907321e9c7e", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-370573912-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2e59203d676d48c8bd8402af4eb58f65", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f856fca-9fb5-41ea-a057-ac4193bd323d", "external-id": "nsx-vlan-transportzone-148", "segmentation_id": 148, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6d0cd52e-ac", "ovs_interfaceid": "6d0cd52e-ace8-4459-860b-f270fa266268", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2168.723883] env[62824]: DEBUG nova.compute.manager [req-27432275-6edd-44d7-8ee0-861ecaae9177 req-a1de06ac-ca1c-42de-a962-b17148da28f0 service nova] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Received event network-vif-plugged-0cefe7bc-5dd6-4f4c-872b-ef1b658a4d60 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2168.723883] env[62824]: DEBUG oslo_concurrency.lockutils [req-27432275-6edd-44d7-8ee0-861ecaae9177 req-a1de06ac-ca1c-42de-a962-b17148da28f0 service nova] Acquiring lock "da46d1d5-f75f-4ef1-b571-fbebab89c2a6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2168.724254] env[62824]: DEBUG oslo_concurrency.lockutils [req-27432275-6edd-44d7-8ee0-861ecaae9177 req-a1de06ac-ca1c-42de-a962-b17148da28f0 service nova] Lock "da46d1d5-f75f-4ef1-b571-fbebab89c2a6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2168.724599] env[62824]: DEBUG oslo_concurrency.lockutils [req-27432275-6edd-44d7-8ee0-861ecaae9177 req-a1de06ac-ca1c-42de-a962-b17148da28f0 service nova] Lock "da46d1d5-f75f-4ef1-b571-fbebab89c2a6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2168.724599] env[62824]: DEBUG nova.compute.manager [req-27432275-6edd-44d7-8ee0-861ecaae9177 req-a1de06ac-ca1c-42de-a962-b17148da28f0 service nova] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] No waiting events found dispatching network-vif-plugged-0cefe7bc-5dd6-4f4c-872b-ef1b658a4d60 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2168.724716] env[62824]: WARNING nova.compute.manager [req-27432275-6edd-44d7-8ee0-861ecaae9177 req-a1de06ac-ca1c-42de-a962-b17148da28f0 service nova] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Received unexpected event network-vif-plugged-0cefe7bc-5dd6-4f4c-872b-ef1b658a4d60 for instance with vm_state building and task_state spawning. [ 2168.724835] env[62824]: DEBUG nova.compute.manager [req-27432275-6edd-44d7-8ee0-861ecaae9177 req-a1de06ac-ca1c-42de-a962-b17148da28f0 service nova] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Received event network-changed-0cefe7bc-5dd6-4f4c-872b-ef1b658a4d60 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2168.725011] env[62824]: DEBUG nova.compute.manager [req-27432275-6edd-44d7-8ee0-861ecaae9177 req-a1de06ac-ca1c-42de-a962-b17148da28f0 service nova] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Refreshing instance network info cache due to event network-changed-0cefe7bc-5dd6-4f4c-872b-ef1b658a4d60. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2168.725212] env[62824]: DEBUG oslo_concurrency.lockutils [req-27432275-6edd-44d7-8ee0-861ecaae9177 req-a1de06ac-ca1c-42de-a962-b17148da28f0 service nova] Acquiring lock "refresh_cache-da46d1d5-f75f-4ef1-b571-fbebab89c2a6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2168.725384] env[62824]: DEBUG oslo_concurrency.lockutils [req-27432275-6edd-44d7-8ee0-861ecaae9177 req-a1de06ac-ca1c-42de-a962-b17148da28f0 service nova] Acquired lock "refresh_cache-da46d1d5-f75f-4ef1-b571-fbebab89c2a6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2168.725499] env[62824]: DEBUG nova.network.neutron [req-27432275-6edd-44d7-8ee0-861ecaae9177 req-a1de06ac-ca1c-42de-a962-b17148da28f0 service nova] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Refreshing network info cache for port 0cefe7bc-5dd6-4f4c-872b-ef1b658a4d60 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2168.770029] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146487, 'name': CreateVM_Task, 'duration_secs': 0.41575} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2168.770029] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2168.770747] env[62824]: DEBUG oslo_concurrency.lockutils [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2168.770911] env[62824]: DEBUG oslo_concurrency.lockutils [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2168.771351] env[62824]: DEBUG oslo_concurrency.lockutils [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2168.771607] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6bef1803-43a9-44bd-9930-398098715a6b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2168.777047] env[62824]: DEBUG oslo_vmware.api [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Waiting for the task: (returnval){ [ 2168.777047] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52947a2c-0fe4-72cb-6623-852b4deb05d7" [ 2168.777047] env[62824]: _type = "Task" [ 2168.777047] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2168.786160] env[62824]: DEBUG oslo_vmware.api [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52947a2c-0fe4-72cb-6623-852b4deb05d7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2168.964711] env[62824]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 2168.964711] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52b57880-aae0-73d3-f814-415d53c6c43e" [ 2168.964711] env[62824]: _type = "HttpNfcLease" [ 2168.964711] env[62824]: } is initializing. {{(pid=62824) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 2168.994699] env[62824]: DEBUG oslo_vmware.api [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Task: {'id': task-2146489, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2169.068157] env[62824]: DEBUG oslo_concurrency.lockutils [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2169.128351] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b30d3e55-ba09-4502-8b58-e6c2d9aaba20 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Lock "49818910-61de-4fb8-bbab-d5d61d2f1ada" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.578s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2169.170170] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ce4bd0ad-2f42-4d62-b47f-581ab3fbf6be tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Releasing lock "refresh_cache-2afd1f18-234c-40b5-9ede-7413ad30dafe" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2169.172137] env[62824]: DEBUG oslo_concurrency.lockutils [req-ca04d89a-90ed-454e-bac6-dfc6f03c62fd req-4410380a-6bdb-4ef0-85c1-959af4316f52 service nova] Acquired lock "refresh_cache-2afd1f18-234c-40b5-9ede-7413ad30dafe" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2169.172137] env[62824]: DEBUG nova.network.neutron [req-ca04d89a-90ed-454e-bac6-dfc6f03c62fd req-4410380a-6bdb-4ef0-85c1-959af4316f52 service nova] [instance: 2afd1f18-234c-40b5-9ede-7413ad30dafe] Refreshing network info cache for port 6d0cd52e-ace8-4459-860b-f270fa266268 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2169.291016] env[62824]: DEBUG oslo_vmware.api [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52947a2c-0fe4-72cb-6623-852b4deb05d7, 'name': SearchDatastore_Task, 'duration_secs': 0.021539} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2169.291509] env[62824]: DEBUG oslo_concurrency.lockutils [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2169.291789] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2169.292238] env[62824]: DEBUG oslo_concurrency.lockutils [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2169.292393] env[62824]: DEBUG oslo_concurrency.lockutils [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2169.292577] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2169.295158] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f3442610-6977-4411-b6fe-66363e090d7f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2169.306313] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2169.306505] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2169.309466] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-68c12b8c-e0a4-445c-97b7-6a8808003810 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2169.316398] env[62824]: DEBUG oslo_vmware.api [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Waiting for the task: (returnval){ [ 2169.316398] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52b8574a-8895-6ddb-03db-1ef0c9392b21" [ 2169.316398] env[62824]: _type = "Task" [ 2169.316398] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2169.325667] env[62824]: DEBUG oslo_vmware.api [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52b8574a-8895-6ddb-03db-1ef0c9392b21, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2169.471653] env[62824]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 2169.471653] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52b57880-aae0-73d3-f814-415d53c6c43e" [ 2169.471653] env[62824]: _type = "HttpNfcLease" [ 2169.471653] env[62824]: } is ready. {{(pid=62824) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 2169.471949] env[62824]: DEBUG oslo_vmware.rw_handles [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 2169.471949] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52b57880-aae0-73d3-f814-415d53c6c43e" [ 2169.471949] env[62824]: _type = "HttpNfcLease" [ 2169.471949] env[62824]: }. {{(pid=62824) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 2169.472706] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61d8f803-b1aa-4abf-baab-a28002723c11 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2169.494269] env[62824]: DEBUG oslo_vmware.rw_handles [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/522df4dd-06d7-a640-2441-f65776f7ff22/disk-0.vmdk from lease info. {{(pid=62824) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 2169.494476] env[62824]: DEBUG oslo_vmware.rw_handles [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Creating HTTP connection to write to file with size = 21334016 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/522df4dd-06d7-a640-2441-f65776f7ff22/disk-0.vmdk. {{(pid=62824) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 2169.562924] env[62824]: DEBUG nova.network.neutron [req-27432275-6edd-44d7-8ee0-861ecaae9177 req-a1de06ac-ca1c-42de-a962-b17148da28f0 service nova] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Updated VIF entry in instance network info cache for port 0cefe7bc-5dd6-4f4c-872b-ef1b658a4d60. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2169.563429] env[62824]: DEBUG nova.network.neutron [req-27432275-6edd-44d7-8ee0-861ecaae9177 req-a1de06ac-ca1c-42de-a962-b17148da28f0 service nova] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Updating instance_info_cache with network_info: [{"id": "0cefe7bc-5dd6-4f4c-872b-ef1b658a4d60", "address": "fa:16:3e:e0:51:b5", "network": {"id": "3210f2eb-2afb-4bce-bb59-f66e85888333", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-216881410-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "c8133feb4bde41e9898b3f7019c24d30", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d2bf584a-b4a3-4e7a-b0b7-eb8a2bc5a11d", "external-id": "nsx-vlan-transportzone-286", "segmentation_id": 286, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0cefe7bc-5d", "ovs_interfaceid": "0cefe7bc-5dd6-4f4c-872b-ef1b658a4d60", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2169.568185] env[62824]: DEBUG oslo_vmware.api [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Task: {'id': task-2146489, 'name': PowerOnVM_Task, 'duration_secs': 0.541734} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2169.570554] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] [instance: ab37b1a2-0012-40fb-9341-b613525e89cf] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2169.570817] env[62824]: INFO nova.compute.manager [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] [instance: ab37b1a2-0012-40fb-9341-b613525e89cf] Took 8.18 seconds to spawn the instance on the hypervisor. [ 2169.571112] env[62824]: DEBUG nova.compute.manager [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] [instance: ab37b1a2-0012-40fb-9341-b613525e89cf] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2169.577082] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8db8d0f4-008c-4ecb-9e56-c7ea6f0b76fa {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2169.584658] env[62824]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-316f5fff-d728-46cd-b11a-c49f44420771 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2169.599342] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a013427d-881a-459c-8724-2d9c935c19fd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2169.608150] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1abe1bd-988a-40b1-9dfe-78d701aff601 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2169.641899] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-893e86ab-4967-46f8-81b2-777c1a049204 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2169.654857] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b47bb139-f8c5-4162-9f56-51e0c35b826f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2169.674169] env[62824]: DEBUG nova.compute.provider_tree [None req-d39ad21e-d65d-4fc4-9829-53a13e0b4169 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2169.677816] env[62824]: DEBUG nova.compute.manager [None req-ce4bd0ad-2f42-4d62-b47f-581ab3fbf6be tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 2afd1f18-234c-40b5-9ede-7413ad30dafe] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2169.678666] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aec578d5-a13a-47ae-baa7-3a31e7540860 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2169.827103] env[62824]: DEBUG oslo_vmware.api [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52b8574a-8895-6ddb-03db-1ef0c9392b21, 'name': SearchDatastore_Task, 'duration_secs': 0.010491} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2169.827859] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1dc3fbbd-52ee-4649-956c-7060db405b72 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2169.833694] env[62824]: DEBUG oslo_vmware.api [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Waiting for the task: (returnval){ [ 2169.833694] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52ea4f56-e975-c9fa-9a7e-3749f42429c2" [ 2169.833694] env[62824]: _type = "Task" [ 2169.833694] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2169.846271] env[62824]: DEBUG oslo_vmware.api [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52ea4f56-e975-c9fa-9a7e-3749f42429c2, 'name': SearchDatastore_Task, 'duration_secs': 0.010117} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2169.846536] env[62824]: DEBUG oslo_concurrency.lockutils [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2169.846806] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] da46d1d5-f75f-4ef1-b571-fbebab89c2a6/da46d1d5-f75f-4ef1-b571-fbebab89c2a6.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2169.847077] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b304a4e2-ad17-4832-aa27-60e408a6fa57 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2169.854772] env[62824]: DEBUG oslo_vmware.api [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Waiting for the task: (returnval){ [ 2169.854772] env[62824]: value = "task-2146490" [ 2169.854772] env[62824]: _type = "Task" [ 2169.854772] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2169.864194] env[62824]: DEBUG oslo_vmware.api [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Task: {'id': task-2146490, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2169.943038] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Acquiring lock "0c9fb5cc-de69-46d8-9962-98e0a84e33c3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2169.943318] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Lock "0c9fb5cc-de69-46d8-9962-98e0a84e33c3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2170.070019] env[62824]: DEBUG oslo_concurrency.lockutils [req-27432275-6edd-44d7-8ee0-861ecaae9177 req-a1de06ac-ca1c-42de-a962-b17148da28f0 service nova] Releasing lock "refresh_cache-da46d1d5-f75f-4ef1-b571-fbebab89c2a6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2170.108313] env[62824]: INFO nova.compute.manager [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] [instance: ab37b1a2-0012-40fb-9341-b613525e89cf] Took 37.58 seconds to build instance. [ 2170.179714] env[62824]: DEBUG nova.scheduler.client.report [None req-d39ad21e-d65d-4fc4-9829-53a13e0b4169 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2170.350499] env[62824]: DEBUG nova.network.neutron [req-ca04d89a-90ed-454e-bac6-dfc6f03c62fd req-4410380a-6bdb-4ef0-85c1-959af4316f52 service nova] [instance: 2afd1f18-234c-40b5-9ede-7413ad30dafe] Updated VIF entry in instance network info cache for port 6d0cd52e-ace8-4459-860b-f270fa266268. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2170.350982] env[62824]: DEBUG nova.network.neutron [req-ca04d89a-90ed-454e-bac6-dfc6f03c62fd req-4410380a-6bdb-4ef0-85c1-959af4316f52 service nova] [instance: 2afd1f18-234c-40b5-9ede-7413ad30dafe] Updating instance_info_cache with network_info: [{"id": "6d0cd52e-ace8-4459-860b-f270fa266268", "address": "fa:16:3e:5b:95:b5", "network": {"id": "60037885-38d9-4586-9fd7-4907321e9c7e", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-370573912-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2e59203d676d48c8bd8402af4eb58f65", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f856fca-9fb5-41ea-a057-ac4193bd323d", "external-id": "nsx-vlan-transportzone-148", "segmentation_id": 148, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6d0cd52e-ac", "ovs_interfaceid": "6d0cd52e-ace8-4459-860b-f270fa266268", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2170.366988] env[62824]: DEBUG oslo_vmware.api [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Task: {'id': task-2146490, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.504589} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2170.367285] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] da46d1d5-f75f-4ef1-b571-fbebab89c2a6/da46d1d5-f75f-4ef1-b571-fbebab89c2a6.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2170.367663] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2170.367768] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-741cb47e-89e2-4fc7-8a04-96585e508989 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2170.376197] env[62824]: DEBUG oslo_vmware.api [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Waiting for the task: (returnval){ [ 2170.376197] env[62824]: value = "task-2146491" [ 2170.376197] env[62824]: _type = "Task" [ 2170.376197] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2170.389415] env[62824]: DEBUG oslo_vmware.api [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Task: {'id': task-2146491, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2170.445616] env[62824]: DEBUG nova.compute.manager [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 0c9fb5cc-de69-46d8-9962-98e0a84e33c3] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2170.539588] env[62824]: DEBUG oslo_vmware.rw_handles [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Completed reading data from the image iterator. {{(pid=62824) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 2170.539845] env[62824]: DEBUG oslo_vmware.rw_handles [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/522df4dd-06d7-a640-2441-f65776f7ff22/disk-0.vmdk. {{(pid=62824) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 2170.540872] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c972fdb-8dd8-4d8d-8da0-9fb45efe93c5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2170.549035] env[62824]: DEBUG oslo_vmware.rw_handles [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/522df4dd-06d7-a640-2441-f65776f7ff22/disk-0.vmdk is in state: ready. {{(pid=62824) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 2170.549035] env[62824]: DEBUG oslo_vmware.rw_handles [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/522df4dd-06d7-a640-2441-f65776f7ff22/disk-0.vmdk. {{(pid=62824) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 2170.549241] env[62824]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-62fa2096-0cc6-4e99-bd1c-1e7fd65ba9c2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2170.614110] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c8484003-23d8-4d14-9270-84956df28644 tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Lock "ab37b1a2-0012-40fb-9341-b613525e89cf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 59.496s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2170.683641] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d39ad21e-d65d-4fc4-9829-53a13e0b4169 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.093s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2170.686136] env[62824]: DEBUG oslo_concurrency.lockutils [None req-123477fc-8c3e-4f2a-83de-cced4f011020 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 29.871s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2170.698217] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b583fa3b-f23b-4556-b380-598db70d5c5b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2170.705851] env[62824]: INFO nova.scheduler.client.report [None req-d39ad21e-d65d-4fc4-9829-53a13e0b4169 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Deleted allocations for instance 59b5b883-4188-471c-8862-444f3ce08cb0 [ 2170.710252] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-ce4bd0ad-2f42-4d62-b47f-581ab3fbf6be tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 2afd1f18-234c-40b5-9ede-7413ad30dafe] Doing hard reboot of VM {{(pid=62824) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 2170.711571] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-0f948092-e2b3-4deb-af04-9da36ce81867 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2170.722615] env[62824]: DEBUG oslo_vmware.api [None req-ce4bd0ad-2f42-4d62-b47f-581ab3fbf6be tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Waiting for the task: (returnval){ [ 2170.722615] env[62824]: value = "task-2146492" [ 2170.722615] env[62824]: _type = "Task" [ 2170.722615] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2170.735155] env[62824]: DEBUG oslo_vmware.api [None req-ce4bd0ad-2f42-4d62-b47f-581ab3fbf6be tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Task: {'id': task-2146492, 'name': ResetVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2170.854316] env[62824]: DEBUG oslo_concurrency.lockutils [req-ca04d89a-90ed-454e-bac6-dfc6f03c62fd req-4410380a-6bdb-4ef0-85c1-959af4316f52 service nova] Releasing lock "refresh_cache-2afd1f18-234c-40b5-9ede-7413ad30dafe" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2170.888934] env[62824]: DEBUG oslo_vmware.api [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Task: {'id': task-2146491, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073936} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2170.889246] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2170.890078] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a48e42fa-808b-4621-b273-b5a6fe1d4347 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2170.916112] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Reconfiguring VM instance instance-00000052 to attach disk [datastore1] da46d1d5-f75f-4ef1-b571-fbebab89c2a6/da46d1d5-f75f-4ef1-b571-fbebab89c2a6.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2170.916112] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f33918ff-46dd-4eb5-ae5a-c91a1c83cddc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2170.937087] env[62824]: DEBUG oslo_vmware.api [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Waiting for the task: (returnval){ [ 2170.937087] env[62824]: value = "task-2146493" [ 2170.937087] env[62824]: _type = "Task" [ 2170.937087] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2170.946162] env[62824]: DEBUG oslo_vmware.api [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Task: {'id': task-2146493, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2170.968727] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2171.036617] env[62824]: DEBUG oslo_vmware.rw_handles [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/522df4dd-06d7-a640-2441-f65776f7ff22/disk-0.vmdk. {{(pid=62824) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 2171.036852] env[62824]: INFO nova.virt.vmwareapi.images [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: da1ad739-b252-4e29-a22a-ab3bcab173ec] Downloaded image file data 758d9d98-91b8-46a1-9c95-5705c074be1a [ 2171.037813] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a41b98f6-33c4-4cd1-8de6-fc3cabd284e9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2171.054078] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f9a991e9-d35a-498e-9240-a2a03becf606 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2171.092132] env[62824]: INFO nova.virt.vmwareapi.images [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: da1ad739-b252-4e29-a22a-ab3bcab173ec] The imported VM was unregistered [ 2171.094751] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: da1ad739-b252-4e29-a22a-ab3bcab173ec] Caching image {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 2171.095029] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Creating directory with path [datastore1] devstack-image-cache_base/758d9d98-91b8-46a1-9c95-5705c074be1a {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2171.095335] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-747f43cc-8ad0-4176-8e3c-0d2758dc22f0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2171.122703] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Created directory with path [datastore1] devstack-image-cache_base/758d9d98-91b8-46a1-9c95-5705c074be1a {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2171.123029] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_cbb54347-fd26-4d25-aa17-fd598a754d99/OSTACK_IMG_cbb54347-fd26-4d25-aa17-fd598a754d99.vmdk to [datastore1] devstack-image-cache_base/758d9d98-91b8-46a1-9c95-5705c074be1a/758d9d98-91b8-46a1-9c95-5705c074be1a.vmdk. {{(pid=62824) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 2171.123369] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-2f3e7dbd-173d-43fa-8781-984b9fcb30a1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2171.131660] env[62824]: DEBUG oslo_vmware.api [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the task: (returnval){ [ 2171.131660] env[62824]: value = "task-2146495" [ 2171.131660] env[62824]: _type = "Task" [ 2171.131660] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2171.140333] env[62824]: DEBUG oslo_vmware.api [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146495, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2171.219528] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d39ad21e-d65d-4fc4-9829-53a13e0b4169 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Lock "59b5b883-4188-471c-8862-444f3ce08cb0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.521s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2171.235466] env[62824]: DEBUG oslo_vmware.api [None req-ce4bd0ad-2f42-4d62-b47f-581ab3fbf6be tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Task: {'id': task-2146492, 'name': ResetVM_Task, 'duration_secs': 0.143065} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2171.236426] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-ce4bd0ad-2f42-4d62-b47f-581ab3fbf6be tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 2afd1f18-234c-40b5-9ede-7413ad30dafe] Did hard reboot of VM {{(pid=62824) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 2171.236426] env[62824]: DEBUG nova.compute.manager [None req-ce4bd0ad-2f42-4d62-b47f-581ab3fbf6be tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 2afd1f18-234c-40b5-9ede-7413ad30dafe] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2171.237027] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9acf1711-fd80-4a61-bf12-121f1ce2cf9f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2171.451508] env[62824]: DEBUG oslo_vmware.api [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Task: {'id': task-2146493, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2171.630392] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f86c493-a241-46fd-9c29-45507ed0852b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2171.645668] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b548a535-306d-4422-a50e-8d746fb87093 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2171.649487] env[62824]: DEBUG oslo_vmware.api [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146495, 'name': MoveVirtualDisk_Task} progress is 15%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2171.678311] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-920f19c3-e952-4cba-a059-ec4c84571da0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2171.688798] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af30b45a-b412-4a37-b8bf-c23920b6ac90 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2171.708462] env[62824]: DEBUG nova.compute.provider_tree [None req-123477fc-8c3e-4f2a-83de-cced4f011020 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2171.753170] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ce4bd0ad-2f42-4d62-b47f-581ab3fbf6be tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Lock "2afd1f18-234c-40b5-9ede-7413ad30dafe" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 4.445s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2171.949543] env[62824]: DEBUG oslo_vmware.api [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Task: {'id': task-2146493, 'name': ReconfigVM_Task, 'duration_secs': 0.862335} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2171.949842] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Reconfigured VM instance instance-00000052 to attach disk [datastore1] da46d1d5-f75f-4ef1-b571-fbebab89c2a6/da46d1d5-f75f-4ef1-b571-fbebab89c2a6.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2171.950525] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6eb6460a-27f8-4ecb-a697-77c85edbf061 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2171.958888] env[62824]: DEBUG oslo_vmware.api [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Waiting for the task: (returnval){ [ 2171.958888] env[62824]: value = "task-2146496" [ 2171.958888] env[62824]: _type = "Task" [ 2171.958888] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2171.971132] env[62824]: DEBUG oslo_vmware.api [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Task: {'id': task-2146496, 'name': Rename_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2172.143776] env[62824]: DEBUG oslo_vmware.api [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146495, 'name': MoveVirtualDisk_Task} progress is 35%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2172.197499] env[62824]: DEBUG oslo_concurrency.lockutils [None req-06ef8715-0d4a-45b8-bcde-6870d18da01a tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Acquiring lock "ab37b1a2-0012-40fb-9341-b613525e89cf" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2172.197865] env[62824]: DEBUG oslo_concurrency.lockutils [None req-06ef8715-0d4a-45b8-bcde-6870d18da01a tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Lock "ab37b1a2-0012-40fb-9341-b613525e89cf" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2172.198104] env[62824]: DEBUG oslo_concurrency.lockutils [None req-06ef8715-0d4a-45b8-bcde-6870d18da01a tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Acquiring lock "ab37b1a2-0012-40fb-9341-b613525e89cf-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2172.198296] env[62824]: DEBUG oslo_concurrency.lockutils [None req-06ef8715-0d4a-45b8-bcde-6870d18da01a tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Lock "ab37b1a2-0012-40fb-9341-b613525e89cf-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2172.198479] env[62824]: DEBUG oslo_concurrency.lockutils [None req-06ef8715-0d4a-45b8-bcde-6870d18da01a tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Lock "ab37b1a2-0012-40fb-9341-b613525e89cf-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2172.201132] env[62824]: INFO nova.compute.manager [None req-06ef8715-0d4a-45b8-bcde-6870d18da01a tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] [instance: ab37b1a2-0012-40fb-9341-b613525e89cf] Terminating instance [ 2172.214506] env[62824]: DEBUG nova.scheduler.client.report [None req-123477fc-8c3e-4f2a-83de-cced4f011020 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2172.470848] env[62824]: DEBUG oslo_vmware.api [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Task: {'id': task-2146496, 'name': Rename_Task, 'duration_secs': 0.43968} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2172.472021] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2172.472021] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2846446d-4919-48de-9814-c3d7fdd4206d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2172.480489] env[62824]: DEBUG oslo_vmware.api [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Waiting for the task: (returnval){ [ 2172.480489] env[62824]: value = "task-2146497" [ 2172.480489] env[62824]: _type = "Task" [ 2172.480489] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2172.492166] env[62824]: DEBUG oslo_vmware.api [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Task: {'id': task-2146497, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2172.633159] env[62824]: DEBUG nova.compute.manager [req-3b127b39-2aad-4784-8766-bf49f228fff7 req-22b813da-6771-47c8-aacf-19909eed57d7 service nova] [instance: 2afd1f18-234c-40b5-9ede-7413ad30dafe] Received event network-changed-6d0cd52e-ace8-4459-860b-f270fa266268 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2172.633159] env[62824]: DEBUG nova.compute.manager [req-3b127b39-2aad-4784-8766-bf49f228fff7 req-22b813da-6771-47c8-aacf-19909eed57d7 service nova] [instance: 2afd1f18-234c-40b5-9ede-7413ad30dafe] Refreshing instance network info cache due to event network-changed-6d0cd52e-ace8-4459-860b-f270fa266268. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2172.633472] env[62824]: DEBUG oslo_concurrency.lockutils [req-3b127b39-2aad-4784-8766-bf49f228fff7 req-22b813da-6771-47c8-aacf-19909eed57d7 service nova] Acquiring lock "refresh_cache-2afd1f18-234c-40b5-9ede-7413ad30dafe" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2172.633472] env[62824]: DEBUG oslo_concurrency.lockutils [req-3b127b39-2aad-4784-8766-bf49f228fff7 req-22b813da-6771-47c8-aacf-19909eed57d7 service nova] Acquired lock "refresh_cache-2afd1f18-234c-40b5-9ede-7413ad30dafe" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2172.633535] env[62824]: DEBUG nova.network.neutron [req-3b127b39-2aad-4784-8766-bf49f228fff7 req-22b813da-6771-47c8-aacf-19909eed57d7 service nova] [instance: 2afd1f18-234c-40b5-9ede-7413ad30dafe] Refreshing network info cache for port 6d0cd52e-ace8-4459-860b-f270fa266268 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2172.648080] env[62824]: DEBUG oslo_vmware.api [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146495, 'name': MoveVirtualDisk_Task} progress is 57%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2172.710018] env[62824]: DEBUG nova.compute.manager [None req-06ef8715-0d4a-45b8-bcde-6870d18da01a tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] [instance: ab37b1a2-0012-40fb-9341-b613525e89cf] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2172.710018] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-06ef8715-0d4a-45b8-bcde-6870d18da01a tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] [instance: ab37b1a2-0012-40fb-9341-b613525e89cf] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2172.711160] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75eda31e-f032-4fd2-adc2-6f993ac1a519 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2172.720041] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-06ef8715-0d4a-45b8-bcde-6870d18da01a tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] [instance: ab37b1a2-0012-40fb-9341-b613525e89cf] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2172.720941] env[62824]: DEBUG oslo_concurrency.lockutils [None req-123477fc-8c3e-4f2a-83de-cced4f011020 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.035s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2172.721696] env[62824]: ERROR nova.compute.manager [None req-123477fc-8c3e-4f2a-83de-cced4f011020 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Instance failed to spawn: nova.exception.InvalidInput: Invalid input received: vif type binding_failed not supported [ 2172.721696] env[62824]: ERROR nova.compute.manager [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Traceback (most recent call last): [ 2172.721696] env[62824]: ERROR nova.compute.manager [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] File "/opt/stack/nova/nova/compute/manager.py", line 7690, in _unshelve_instance [ 2172.721696] env[62824]: ERROR nova.compute.manager [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] self.driver.spawn(context, instance, image_meta, [ 2172.721696] env[62824]: ERROR nova.compute.manager [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 2172.721696] env[62824]: ERROR nova.compute.manager [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] self._vmops.spawn(context, instance, image_meta, injected_files, [ 2172.721696] env[62824]: ERROR nova.compute.manager [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 2172.721696] env[62824]: ERROR nova.compute.manager [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] vm_ref = self.build_virtual_machine(instance, [ 2172.721696] env[62824]: ERROR nova.compute.manager [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 2172.721696] env[62824]: ERROR nova.compute.manager [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] vif_infos = vmwarevif.get_vif_info(self._session, [ 2172.721696] env[62824]: ERROR nova.compute.manager [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 120, in get_vif_info [ 2172.721696] env[62824]: ERROR nova.compute.manager [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] vif_infos.append(get_vif_dict(session, cluster, vif_model, vif)) [ 2172.721696] env[62824]: ERROR nova.compute.manager [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 107, in get_vif_dict [ 2172.721696] env[62824]: ERROR nova.compute.manager [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] ref = get_network_ref(session, cluster, vif) [ 2172.721696] env[62824]: ERROR nova.compute.manager [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 100, in get_network_ref [ 2172.721696] env[62824]: ERROR nova.compute.manager [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] raise exception.InvalidInput(reason=reason) [ 2172.721696] env[62824]: ERROR nova.compute.manager [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] nova.exception.InvalidInput: Invalid input received: vif type binding_failed not supported [ 2172.721696] env[62824]: ERROR nova.compute.manager [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] [ 2172.723483] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ad4fb1ce-418c-4e89-8d39-7f866c5258d8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2172.725717] env[62824]: DEBUG oslo_concurrency.lockutils [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.940s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2172.727401] env[62824]: INFO nova.compute.claims [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 58243465-ba24-4b37-a640-04ab24963230] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2172.737425] env[62824]: DEBUG oslo_vmware.api [None req-06ef8715-0d4a-45b8-bcde-6870d18da01a tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Waiting for the task: (returnval){ [ 2172.737425] env[62824]: value = "task-2146498" [ 2172.737425] env[62824]: _type = "Task" [ 2172.737425] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2172.751831] env[62824]: DEBUG oslo_vmware.api [None req-06ef8715-0d4a-45b8-bcde-6870d18da01a tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Task: {'id': task-2146498, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2172.757457] env[62824]: INFO nova.scheduler.client.report [None req-123477fc-8c3e-4f2a-83de-cced4f011020 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Deleted allocations for instance b8cc8cd3-ea03-40bf-b867-7ad193365552 [ 2172.991745] env[62824]: DEBUG oslo_vmware.api [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Task: {'id': task-2146497, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2173.147723] env[62824]: DEBUG oslo_vmware.api [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146495, 'name': MoveVirtualDisk_Task} progress is 77%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2173.254545] env[62824]: DEBUG oslo_vmware.api [None req-06ef8715-0d4a-45b8-bcde-6870d18da01a tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Task: {'id': task-2146498, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2173.287481] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f32f5e12-bb47-427a-aceb-0383bc172546 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Acquiring lock "2afd1f18-234c-40b5-9ede-7413ad30dafe" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2173.287790] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f32f5e12-bb47-427a-aceb-0383bc172546 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Lock "2afd1f18-234c-40b5-9ede-7413ad30dafe" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2173.288011] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f32f5e12-bb47-427a-aceb-0383bc172546 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Acquiring lock "2afd1f18-234c-40b5-9ede-7413ad30dafe-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2173.288228] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f32f5e12-bb47-427a-aceb-0383bc172546 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Lock "2afd1f18-234c-40b5-9ede-7413ad30dafe-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2173.288404] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f32f5e12-bb47-427a-aceb-0383bc172546 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Lock "2afd1f18-234c-40b5-9ede-7413ad30dafe-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2173.290407] env[62824]: INFO nova.compute.manager [None req-f32f5e12-bb47-427a-aceb-0383bc172546 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 2afd1f18-234c-40b5-9ede-7413ad30dafe] Terminating instance [ 2173.493904] env[62824]: DEBUG oslo_vmware.api [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Task: {'id': task-2146497, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2173.645344] env[62824]: DEBUG oslo_vmware.api [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146495, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2173.668394] env[62824]: DEBUG nova.network.neutron [req-3b127b39-2aad-4784-8766-bf49f228fff7 req-22b813da-6771-47c8-aacf-19909eed57d7 service nova] [instance: 2afd1f18-234c-40b5-9ede-7413ad30dafe] Updated VIF entry in instance network info cache for port 6d0cd52e-ace8-4459-860b-f270fa266268. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2173.668759] env[62824]: DEBUG nova.network.neutron [req-3b127b39-2aad-4784-8766-bf49f228fff7 req-22b813da-6771-47c8-aacf-19909eed57d7 service nova] [instance: 2afd1f18-234c-40b5-9ede-7413ad30dafe] Updating instance_info_cache with network_info: [{"id": "6d0cd52e-ace8-4459-860b-f270fa266268", "address": "fa:16:3e:5b:95:b5", "network": {"id": "60037885-38d9-4586-9fd7-4907321e9c7e", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-370573912-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2e59203d676d48c8bd8402af4eb58f65", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f856fca-9fb5-41ea-a057-ac4193bd323d", "external-id": "nsx-vlan-transportzone-148", "segmentation_id": 148, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6d0cd52e-ac", "ovs_interfaceid": "6d0cd52e-ace8-4459-860b-f270fa266268", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2173.751660] env[62824]: DEBUG oslo_vmware.api [None req-06ef8715-0d4a-45b8-bcde-6870d18da01a tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Task: {'id': task-2146498, 'name': PowerOffVM_Task, 'duration_secs': 1.00338} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2173.751945] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-06ef8715-0d4a-45b8-bcde-6870d18da01a tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] [instance: ab37b1a2-0012-40fb-9341-b613525e89cf] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2173.752130] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-06ef8715-0d4a-45b8-bcde-6870d18da01a tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] [instance: ab37b1a2-0012-40fb-9341-b613525e89cf] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2173.752384] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-788afbe9-4895-4711-9281-7c481c03cb9a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2173.797351] env[62824]: DEBUG nova.compute.manager [None req-f32f5e12-bb47-427a-aceb-0383bc172546 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 2afd1f18-234c-40b5-9ede-7413ad30dafe] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2173.797576] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-f32f5e12-bb47-427a-aceb-0383bc172546 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 2afd1f18-234c-40b5-9ede-7413ad30dafe] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2173.798481] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12b5186d-ffb5-421e-805e-5b1c5bf4aef1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2173.809979] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-f32f5e12-bb47-427a-aceb-0383bc172546 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 2afd1f18-234c-40b5-9ede-7413ad30dafe] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2173.809979] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-02ebb569-5b7a-423c-9650-79cd5a1d9dac {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2173.815757] env[62824]: DEBUG oslo_vmware.api [None req-f32f5e12-bb47-427a-aceb-0383bc172546 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Waiting for the task: (returnval){ [ 2173.815757] env[62824]: value = "task-2146500" [ 2173.815757] env[62824]: _type = "Task" [ 2173.815757] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2173.829961] env[62824]: DEBUG oslo_vmware.api [None req-f32f5e12-bb47-427a-aceb-0383bc172546 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Task: {'id': task-2146500, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2173.857411] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-06ef8715-0d4a-45b8-bcde-6870d18da01a tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] [instance: ab37b1a2-0012-40fb-9341-b613525e89cf] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2173.857728] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-06ef8715-0d4a-45b8-bcde-6870d18da01a tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] [instance: ab37b1a2-0012-40fb-9341-b613525e89cf] Deleting contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2173.858010] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-06ef8715-0d4a-45b8-bcde-6870d18da01a tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Deleting the datastore file [datastore1] ab37b1a2-0012-40fb-9341-b613525e89cf {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2173.858357] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b7b4302c-5ef4-47ef-8e73-15e73a7a5c55 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2173.872030] env[62824]: DEBUG oslo_vmware.api [None req-06ef8715-0d4a-45b8-bcde-6870d18da01a tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Waiting for the task: (returnval){ [ 2173.872030] env[62824]: value = "task-2146501" [ 2173.872030] env[62824]: _type = "Task" [ 2173.872030] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2173.882427] env[62824]: DEBUG oslo_vmware.api [None req-06ef8715-0d4a-45b8-bcde-6870d18da01a tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Task: {'id': task-2146501, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2173.911572] env[62824]: DEBUG oslo_concurrency.lockutils [None req-123477fc-8c3e-4f2a-83de-cced4f011020 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Lock "b8cc8cd3-ea03-40bf-b867-7ad193365552" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 89.965s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2173.998910] env[62824]: DEBUG oslo_vmware.api [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Task: {'id': task-2146497, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2174.146962] env[62824]: DEBUG oslo_vmware.api [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146495, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.584895} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2174.149720] env[62824]: INFO nova.virt.vmwareapi.ds_util [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_cbb54347-fd26-4d25-aa17-fd598a754d99/OSTACK_IMG_cbb54347-fd26-4d25-aa17-fd598a754d99.vmdk to [datastore1] devstack-image-cache_base/758d9d98-91b8-46a1-9c95-5705c074be1a/758d9d98-91b8-46a1-9c95-5705c074be1a.vmdk. [ 2174.149984] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: da1ad739-b252-4e29-a22a-ab3bcab173ec] Cleaning up location [datastore1] OSTACK_IMG_cbb54347-fd26-4d25-aa17-fd598a754d99 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 2174.150122] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_cbb54347-fd26-4d25-aa17-fd598a754d99 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2174.150667] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7e4504a6-bb79-43c7-be63-60e1d08608d0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2174.158308] env[62824]: DEBUG oslo_vmware.api [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the task: (returnval){ [ 2174.158308] env[62824]: value = "task-2146502" [ 2174.158308] env[62824]: _type = "Task" [ 2174.158308] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2174.169059] env[62824]: DEBUG oslo_vmware.api [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146502, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2174.171608] env[62824]: DEBUG oslo_concurrency.lockutils [req-3b127b39-2aad-4784-8766-bf49f228fff7 req-22b813da-6771-47c8-aacf-19909eed57d7 service nova] Releasing lock "refresh_cache-2afd1f18-234c-40b5-9ede-7413ad30dafe" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2174.185361] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5eb1777d-6ae2-44f6-b618-90ff672d0f12 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2174.193719] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-902e7df8-bef5-4f89-8798-f726b99b7147 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2174.227402] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94a2b30c-cdc0-4f41-842c-3465332437fd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2174.235504] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea160779-534f-4d81-901b-6db4ad2bf824 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2174.249762] env[62824]: DEBUG nova.compute.provider_tree [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2174.327098] env[62824]: DEBUG oslo_vmware.api [None req-f32f5e12-bb47-427a-aceb-0383bc172546 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Task: {'id': task-2146500, 'name': PowerOffVM_Task, 'duration_secs': 0.210002} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2174.328424] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-f32f5e12-bb47-427a-aceb-0383bc172546 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 2afd1f18-234c-40b5-9ede-7413ad30dafe] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2174.328626] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-f32f5e12-bb47-427a-aceb-0383bc172546 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 2afd1f18-234c-40b5-9ede-7413ad30dafe] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2174.331073] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6a66e577-e91d-4df7-9928-aa791d57d937 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2174.332831] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Acquiring lock "c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2174.333061] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Lock "c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2174.381076] env[62824]: DEBUG oslo_vmware.api [None req-06ef8715-0d4a-45b8-bcde-6870d18da01a tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Task: {'id': task-2146501, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.201446} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2174.381326] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-06ef8715-0d4a-45b8-bcde-6870d18da01a tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2174.381512] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-06ef8715-0d4a-45b8-bcde-6870d18da01a tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] [instance: ab37b1a2-0012-40fb-9341-b613525e89cf] Deleted contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2174.381688] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-06ef8715-0d4a-45b8-bcde-6870d18da01a tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] [instance: ab37b1a2-0012-40fb-9341-b613525e89cf] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2174.381861] env[62824]: INFO nova.compute.manager [None req-06ef8715-0d4a-45b8-bcde-6870d18da01a tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] [instance: ab37b1a2-0012-40fb-9341-b613525e89cf] Took 1.67 seconds to destroy the instance on the hypervisor. [ 2174.382127] env[62824]: DEBUG oslo.service.loopingcall [None req-06ef8715-0d4a-45b8-bcde-6870d18da01a tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2174.382324] env[62824]: DEBUG nova.compute.manager [-] [instance: ab37b1a2-0012-40fb-9341-b613525e89cf] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2174.382494] env[62824]: DEBUG nova.network.neutron [-] [instance: ab37b1a2-0012-40fb-9341-b613525e89cf] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2174.439656] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-f32f5e12-bb47-427a-aceb-0383bc172546 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 2afd1f18-234c-40b5-9ede-7413ad30dafe] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2174.439869] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-f32f5e12-bb47-427a-aceb-0383bc172546 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 2afd1f18-234c-40b5-9ede-7413ad30dafe] Deleting contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2174.440092] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-f32f5e12-bb47-427a-aceb-0383bc172546 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Deleting the datastore file [datastore2] 2afd1f18-234c-40b5-9ede-7413ad30dafe {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2174.440354] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-76b5e078-66fb-4659-9d96-30c7ab2e6ff1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2174.447409] env[62824]: DEBUG oslo_vmware.api [None req-f32f5e12-bb47-427a-aceb-0383bc172546 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Waiting for the task: (returnval){ [ 2174.447409] env[62824]: value = "task-2146504" [ 2174.447409] env[62824]: _type = "Task" [ 2174.447409] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2174.456013] env[62824]: DEBUG oslo_vmware.api [None req-f32f5e12-bb47-427a-aceb-0383bc172546 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Task: {'id': task-2146504, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2174.493163] env[62824]: DEBUG oslo_vmware.api [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Task: {'id': task-2146497, 'name': PowerOnVM_Task, 'duration_secs': 1.619798} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2174.493433] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2174.493647] env[62824]: INFO nova.compute.manager [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Took 7.97 seconds to spawn the instance on the hypervisor. [ 2174.493860] env[62824]: DEBUG nova.compute.manager [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2174.496029] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7fc96fb-7194-479e-bf2c-a59b958c4f31 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2174.664856] env[62824]: DEBUG nova.compute.manager [req-9829609b-b59a-4d56-962a-dd908d7fa060 req-cf6092f0-53b3-4edb-a294-3a79df281f28 service nova] [instance: ab37b1a2-0012-40fb-9341-b613525e89cf] Received event network-vif-deleted-a7172f1f-6473-48fe-bec7-5c84411c8fd2 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2174.665133] env[62824]: INFO nova.compute.manager [req-9829609b-b59a-4d56-962a-dd908d7fa060 req-cf6092f0-53b3-4edb-a294-3a79df281f28 service nova] [instance: ab37b1a2-0012-40fb-9341-b613525e89cf] Neutron deleted interface a7172f1f-6473-48fe-bec7-5c84411c8fd2; detaching it from the instance and deleting it from the info cache [ 2174.665331] env[62824]: DEBUG nova.network.neutron [req-9829609b-b59a-4d56-962a-dd908d7fa060 req-cf6092f0-53b3-4edb-a294-3a79df281f28 service nova] [instance: ab37b1a2-0012-40fb-9341-b613525e89cf] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2174.669884] env[62824]: DEBUG oslo_vmware.api [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146502, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.040003} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2174.670315] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2174.670482] env[62824]: DEBUG oslo_concurrency.lockutils [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Releasing lock "[datastore1] devstack-image-cache_base/758d9d98-91b8-46a1-9c95-5705c074be1a/758d9d98-91b8-46a1-9c95-5705c074be1a.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2174.670720] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/758d9d98-91b8-46a1-9c95-5705c074be1a/758d9d98-91b8-46a1-9c95-5705c074be1a.vmdk to [datastore1] da1ad739-b252-4e29-a22a-ab3bcab173ec/da1ad739-b252-4e29-a22a-ab3bcab173ec.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2174.670999] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a741e302-9303-4982-a46c-4d083d7aa473 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2174.680833] env[62824]: DEBUG oslo_vmware.api [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the task: (returnval){ [ 2174.680833] env[62824]: value = "task-2146505" [ 2174.680833] env[62824]: _type = "Task" [ 2174.680833] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2174.689000] env[62824]: DEBUG oslo_vmware.api [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146505, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2174.752472] env[62824]: DEBUG nova.scheduler.client.report [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2174.835934] env[62824]: DEBUG nova.compute.manager [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2174.961830] env[62824]: DEBUG oslo_vmware.api [None req-f32f5e12-bb47-427a-aceb-0383bc172546 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Task: {'id': task-2146504, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.136011} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2174.961830] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-f32f5e12-bb47-427a-aceb-0383bc172546 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2174.961830] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-f32f5e12-bb47-427a-aceb-0383bc172546 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 2afd1f18-234c-40b5-9ede-7413ad30dafe] Deleted contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2174.962361] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-f32f5e12-bb47-427a-aceb-0383bc172546 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 2afd1f18-234c-40b5-9ede-7413ad30dafe] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2174.962361] env[62824]: INFO nova.compute.manager [None req-f32f5e12-bb47-427a-aceb-0383bc172546 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 2afd1f18-234c-40b5-9ede-7413ad30dafe] Took 1.16 seconds to destroy the instance on the hypervisor. [ 2174.962706] env[62824]: DEBUG oslo.service.loopingcall [None req-f32f5e12-bb47-427a-aceb-0383bc172546 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2174.962882] env[62824]: DEBUG nova.compute.manager [-] [instance: 2afd1f18-234c-40b5-9ede-7413ad30dafe] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2174.962969] env[62824]: DEBUG nova.network.neutron [-] [instance: 2afd1f18-234c-40b5-9ede-7413ad30dafe] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2175.012157] env[62824]: INFO nova.compute.manager [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Took 37.41 seconds to build instance. [ 2175.142280] env[62824]: DEBUG nova.network.neutron [-] [instance: ab37b1a2-0012-40fb-9341-b613525e89cf] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2175.171873] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f9b163c4-3cdd-480f-87f1-b110e7b00c62 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2175.186966] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b06511c-cabb-4cd3-8629-79e29277bf5d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2175.213409] env[62824]: DEBUG oslo_vmware.api [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146505, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2175.237388] env[62824]: DEBUG nova.compute.manager [req-9829609b-b59a-4d56-962a-dd908d7fa060 req-cf6092f0-53b3-4edb-a294-3a79df281f28 service nova] [instance: ab37b1a2-0012-40fb-9341-b613525e89cf] Detach interface failed, port_id=a7172f1f-6473-48fe-bec7-5c84411c8fd2, reason: Instance ab37b1a2-0012-40fb-9341-b613525e89cf could not be found. {{(pid=62824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11482}} [ 2175.257821] env[62824]: DEBUG oslo_concurrency.lockutils [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.532s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2175.258441] env[62824]: DEBUG nova.compute.manager [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 58243465-ba24-4b37-a640-04ab24963230] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2175.267828] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0f6836d6-2e7d-41c5-ba6f-3d4c9dc474aa tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.378s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2175.268124] env[62824]: DEBUG nova.objects.instance [None req-0f6836d6-2e7d-41c5-ba6f-3d4c9dc474aa tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Lazy-loading 'resources' on Instance uuid 29140542-dc4b-411a-ac10-7d84086eabbb {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2175.363249] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2175.368961] env[62824]: INFO nova.compute.manager [None req-0c0f7a2a-4ac1-4123-9e01-b25a8857edf5 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Rescuing [ 2175.369304] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0c0f7a2a-4ac1-4123-9e01-b25a8857edf5 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Acquiring lock "refresh_cache-da46d1d5-f75f-4ef1-b571-fbebab89c2a6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2175.369493] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0c0f7a2a-4ac1-4123-9e01-b25a8857edf5 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Acquired lock "refresh_cache-da46d1d5-f75f-4ef1-b571-fbebab89c2a6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2175.369673] env[62824]: DEBUG nova.network.neutron [None req-0c0f7a2a-4ac1-4123-9e01-b25a8857edf5 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2175.423398] env[62824]: INFO nova.compute.manager [None req-123477fc-8c3e-4f2a-83de-cced4f011020 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Successfully reverted task state from spawning on failure for instance. [ 2175.432329] env[62824]: ERROR oslo_messaging.rpc.server [None req-123477fc-8c3e-4f2a-83de-cced4f011020 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Exception during message handling: nova.exception.InvalidInput: Invalid input received: vif type binding_failed not supported [ 2175.432329] env[62824]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 2175.432329] env[62824]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/server.py", line 174, in _process_incoming [ 2175.432329] env[62824]: ERROR oslo_messaging.rpc.server res = self.dispatcher.dispatch(message) [ 2175.432329] env[62824]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 309, in dispatch [ 2175.432329] env[62824]: ERROR oslo_messaging.rpc.server return self._do_dispatch(endpoint, method, ctxt, args) [ 2175.432329] env[62824]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 229, in _do_dispatch [ 2175.432329] env[62824]: ERROR oslo_messaging.rpc.server result = func(ctxt, **new_args) [ 2175.432329] env[62824]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 65, in wrapped [ 2175.432329] env[62824]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 2175.432329] env[62824]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 2175.432329] env[62824]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 2175.432329] env[62824]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 2175.432329] env[62824]: ERROR oslo_messaging.rpc.server raise self.value [ 2175.432329] env[62824]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 63, in wrapped [ 2175.432329] env[62824]: ERROR oslo_messaging.rpc.server return f(self, context, *args, **kw) [ 2175.432329] env[62824]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 168, in decorated_function [ 2175.432329] env[62824]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 2175.432329] env[62824]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 2175.432329] env[62824]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 2175.432329] env[62824]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 2175.432329] env[62824]: ERROR oslo_messaging.rpc.server raise self.value [ 2175.432329] env[62824]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 159, in decorated_function [ 2175.432329] env[62824]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 2175.432329] env[62824]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/utils.py", line 1484, in decorated_function [ 2175.432329] env[62824]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 2175.432329] env[62824]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 215, in decorated_function [ 2175.432329] env[62824]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 2175.432329] env[62824]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 2175.432329] env[62824]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 2175.432329] env[62824]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 2175.432329] env[62824]: ERROR oslo_messaging.rpc.server raise self.value [ 2175.432329] env[62824]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 205, in decorated_function [ 2175.432329] env[62824]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 2175.432329] env[62824]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 7610, in unshelve_instance [ 2175.432329] env[62824]: ERROR oslo_messaging.rpc.server do_unshelve_instance() [ 2175.432329] env[62824]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 2175.432329] env[62824]: ERROR oslo_messaging.rpc.server return f(*args, **kwargs) [ 2175.432329] env[62824]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 7607, in do_unshelve_instance [ 2175.432329] env[62824]: ERROR oslo_messaging.rpc.server self._unshelve_instance( [ 2175.432329] env[62824]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 7698, in _unshelve_instance [ 2175.432329] env[62824]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(logger=LOG): [ 2175.432329] env[62824]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 2175.432329] env[62824]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 2175.432329] env[62824]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 2175.432329] env[62824]: ERROR oslo_messaging.rpc.server raise self.value [ 2175.432329] env[62824]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 7690, in _unshelve_instance [ 2175.432329] env[62824]: ERROR oslo_messaging.rpc.server self.driver.spawn(context, instance, image_meta, [ 2175.432329] env[62824]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 2175.432329] env[62824]: ERROR oslo_messaging.rpc.server self._vmops.spawn(context, instance, image_meta, injected_files, [ 2175.432329] env[62824]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 2175.432329] env[62824]: ERROR oslo_messaging.rpc.server vm_ref = self.build_virtual_machine(instance, [ 2175.432329] env[62824]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 2175.434975] env[62824]: ERROR oslo_messaging.rpc.server vif_infos = vmwarevif.get_vif_info(self._session, [ 2175.434975] env[62824]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 120, in get_vif_info [ 2175.434975] env[62824]: ERROR oslo_messaging.rpc.server vif_infos.append(get_vif_dict(session, cluster, vif_model, vif)) [ 2175.434975] env[62824]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 107, in get_vif_dict [ 2175.434975] env[62824]: ERROR oslo_messaging.rpc.server ref = get_network_ref(session, cluster, vif) [ 2175.434975] env[62824]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 100, in get_network_ref [ 2175.434975] env[62824]: ERROR oslo_messaging.rpc.server raise exception.InvalidInput(reason=reason) [ 2175.434975] env[62824]: ERROR oslo_messaging.rpc.server nova.exception.InvalidInput: Invalid input received: vif type binding_failed not supported [ 2175.434975] env[62824]: ERROR oslo_messaging.rpc.server [ 2175.514654] env[62824]: DEBUG oslo_concurrency.lockutils [None req-775e01c5-04fa-41a4-a85e-e444a6522a6c tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Lock "da46d1d5-f75f-4ef1-b571-fbebab89c2a6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 55.392s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2175.645318] env[62824]: INFO nova.compute.manager [-] [instance: ab37b1a2-0012-40fb-9341-b613525e89cf] Took 1.26 seconds to deallocate network for instance. [ 2175.709548] env[62824]: DEBUG oslo_vmware.api [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146505, 'name': CopyVirtualDisk_Task} progress is 43%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2175.769820] env[62824]: DEBUG nova.compute.utils [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2175.772683] env[62824]: DEBUG nova.compute.manager [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 58243465-ba24-4b37-a640-04ab24963230] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2175.773441] env[62824]: DEBUG nova.network.neutron [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 58243465-ba24-4b37-a640-04ab24963230] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2175.838773] env[62824]: DEBUG nova.policy [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c379c824a178444e99bb121716fc7887', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0c11aef7fb6247cb9b5272a6a063cd12', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2175.934362] env[62824]: DEBUG nova.network.neutron [-] [instance: 2afd1f18-234c-40b5-9ede-7413ad30dafe] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2176.133434] env[62824]: DEBUG nova.network.neutron [None req-0c0f7a2a-4ac1-4123-9e01-b25a8857edf5 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Updating instance_info_cache with network_info: [{"id": "0cefe7bc-5dd6-4f4c-872b-ef1b658a4d60", "address": "fa:16:3e:e0:51:b5", "network": {"id": "3210f2eb-2afb-4bce-bb59-f66e85888333", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-216881410-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "c8133feb4bde41e9898b3f7019c24d30", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d2bf584a-b4a3-4e7a-b0b7-eb8a2bc5a11d", "external-id": "nsx-vlan-transportzone-286", "segmentation_id": 286, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0cefe7bc-5d", "ovs_interfaceid": "0cefe7bc-5dd6-4f4c-872b-ef1b658a4d60", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2176.154045] env[62824]: DEBUG oslo_concurrency.lockutils [None req-06ef8715-0d4a-45b8-bcde-6870d18da01a tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2176.210025] env[62824]: DEBUG oslo_vmware.api [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146505, 'name': CopyVirtualDisk_Task} progress is 66%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2176.282542] env[62824]: DEBUG nova.compute.manager [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 58243465-ba24-4b37-a640-04ab24963230] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2176.285968] env[62824]: DEBUG nova.network.neutron [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 58243465-ba24-4b37-a640-04ab24963230] Successfully created port: 63af4d3b-ce05-431b-84e1-e3e3fbdf673d {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2176.315101] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc851e64-77b8-4d7c-bccc-1a97663ef484 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2176.326737] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28f90669-8286-40a0-84c2-71c94f04278c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2176.359461] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03efe6e6-d92f-4513-959f-00884ba52c6e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2176.369631] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddfb25e3-355f-4abe-a748-86446cbeb649 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2176.385495] env[62824]: DEBUG nova.compute.provider_tree [None req-0f6836d6-2e7d-41c5-ba6f-3d4c9dc474aa tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2176.438538] env[62824]: INFO nova.compute.manager [-] [instance: 2afd1f18-234c-40b5-9ede-7413ad30dafe] Took 1.47 seconds to deallocate network for instance. [ 2176.635712] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0c0f7a2a-4ac1-4123-9e01-b25a8857edf5 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Releasing lock "refresh_cache-da46d1d5-f75f-4ef1-b571-fbebab89c2a6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2176.693657] env[62824]: DEBUG nova.compute.manager [req-a9a00997-f511-4ca8-9375-e668743983de req-a0029b59-134d-4235-bfd7-092d738a1094 service nova] [instance: 2afd1f18-234c-40b5-9ede-7413ad30dafe] Received event network-vif-deleted-6d0cd52e-ace8-4459-860b-f270fa266268 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2176.709888] env[62824]: DEBUG oslo_vmware.api [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146505, 'name': CopyVirtualDisk_Task} progress is 85%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2176.895441] env[62824]: DEBUG nova.scheduler.client.report [None req-0f6836d6-2e7d-41c5-ba6f-3d4c9dc474aa tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2176.944041] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f32f5e12-bb47-427a-aceb-0383bc172546 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2177.210983] env[62824]: DEBUG oslo_vmware.api [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146505, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.353554} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2177.211272] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/758d9d98-91b8-46a1-9c95-5705c074be1a/758d9d98-91b8-46a1-9c95-5705c074be1a.vmdk to [datastore1] da1ad739-b252-4e29-a22a-ab3bcab173ec/da1ad739-b252-4e29-a22a-ab3bcab173ec.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2177.212070] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2a4286c-1b18-46f4-89cb-58427878f156 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2177.235765] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: da1ad739-b252-4e29-a22a-ab3bcab173ec] Reconfiguring VM instance instance-00000051 to attach disk [datastore1] da1ad739-b252-4e29-a22a-ab3bcab173ec/da1ad739-b252-4e29-a22a-ab3bcab173ec.vmdk or device None with type streamOptimized {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2177.236060] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-86427c25-c5f2-44b2-86c5-c3027934d0ca {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2177.256870] env[62824]: DEBUG oslo_vmware.api [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the task: (returnval){ [ 2177.256870] env[62824]: value = "task-2146506" [ 2177.256870] env[62824]: _type = "Task" [ 2177.256870] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2177.265643] env[62824]: DEBUG oslo_vmware.api [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146506, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2177.295502] env[62824]: DEBUG nova.compute.manager [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 58243465-ba24-4b37-a640-04ab24963230] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2177.323208] env[62824]: DEBUG nova.virt.hardware [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2177.323562] env[62824]: DEBUG nova.virt.hardware [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2177.323652] env[62824]: DEBUG nova.virt.hardware [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2177.323844] env[62824]: DEBUG nova.virt.hardware [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2177.324013] env[62824]: DEBUG nova.virt.hardware [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2177.324178] env[62824]: DEBUG nova.virt.hardware [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2177.324405] env[62824]: DEBUG nova.virt.hardware [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2177.324582] env[62824]: DEBUG nova.virt.hardware [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2177.324768] env[62824]: DEBUG nova.virt.hardware [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2177.324965] env[62824]: DEBUG nova.virt.hardware [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2177.325186] env[62824]: DEBUG nova.virt.hardware [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2177.326133] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e47c893-c02c-4859-b74c-c350f708b5be {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2177.334758] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff5ef4fd-8052-486b-bcfc-21d761e2332d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2177.401197] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0f6836d6-2e7d-41c5-ba6f-3d4c9dc474aa tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.133s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2177.403888] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ae8aad58-b430-4dd1-8bc0-b6bcc01e69fc tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.592s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2177.404229] env[62824]: DEBUG nova.objects.instance [None req-ae8aad58-b430-4dd1-8bc0-b6bcc01e69fc tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Lazy-loading 'resources' on Instance uuid dadf590a-7288-4dd2-90de-125fe272f4ba {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2177.431063] env[62824]: INFO nova.scheduler.client.report [None req-0f6836d6-2e7d-41c5-ba6f-3d4c9dc474aa tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Deleted allocations for instance 29140542-dc4b-411a-ac10-7d84086eabbb [ 2177.767164] env[62824]: DEBUG oslo_vmware.api [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146506, 'name': ReconfigVM_Task, 'duration_secs': 0.302718} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2177.767469] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: da1ad739-b252-4e29-a22a-ab3bcab173ec] Reconfigured VM instance instance-00000051 to attach disk [datastore1] da1ad739-b252-4e29-a22a-ab3bcab173ec/da1ad739-b252-4e29-a22a-ab3bcab173ec.vmdk or device None with type streamOptimized {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2177.768201] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3d0fbd3e-395e-44d4-a798-002f904c6d7a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2177.776109] env[62824]: DEBUG oslo_vmware.api [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the task: (returnval){ [ 2177.776109] env[62824]: value = "task-2146507" [ 2177.776109] env[62824]: _type = "Task" [ 2177.776109] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2177.791920] env[62824]: DEBUG oslo_vmware.api [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146507, 'name': Rename_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2177.939445] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0f6836d6-2e7d-41c5-ba6f-3d4c9dc474aa tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Lock "29140542-dc4b-411a-ac10-7d84086eabbb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.599s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2177.946925] env[62824]: DEBUG nova.network.neutron [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 58243465-ba24-4b37-a640-04ab24963230] Successfully updated port: 63af4d3b-ce05-431b-84e1-e3e3fbdf673d {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2178.182512] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c0f7a2a-4ac1-4123-9e01-b25a8857edf5 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2178.183075] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fba9994a-7afd-47ea-a5f3-21fe911a7296 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2178.192110] env[62824]: DEBUG oslo_vmware.api [None req-0c0f7a2a-4ac1-4123-9e01-b25a8857edf5 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Waiting for the task: (returnval){ [ 2178.192110] env[62824]: value = "task-2146508" [ 2178.192110] env[62824]: _type = "Task" [ 2178.192110] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2178.204038] env[62824]: DEBUG oslo_vmware.api [None req-0c0f7a2a-4ac1-4123-9e01-b25a8857edf5 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Task: {'id': task-2146508, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2178.287137] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbe60260-05e2-4632-83ff-45d9e9a31f0c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2178.293072] env[62824]: DEBUG oslo_vmware.api [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146507, 'name': Rename_Task, 'duration_secs': 0.13713} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2178.294238] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: da1ad739-b252-4e29-a22a-ab3bcab173ec] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2178.295234] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-739a2f8c-ef53-486f-a5d0-5af3f37e417e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2178.302140] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26edef2c-33c5-49ad-b4fe-0f73cdb91e50 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2178.308027] env[62824]: DEBUG oslo_vmware.api [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the task: (returnval){ [ 2178.308027] env[62824]: value = "task-2146509" [ 2178.308027] env[62824]: _type = "Task" [ 2178.308027] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2178.343832] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c91bc1b7-a748-41df-bddf-3dcdd733b56d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2178.352592] env[62824]: DEBUG oslo_vmware.api [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146509, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2178.360553] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e6a3b7c-c93d-48a0-af8c-85b87c312466 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2178.376919] env[62824]: DEBUG nova.compute.provider_tree [None req-ae8aad58-b430-4dd1-8bc0-b6bcc01e69fc tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2178.451878] env[62824]: DEBUG oslo_concurrency.lockutils [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquiring lock "refresh_cache-58243465-ba24-4b37-a640-04ab24963230" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2178.451878] env[62824]: DEBUG oslo_concurrency.lockutils [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquired lock "refresh_cache-58243465-ba24-4b37-a640-04ab24963230" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2178.451878] env[62824]: DEBUG nova.network.neutron [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 58243465-ba24-4b37-a640-04ab24963230] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2178.703246] env[62824]: DEBUG oslo_vmware.api [None req-0c0f7a2a-4ac1-4123-9e01-b25a8857edf5 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Task: {'id': task-2146508, 'name': PowerOffVM_Task, 'duration_secs': 0.215155} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2178.703246] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c0f7a2a-4ac1-4123-9e01-b25a8857edf5 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2178.704080] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5b68823-a9d8-4036-93a4-a8c179aa10b8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2178.724823] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebe24b6f-333e-4792-8156-46f9760d6ccf {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2178.738190] env[62824]: DEBUG nova.compute.manager [req-6ae52aba-e217-4750-91fb-4f2425accbbc req-93a1ba76-f268-41f3-ae8b-8584de9b9de8 service nova] [instance: 58243465-ba24-4b37-a640-04ab24963230] Received event network-vif-plugged-63af4d3b-ce05-431b-84e1-e3e3fbdf673d {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2178.738398] env[62824]: DEBUG oslo_concurrency.lockutils [req-6ae52aba-e217-4750-91fb-4f2425accbbc req-93a1ba76-f268-41f3-ae8b-8584de9b9de8 service nova] Acquiring lock "58243465-ba24-4b37-a640-04ab24963230-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2178.738778] env[62824]: DEBUG oslo_concurrency.lockutils [req-6ae52aba-e217-4750-91fb-4f2425accbbc req-93a1ba76-f268-41f3-ae8b-8584de9b9de8 service nova] Lock "58243465-ba24-4b37-a640-04ab24963230-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2178.738778] env[62824]: DEBUG oslo_concurrency.lockutils [req-6ae52aba-e217-4750-91fb-4f2425accbbc req-93a1ba76-f268-41f3-ae8b-8584de9b9de8 service nova] Lock "58243465-ba24-4b37-a640-04ab24963230-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2178.739081] env[62824]: DEBUG nova.compute.manager [req-6ae52aba-e217-4750-91fb-4f2425accbbc req-93a1ba76-f268-41f3-ae8b-8584de9b9de8 service nova] [instance: 58243465-ba24-4b37-a640-04ab24963230] No waiting events found dispatching network-vif-plugged-63af4d3b-ce05-431b-84e1-e3e3fbdf673d {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2178.739161] env[62824]: WARNING nova.compute.manager [req-6ae52aba-e217-4750-91fb-4f2425accbbc req-93a1ba76-f268-41f3-ae8b-8584de9b9de8 service nova] [instance: 58243465-ba24-4b37-a640-04ab24963230] Received unexpected event network-vif-plugged-63af4d3b-ce05-431b-84e1-e3e3fbdf673d for instance with vm_state building and task_state spawning. [ 2178.739287] env[62824]: DEBUG nova.compute.manager [req-6ae52aba-e217-4750-91fb-4f2425accbbc req-93a1ba76-f268-41f3-ae8b-8584de9b9de8 service nova] [instance: 58243465-ba24-4b37-a640-04ab24963230] Received event network-changed-63af4d3b-ce05-431b-84e1-e3e3fbdf673d {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2178.739440] env[62824]: DEBUG nova.compute.manager [req-6ae52aba-e217-4750-91fb-4f2425accbbc req-93a1ba76-f268-41f3-ae8b-8584de9b9de8 service nova] [instance: 58243465-ba24-4b37-a640-04ab24963230] Refreshing instance network info cache due to event network-changed-63af4d3b-ce05-431b-84e1-e3e3fbdf673d. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2178.739620] env[62824]: DEBUG oslo_concurrency.lockutils [req-6ae52aba-e217-4750-91fb-4f2425accbbc req-93a1ba76-f268-41f3-ae8b-8584de9b9de8 service nova] Acquiring lock "refresh_cache-58243465-ba24-4b37-a640-04ab24963230" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2178.763027] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c0f7a2a-4ac1-4123-9e01-b25a8857edf5 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2178.763027] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f80b92b4-dc7f-48dd-ba93-d349f261e15f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2178.771753] env[62824]: DEBUG oslo_vmware.api [None req-0c0f7a2a-4ac1-4123-9e01-b25a8857edf5 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Waiting for the task: (returnval){ [ 2178.771753] env[62824]: value = "task-2146510" [ 2178.771753] env[62824]: _type = "Task" [ 2178.771753] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2178.782316] env[62824]: DEBUG oslo_vmware.api [None req-0c0f7a2a-4ac1-4123-9e01-b25a8857edf5 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Task: {'id': task-2146510, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2178.818910] env[62824]: DEBUG oslo_vmware.api [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146509, 'name': PowerOnVM_Task, 'duration_secs': 0.465022} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2178.819219] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: da1ad739-b252-4e29-a22a-ab3bcab173ec] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2178.819409] env[62824]: INFO nova.compute.manager [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: da1ad739-b252-4e29-a22a-ab3bcab173ec] Took 14.86 seconds to spawn the instance on the hypervisor. [ 2178.819585] env[62824]: DEBUG nova.compute.manager [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: da1ad739-b252-4e29-a22a-ab3bcab173ec] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2178.820433] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-186262b4-811a-4abd-ae3f-fce65748247a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2178.880881] env[62824]: DEBUG nova.scheduler.client.report [None req-ae8aad58-b430-4dd1-8bc0-b6bcc01e69fc tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2178.998265] env[62824]: DEBUG nova.network.neutron [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 58243465-ba24-4b37-a640-04ab24963230] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2179.192273] env[62824]: DEBUG nova.network.neutron [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 58243465-ba24-4b37-a640-04ab24963230] Updating instance_info_cache with network_info: [{"id": "63af4d3b-ce05-431b-84e1-e3e3fbdf673d", "address": "fa:16:3e:da:11:f0", "network": {"id": "b995934c-4934-4469-884c-73cb4de1796f", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1892307918-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c11aef7fb6247cb9b5272a6a063cd12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15165046-2de9-4ada-9e99-0126e20854a9", "external-id": "nsx-vlan-transportzone-974", "segmentation_id": 974, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap63af4d3b-ce", "ovs_interfaceid": "63af4d3b-ce05-431b-84e1-e3e3fbdf673d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2179.283318] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c0f7a2a-4ac1-4123-9e01-b25a8857edf5 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] VM already powered off {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 2179.283563] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-0c0f7a2a-4ac1-4123-9e01-b25a8857edf5 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2179.283809] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0c0f7a2a-4ac1-4123-9e01-b25a8857edf5 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2179.283962] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0c0f7a2a-4ac1-4123-9e01-b25a8857edf5 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2179.284157] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-0c0f7a2a-4ac1-4123-9e01-b25a8857edf5 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2179.284403] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5f177580-2ade-4e63-adbf-5997cc329a14 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2179.294210] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-0c0f7a2a-4ac1-4123-9e01-b25a8857edf5 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2179.294372] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-0c0f7a2a-4ac1-4123-9e01-b25a8857edf5 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2179.295175] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-54145318-4272-4ce1-bcd2-2a7133f01f66 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2179.301430] env[62824]: DEBUG oslo_vmware.api [None req-0c0f7a2a-4ac1-4123-9e01-b25a8857edf5 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Waiting for the task: (returnval){ [ 2179.301430] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5287a122-dd65-2c07-e358-907e41561e50" [ 2179.301430] env[62824]: _type = "Task" [ 2179.301430] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2179.309770] env[62824]: DEBUG oslo_vmware.api [None req-0c0f7a2a-4ac1-4123-9e01-b25a8857edf5 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5287a122-dd65-2c07-e358-907e41561e50, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2179.336945] env[62824]: INFO nova.compute.manager [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: da1ad739-b252-4e29-a22a-ab3bcab173ec] Took 44.79 seconds to build instance. [ 2179.386176] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ae8aad58-b430-4dd1-8bc0-b6bcc01e69fc tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.982s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2179.388667] env[62824]: DEBUG oslo_concurrency.lockutils [None req-947b662c-414a-42b6-bd23-f5e60b455b8c tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.937s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2179.388905] env[62824]: DEBUG nova.objects.instance [None req-947b662c-414a-42b6-bd23-f5e60b455b8c tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Lazy-loading 'resources' on Instance uuid 02405b17-7022-4e2c-a357-415de88f63d0 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2179.412096] env[62824]: INFO nova.scheduler.client.report [None req-ae8aad58-b430-4dd1-8bc0-b6bcc01e69fc tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Deleted allocations for instance dadf590a-7288-4dd2-90de-125fe272f4ba [ 2179.695232] env[62824]: DEBUG oslo_concurrency.lockutils [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Releasing lock "refresh_cache-58243465-ba24-4b37-a640-04ab24963230" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2179.695580] env[62824]: DEBUG nova.compute.manager [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 58243465-ba24-4b37-a640-04ab24963230] Instance network_info: |[{"id": "63af4d3b-ce05-431b-84e1-e3e3fbdf673d", "address": "fa:16:3e:da:11:f0", "network": {"id": "b995934c-4934-4469-884c-73cb4de1796f", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1892307918-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c11aef7fb6247cb9b5272a6a063cd12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15165046-2de9-4ada-9e99-0126e20854a9", "external-id": "nsx-vlan-transportzone-974", "segmentation_id": 974, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap63af4d3b-ce", "ovs_interfaceid": "63af4d3b-ce05-431b-84e1-e3e3fbdf673d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2179.695932] env[62824]: DEBUG oslo_concurrency.lockutils [req-6ae52aba-e217-4750-91fb-4f2425accbbc req-93a1ba76-f268-41f3-ae8b-8584de9b9de8 service nova] Acquired lock "refresh_cache-58243465-ba24-4b37-a640-04ab24963230" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2179.696139] env[62824]: DEBUG nova.network.neutron [req-6ae52aba-e217-4750-91fb-4f2425accbbc req-93a1ba76-f268-41f3-ae8b-8584de9b9de8 service nova] [instance: 58243465-ba24-4b37-a640-04ab24963230] Refreshing network info cache for port 63af4d3b-ce05-431b-84e1-e3e3fbdf673d {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2179.697664] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 58243465-ba24-4b37-a640-04ab24963230] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:da:11:f0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '15165046-2de9-4ada-9e99-0126e20854a9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '63af4d3b-ce05-431b-84e1-e3e3fbdf673d', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2179.705362] env[62824]: DEBUG oslo.service.loopingcall [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2179.708324] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 58243465-ba24-4b37-a640-04ab24963230] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2179.709045] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-31b4e191-c06d-42a4-a084-831156457270 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2179.724162] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0c3b6169-674a-4946-8b27-bab89f3484d0 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Acquiring lock "d067fa97-cedc-4e3d-9be4-d860a79a7723" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2179.724351] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0c3b6169-674a-4946-8b27-bab89f3484d0 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Lock "d067fa97-cedc-4e3d-9be4-d860a79a7723" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2179.724575] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0c3b6169-674a-4946-8b27-bab89f3484d0 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Acquiring lock "d067fa97-cedc-4e3d-9be4-d860a79a7723-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2179.724771] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0c3b6169-674a-4946-8b27-bab89f3484d0 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Lock "d067fa97-cedc-4e3d-9be4-d860a79a7723-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2179.724952] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0c3b6169-674a-4946-8b27-bab89f3484d0 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Lock "d067fa97-cedc-4e3d-9be4-d860a79a7723-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2179.727276] env[62824]: INFO nova.compute.manager [None req-0c3b6169-674a-4946-8b27-bab89f3484d0 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: d067fa97-cedc-4e3d-9be4-d860a79a7723] Terminating instance [ 2179.733210] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2179.733210] env[62824]: value = "task-2146511" [ 2179.733210] env[62824]: _type = "Task" [ 2179.733210] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2179.748596] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146511, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2179.815199] env[62824]: DEBUG oslo_vmware.api [None req-0c0f7a2a-4ac1-4123-9e01-b25a8857edf5 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5287a122-dd65-2c07-e358-907e41561e50, 'name': SearchDatastore_Task, 'duration_secs': 0.010556} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2179.816125] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ffc2cd68-7793-40c8-9d0b-de7194e37bbf {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2179.822462] env[62824]: DEBUG oslo_vmware.api [None req-0c0f7a2a-4ac1-4123-9e01-b25a8857edf5 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Waiting for the task: (returnval){ [ 2179.822462] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]525dac76-acaf-159d-4159-c25208dd7519" [ 2179.822462] env[62824]: _type = "Task" [ 2179.822462] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2179.831206] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0235e5cb-af08-4951-9315-366312405d6f tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Acquiring lock "4c2c2068-664d-404f-a99e-8fc7719f43e5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2179.831457] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0235e5cb-af08-4951-9315-366312405d6f tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Lock "4c2c2068-664d-404f-a99e-8fc7719f43e5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2179.831681] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0235e5cb-af08-4951-9315-366312405d6f tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Acquiring lock "4c2c2068-664d-404f-a99e-8fc7719f43e5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2179.831870] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0235e5cb-af08-4951-9315-366312405d6f tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Lock "4c2c2068-664d-404f-a99e-8fc7719f43e5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2179.832076] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0235e5cb-af08-4951-9315-366312405d6f tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Lock "4c2c2068-664d-404f-a99e-8fc7719f43e5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2179.833619] env[62824]: DEBUG oslo_vmware.api [None req-0c0f7a2a-4ac1-4123-9e01-b25a8857edf5 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]525dac76-acaf-159d-4159-c25208dd7519, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2179.834081] env[62824]: INFO nova.compute.manager [None req-0235e5cb-af08-4951-9315-366312405d6f tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 4c2c2068-664d-404f-a99e-8fc7719f43e5] Terminating instance [ 2179.838820] env[62824]: DEBUG oslo_concurrency.lockutils [None req-242fa478-ba6e-4b9e-9df2-ace707210fe0 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Lock "da1ad739-b252-4e29-a22a-ab3bcab173ec" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 60.701s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2179.922658] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ae8aad58-b430-4dd1-8bc0-b6bcc01e69fc tempest-ServersAaction247Test-1341876915 tempest-ServersAaction247Test-1341876915-project-member] Lock "dadf590a-7288-4dd2-90de-125fe272f4ba" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.354s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2180.231606] env[62824]: DEBUG nova.compute.manager [None req-0c3b6169-674a-4946-8b27-bab89f3484d0 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: d067fa97-cedc-4e3d-9be4-d860a79a7723] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2180.231951] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-0c3b6169-674a-4946-8b27-bab89f3484d0 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: d067fa97-cedc-4e3d-9be4-d860a79a7723] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2180.232862] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2612876-cfbe-4d74-aaa8-e8a4e45ad6bc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2180.246575] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146511, 'name': CreateVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2180.248918] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c3b6169-674a-4946-8b27-bab89f3484d0 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: d067fa97-cedc-4e3d-9be4-d860a79a7723] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2180.249287] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0f14af33-6c22-4e74-a966-918e0acbf5f3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2180.256605] env[62824]: DEBUG oslo_vmware.api [None req-0c3b6169-674a-4946-8b27-bab89f3484d0 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Waiting for the task: (returnval){ [ 2180.256605] env[62824]: value = "task-2146512" [ 2180.256605] env[62824]: _type = "Task" [ 2180.256605] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2180.266039] env[62824]: DEBUG oslo_vmware.api [None req-0c3b6169-674a-4946-8b27-bab89f3484d0 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Task: {'id': task-2146512, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2180.270774] env[62824]: DEBUG nova.network.neutron [req-6ae52aba-e217-4750-91fb-4f2425accbbc req-93a1ba76-f268-41f3-ae8b-8584de9b9de8 service nova] [instance: 58243465-ba24-4b37-a640-04ab24963230] Updated VIF entry in instance network info cache for port 63af4d3b-ce05-431b-84e1-e3e3fbdf673d. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2180.271155] env[62824]: DEBUG nova.network.neutron [req-6ae52aba-e217-4750-91fb-4f2425accbbc req-93a1ba76-f268-41f3-ae8b-8584de9b9de8 service nova] [instance: 58243465-ba24-4b37-a640-04ab24963230] Updating instance_info_cache with network_info: [{"id": "63af4d3b-ce05-431b-84e1-e3e3fbdf673d", "address": "fa:16:3e:da:11:f0", "network": {"id": "b995934c-4934-4469-884c-73cb4de1796f", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1892307918-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c11aef7fb6247cb9b5272a6a063cd12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15165046-2de9-4ada-9e99-0126e20854a9", "external-id": "nsx-vlan-transportzone-974", "segmentation_id": 974, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap63af4d3b-ce", "ovs_interfaceid": "63af4d3b-ce05-431b-84e1-e3e3fbdf673d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2180.336174] env[62824]: DEBUG oslo_vmware.api [None req-0c0f7a2a-4ac1-4123-9e01-b25a8857edf5 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]525dac76-acaf-159d-4159-c25208dd7519, 'name': SearchDatastore_Task, 'duration_secs': 0.010484} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2180.336174] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0c0f7a2a-4ac1-4123-9e01-b25a8857edf5 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2180.336496] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-0c0f7a2a-4ac1-4123-9e01-b25a8857edf5 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] da46d1d5-f75f-4ef1-b571-fbebab89c2a6/9e2a7d30-212d-4ab8-9606-c5c6d52629e8-rescue.vmdk. {{(pid=62824) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 2180.337199] env[62824]: DEBUG nova.compute.manager [None req-0235e5cb-af08-4951-9315-366312405d6f tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 4c2c2068-664d-404f-a99e-8fc7719f43e5] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2180.337402] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-0235e5cb-af08-4951-9315-366312405d6f tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 4c2c2068-664d-404f-a99e-8fc7719f43e5] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2180.337655] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2dc64443-362c-4e12-9d43-06b13c9640f3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2180.340896] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1c2428f-52a0-4c60-8721-7b16a3e3378b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2180.355858] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-0235e5cb-af08-4951-9315-366312405d6f tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 4c2c2068-664d-404f-a99e-8fc7719f43e5] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2180.356754] env[62824]: DEBUG oslo_vmware.api [None req-0c0f7a2a-4ac1-4123-9e01-b25a8857edf5 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Waiting for the task: (returnval){ [ 2180.356754] env[62824]: value = "task-2146513" [ 2180.356754] env[62824]: _type = "Task" [ 2180.356754] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2180.356754] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2e1c7d0f-6d5a-477f-8694-e5d5eb805a3a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2180.369293] env[62824]: DEBUG oslo_vmware.api [None req-0c0f7a2a-4ac1-4123-9e01-b25a8857edf5 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Task: {'id': task-2146513, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2180.370951] env[62824]: DEBUG oslo_vmware.api [None req-0235e5cb-af08-4951-9315-366312405d6f tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Waiting for the task: (returnval){ [ 2180.370951] env[62824]: value = "task-2146514" [ 2180.370951] env[62824]: _type = "Task" [ 2180.370951] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2180.379533] env[62824]: DEBUG oslo_vmware.api [None req-0235e5cb-af08-4951-9315-366312405d6f tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Task: {'id': task-2146514, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2180.459019] env[62824]: DEBUG oslo_concurrency.lockutils [None req-cac6081d-dcbb-4024-bf06-8c5d7fd6c54c tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Acquiring lock "da1ad739-b252-4e29-a22a-ab3bcab173ec" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2180.459019] env[62824]: DEBUG oslo_concurrency.lockutils [None req-cac6081d-dcbb-4024-bf06-8c5d7fd6c54c tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Lock "da1ad739-b252-4e29-a22a-ab3bcab173ec" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.002s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2180.459019] env[62824]: DEBUG oslo_concurrency.lockutils [None req-cac6081d-dcbb-4024-bf06-8c5d7fd6c54c tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Acquiring lock "da1ad739-b252-4e29-a22a-ab3bcab173ec-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2180.459019] env[62824]: DEBUG oslo_concurrency.lockutils [None req-cac6081d-dcbb-4024-bf06-8c5d7fd6c54c tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Lock "da1ad739-b252-4e29-a22a-ab3bcab173ec-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2180.459019] env[62824]: DEBUG oslo_concurrency.lockutils [None req-cac6081d-dcbb-4024-bf06-8c5d7fd6c54c tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Lock "da1ad739-b252-4e29-a22a-ab3bcab173ec-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2180.462019] env[62824]: INFO nova.compute.manager [None req-cac6081d-dcbb-4024-bf06-8c5d7fd6c54c tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: da1ad739-b252-4e29-a22a-ab3bcab173ec] Terminating instance [ 2180.476163] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6929d514-b255-4595-bafa-14a033b19b09 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2180.485723] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90281713-aa98-48c4-9e09-9b1ca55a353b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2180.518676] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0b0c433-4656-48f4-b4ef-db94e552b438 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2180.528114] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88a83723-2cb1-450d-9618-98bc77dcbcef {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2180.545991] env[62824]: DEBUG nova.compute.provider_tree [None req-947b662c-414a-42b6-bd23-f5e60b455b8c tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2180.755581] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146511, 'name': CreateVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2180.770120] env[62824]: DEBUG oslo_vmware.api [None req-0c3b6169-674a-4946-8b27-bab89f3484d0 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Task: {'id': task-2146512, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2180.778101] env[62824]: DEBUG oslo_concurrency.lockutils [req-6ae52aba-e217-4750-91fb-4f2425accbbc req-93a1ba76-f268-41f3-ae8b-8584de9b9de8 service nova] Releasing lock "refresh_cache-58243465-ba24-4b37-a640-04ab24963230" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2180.869090] env[62824]: DEBUG oslo_vmware.api [None req-0c0f7a2a-4ac1-4123-9e01-b25a8857edf5 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Task: {'id': task-2146513, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.461954} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2180.869371] env[62824]: INFO nova.virt.vmwareapi.ds_util [None req-0c0f7a2a-4ac1-4123-9e01-b25a8857edf5 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] da46d1d5-f75f-4ef1-b571-fbebab89c2a6/9e2a7d30-212d-4ab8-9606-c5c6d52629e8-rescue.vmdk. [ 2180.870458] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ddcbe30-ecb7-48b4-9f0e-a20e540a4cb0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2180.900914] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-0c0f7a2a-4ac1-4123-9e01-b25a8857edf5 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Reconfiguring VM instance instance-00000052 to attach disk [datastore1] da46d1d5-f75f-4ef1-b571-fbebab89c2a6/9e2a7d30-212d-4ab8-9606-c5c6d52629e8-rescue.vmdk or device None with type thin {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2180.904120] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-db865d0d-d5a4-4443-bc96-0eee6d8499cc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2180.917031] env[62824]: DEBUG oslo_vmware.api [None req-0235e5cb-af08-4951-9315-366312405d6f tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Task: {'id': task-2146514, 'name': PowerOffVM_Task, 'duration_secs': 0.196782} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2180.917320] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-0235e5cb-af08-4951-9315-366312405d6f tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 4c2c2068-664d-404f-a99e-8fc7719f43e5] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2180.917489] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-0235e5cb-af08-4951-9315-366312405d6f tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 4c2c2068-664d-404f-a99e-8fc7719f43e5] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2180.918491] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8b98a5e5-ef49-42f2-9bd9-4b85acf1ef64 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2180.923621] env[62824]: DEBUG oslo_vmware.api [None req-0c0f7a2a-4ac1-4123-9e01-b25a8857edf5 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Waiting for the task: (returnval){ [ 2180.923621] env[62824]: value = "task-2146515" [ 2180.923621] env[62824]: _type = "Task" [ 2180.923621] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2180.934688] env[62824]: DEBUG oslo_vmware.api [None req-0c0f7a2a-4ac1-4123-9e01-b25a8857edf5 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Task: {'id': task-2146515, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2180.966671] env[62824]: DEBUG nova.compute.manager [None req-cac6081d-dcbb-4024-bf06-8c5d7fd6c54c tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: da1ad739-b252-4e29-a22a-ab3bcab173ec] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2180.966953] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-cac6081d-dcbb-4024-bf06-8c5d7fd6c54c tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: da1ad739-b252-4e29-a22a-ab3bcab173ec] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2180.967895] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff97584f-42d0-422f-a1cb-cc64088ccee6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2180.977625] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-cac6081d-dcbb-4024-bf06-8c5d7fd6c54c tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: da1ad739-b252-4e29-a22a-ab3bcab173ec] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2180.977941] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3d58e1cd-aa3e-4d13-b674-4aba33048bb1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2180.985772] env[62824]: DEBUG oslo_vmware.api [None req-cac6081d-dcbb-4024-bf06-8c5d7fd6c54c tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the task: (returnval){ [ 2180.985772] env[62824]: value = "task-2146517" [ 2180.985772] env[62824]: _type = "Task" [ 2180.985772] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2180.995473] env[62824]: DEBUG oslo_vmware.api [None req-cac6081d-dcbb-4024-bf06-8c5d7fd6c54c tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146517, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2181.042685] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-0235e5cb-af08-4951-9315-366312405d6f tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 4c2c2068-664d-404f-a99e-8fc7719f43e5] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2181.043256] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-0235e5cb-af08-4951-9315-366312405d6f tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 4c2c2068-664d-404f-a99e-8fc7719f43e5] Deleting contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2181.043566] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-0235e5cb-af08-4951-9315-366312405d6f tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Deleting the datastore file [datastore2] 4c2c2068-664d-404f-a99e-8fc7719f43e5 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2181.043997] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bad48a50-b943-4ec0-9b98-c4ebd87c7508 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2181.049892] env[62824]: DEBUG nova.scheduler.client.report [None req-947b662c-414a-42b6-bd23-f5e60b455b8c tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2181.055543] env[62824]: DEBUG oslo_vmware.api [None req-0235e5cb-af08-4951-9315-366312405d6f tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Waiting for the task: (returnval){ [ 2181.055543] env[62824]: value = "task-2146518" [ 2181.055543] env[62824]: _type = "Task" [ 2181.055543] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2181.066217] env[62824]: DEBUG oslo_vmware.api [None req-0235e5cb-af08-4951-9315-366312405d6f tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Task: {'id': task-2146518, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2181.248684] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146511, 'name': CreateVM_Task, 'duration_secs': 1.401476} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2181.248996] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 58243465-ba24-4b37-a640-04ab24963230] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2181.249872] env[62824]: DEBUG oslo_concurrency.lockutils [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2181.250171] env[62824]: DEBUG oslo_concurrency.lockutils [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2181.250449] env[62824]: DEBUG oslo_concurrency.lockutils [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2181.250743] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-94e58207-d267-4178-a097-c8ab96b5ec4f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2181.256549] env[62824]: DEBUG oslo_vmware.api [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 2181.256549] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]525492a8-5574-6c1d-595c-50f6985dd672" [ 2181.256549] env[62824]: _type = "Task" [ 2181.256549] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2181.269899] env[62824]: DEBUG oslo_vmware.api [None req-0c3b6169-674a-4946-8b27-bab89f3484d0 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Task: {'id': task-2146512, 'name': PowerOffVM_Task, 'duration_secs': 0.974231} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2181.273421] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c3b6169-674a-4946-8b27-bab89f3484d0 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: d067fa97-cedc-4e3d-9be4-d860a79a7723] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2181.273615] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-0c3b6169-674a-4946-8b27-bab89f3484d0 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: d067fa97-cedc-4e3d-9be4-d860a79a7723] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2181.273879] env[62824]: DEBUG oslo_vmware.api [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]525492a8-5574-6c1d-595c-50f6985dd672, 'name': SearchDatastore_Task, 'duration_secs': 0.010805} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2181.274096] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4ad18f58-1b23-4c78-aa17-9ca03b207226 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2181.275666] env[62824]: DEBUG oslo_concurrency.lockutils [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2181.275929] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 58243465-ba24-4b37-a640-04ab24963230] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2181.276186] env[62824]: DEBUG oslo_concurrency.lockutils [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2181.276333] env[62824]: DEBUG oslo_concurrency.lockutils [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2181.276511] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2181.276744] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-51cbbf6c-cfb5-44f8-8785-4439a17b669d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2181.285894] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2181.286114] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2181.286836] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f28d6dba-78e1-4adc-9026-3706de59188b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2181.292417] env[62824]: DEBUG oslo_vmware.api [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 2181.292417] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52bc47b8-59d2-42f8-417e-58cb9c8c8e1a" [ 2181.292417] env[62824]: _type = "Task" [ 2181.292417] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2181.300844] env[62824]: DEBUG oslo_vmware.api [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52bc47b8-59d2-42f8-417e-58cb9c8c8e1a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2181.433970] env[62824]: DEBUG oslo_vmware.api [None req-0c0f7a2a-4ac1-4123-9e01-b25a8857edf5 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Task: {'id': task-2146515, 'name': ReconfigVM_Task, 'duration_secs': 0.328996} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2181.434497] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-0c0f7a2a-4ac1-4123-9e01-b25a8857edf5 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Reconfigured VM instance instance-00000052 to attach disk [datastore1] da46d1d5-f75f-4ef1-b571-fbebab89c2a6/9e2a7d30-212d-4ab8-9606-c5c6d52629e8-rescue.vmdk or device None with type thin {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2181.437607] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6acc4a1-e74d-431a-b157-3e7bf9f448ac {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2181.466733] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c7c9a4de-94cd-4793-8aa3-ec28c0bd90af {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2181.483809] env[62824]: DEBUG oslo_vmware.api [None req-0c0f7a2a-4ac1-4123-9e01-b25a8857edf5 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Waiting for the task: (returnval){ [ 2181.483809] env[62824]: value = "task-2146520" [ 2181.483809] env[62824]: _type = "Task" [ 2181.483809] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2181.495375] env[62824]: DEBUG oslo_vmware.api [None req-0c0f7a2a-4ac1-4123-9e01-b25a8857edf5 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Task: {'id': task-2146520, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2181.499320] env[62824]: DEBUG oslo_vmware.api [None req-cac6081d-dcbb-4024-bf06-8c5d7fd6c54c tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146517, 'name': PowerOffVM_Task, 'duration_secs': 0.244548} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2181.499600] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-cac6081d-dcbb-4024-bf06-8c5d7fd6c54c tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: da1ad739-b252-4e29-a22a-ab3bcab173ec] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2181.499768] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-cac6081d-dcbb-4024-bf06-8c5d7fd6c54c tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: da1ad739-b252-4e29-a22a-ab3bcab173ec] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2181.500323] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0dcf2711-2cce-439c-80c9-08725c41aaa8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2181.538592] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-0c3b6169-674a-4946-8b27-bab89f3484d0 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: d067fa97-cedc-4e3d-9be4-d860a79a7723] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2181.538820] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-0c3b6169-674a-4946-8b27-bab89f3484d0 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: d067fa97-cedc-4e3d-9be4-d860a79a7723] Deleting contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2181.539010] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-0c3b6169-674a-4946-8b27-bab89f3484d0 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Deleting the datastore file [datastore2] d067fa97-cedc-4e3d-9be4-d860a79a7723 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2181.539310] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f0b203c7-a032-43a2-bd11-ed225e718d3b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2181.548529] env[62824]: DEBUG oslo_vmware.api [None req-0c3b6169-674a-4946-8b27-bab89f3484d0 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Waiting for the task: (returnval){ [ 2181.548529] env[62824]: value = "task-2146522" [ 2181.548529] env[62824]: _type = "Task" [ 2181.548529] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2181.558132] env[62824]: DEBUG oslo_vmware.api [None req-0c3b6169-674a-4946-8b27-bab89f3484d0 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Task: {'id': task-2146522, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2181.562502] env[62824]: DEBUG oslo_concurrency.lockutils [None req-947b662c-414a-42b6-bd23-f5e60b455b8c tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.174s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2181.564709] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 23.613s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2181.564889] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2181.565076] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62824) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 2181.565363] env[62824]: DEBUG oslo_concurrency.lockutils [None req-aefa2097-549e-43b5-9152-0e8ce3b5da7a tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.537s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2181.565656] env[62824]: DEBUG nova.objects.instance [None req-aefa2097-549e-43b5-9152-0e8ce3b5da7a tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Lazy-loading 'resources' on Instance uuid 7a2a51c4-e558-4f5f-b82c-718bc12c1df5 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2181.570021] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c622f2e7-2c0a-487d-8845-497cb1788a89 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2181.577340] env[62824]: DEBUG oslo_vmware.api [None req-0235e5cb-af08-4951-9315-366312405d6f tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Task: {'id': task-2146518, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.141937} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2181.577941] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-0235e5cb-af08-4951-9315-366312405d6f tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2181.578248] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-0235e5cb-af08-4951-9315-366312405d6f tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 4c2c2068-664d-404f-a99e-8fc7719f43e5] Deleted contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2181.578631] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-0235e5cb-af08-4951-9315-366312405d6f tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 4c2c2068-664d-404f-a99e-8fc7719f43e5] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2181.578824] env[62824]: INFO nova.compute.manager [None req-0235e5cb-af08-4951-9315-366312405d6f tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: 4c2c2068-664d-404f-a99e-8fc7719f43e5] Took 1.24 seconds to destroy the instance on the hypervisor. [ 2181.579095] env[62824]: DEBUG oslo.service.loopingcall [None req-0235e5cb-af08-4951-9315-366312405d6f tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2181.579582] env[62824]: DEBUG nova.compute.manager [-] [instance: 4c2c2068-664d-404f-a99e-8fc7719f43e5] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2181.579700] env[62824]: DEBUG nova.network.neutron [-] [instance: 4c2c2068-664d-404f-a99e-8fc7719f43e5] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2181.585057] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e96db6a-8a5c-4ba0-9d7f-eb6cef1885d8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2181.591548] env[62824]: INFO nova.scheduler.client.report [None req-947b662c-414a-42b6-bd23-f5e60b455b8c tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Deleted allocations for instance 02405b17-7022-4e2c-a357-415de88f63d0 [ 2181.604452] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e8e5a55-9aa4-4ad1-a6e5-8dcdcd39a1d2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2181.612609] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86563a3a-161f-410c-b955-81a9f510601a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2181.652131] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179000MB free_disk=174GB free_vcpus=48 pci_devices=None {{(pid=62824) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 2181.652332] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2181.735829] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-cac6081d-dcbb-4024-bf06-8c5d7fd6c54c tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: da1ad739-b252-4e29-a22a-ab3bcab173ec] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2181.736115] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-cac6081d-dcbb-4024-bf06-8c5d7fd6c54c tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: da1ad739-b252-4e29-a22a-ab3bcab173ec] Deleting contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2181.736311] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-cac6081d-dcbb-4024-bf06-8c5d7fd6c54c tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Deleting the datastore file [datastore1] da1ad739-b252-4e29-a22a-ab3bcab173ec {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2181.736619] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ca511b7e-cf77-40e5-b0ce-503ef8130236 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2181.744547] env[62824]: DEBUG oslo_vmware.api [None req-cac6081d-dcbb-4024-bf06-8c5d7fd6c54c tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the task: (returnval){ [ 2181.744547] env[62824]: value = "task-2146523" [ 2181.744547] env[62824]: _type = "Task" [ 2181.744547] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2181.752843] env[62824]: DEBUG oslo_vmware.api [None req-cac6081d-dcbb-4024-bf06-8c5d7fd6c54c tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146523, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2181.807055] env[62824]: DEBUG oslo_vmware.api [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52bc47b8-59d2-42f8-417e-58cb9c8c8e1a, 'name': SearchDatastore_Task, 'duration_secs': 0.009419} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2181.807938] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-11dfdd5f-6c10-43e4-b2b1-982cfad3e5b0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2181.814839] env[62824]: DEBUG oslo_vmware.api [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 2181.814839] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]525f46f0-346f-45e5-09ad-ac7432fce2c1" [ 2181.814839] env[62824]: _type = "Task" [ 2181.814839] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2181.823603] env[62824]: DEBUG oslo_vmware.api [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]525f46f0-346f-45e5-09ad-ac7432fce2c1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2181.997362] env[62824]: DEBUG oslo_vmware.api [None req-0c0f7a2a-4ac1-4123-9e01-b25a8857edf5 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Task: {'id': task-2146520, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2182.059580] env[62824]: DEBUG oslo_vmware.api [None req-0c3b6169-674a-4946-8b27-bab89f3484d0 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Task: {'id': task-2146522, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.146203} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2182.059883] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-0c3b6169-674a-4946-8b27-bab89f3484d0 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2182.060073] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-0c3b6169-674a-4946-8b27-bab89f3484d0 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: d067fa97-cedc-4e3d-9be4-d860a79a7723] Deleted contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2182.060316] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-0c3b6169-674a-4946-8b27-bab89f3484d0 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: d067fa97-cedc-4e3d-9be4-d860a79a7723] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2182.060517] env[62824]: INFO nova.compute.manager [None req-0c3b6169-674a-4946-8b27-bab89f3484d0 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] [instance: d067fa97-cedc-4e3d-9be4-d860a79a7723] Took 1.83 seconds to destroy the instance on the hypervisor. [ 2182.060783] env[62824]: DEBUG oslo.service.loopingcall [None req-0c3b6169-674a-4946-8b27-bab89f3484d0 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2182.061410] env[62824]: DEBUG nova.compute.manager [-] [instance: d067fa97-cedc-4e3d-9be4-d860a79a7723] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2182.061532] env[62824]: DEBUG nova.network.neutron [-] [instance: d067fa97-cedc-4e3d-9be4-d860a79a7723] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2182.084167] env[62824]: DEBUG nova.compute.manager [req-fb0f114b-7718-4642-bbce-5d3b0009ba46 req-f1c3db5f-a304-45f0-a87d-69350309516f service nova] [instance: 4c2c2068-664d-404f-a99e-8fc7719f43e5] Received event network-vif-deleted-96a12f9d-8a1e-4df8-b5fa-24c8b92c413f {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2182.084456] env[62824]: INFO nova.compute.manager [req-fb0f114b-7718-4642-bbce-5d3b0009ba46 req-f1c3db5f-a304-45f0-a87d-69350309516f service nova] [instance: 4c2c2068-664d-404f-a99e-8fc7719f43e5] Neutron deleted interface 96a12f9d-8a1e-4df8-b5fa-24c8b92c413f; detaching it from the instance and deleting it from the info cache [ 2182.084578] env[62824]: DEBUG nova.network.neutron [req-fb0f114b-7718-4642-bbce-5d3b0009ba46 req-f1c3db5f-a304-45f0-a87d-69350309516f service nova] [instance: 4c2c2068-664d-404f-a99e-8fc7719f43e5] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2182.112995] env[62824]: DEBUG oslo_concurrency.lockutils [None req-947b662c-414a-42b6-bd23-f5e60b455b8c tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Lock "02405b17-7022-4e2c-a357-415de88f63d0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.432s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2182.257318] env[62824]: DEBUG oslo_vmware.api [None req-cac6081d-dcbb-4024-bf06-8c5d7fd6c54c tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146523, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.150975} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2182.257713] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-cac6081d-dcbb-4024-bf06-8c5d7fd6c54c tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2182.258028] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-cac6081d-dcbb-4024-bf06-8c5d7fd6c54c tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: da1ad739-b252-4e29-a22a-ab3bcab173ec] Deleted contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2182.258259] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-cac6081d-dcbb-4024-bf06-8c5d7fd6c54c tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: da1ad739-b252-4e29-a22a-ab3bcab173ec] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2182.258507] env[62824]: INFO nova.compute.manager [None req-cac6081d-dcbb-4024-bf06-8c5d7fd6c54c tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: da1ad739-b252-4e29-a22a-ab3bcab173ec] Took 1.29 seconds to destroy the instance on the hypervisor. [ 2182.258810] env[62824]: DEBUG oslo.service.loopingcall [None req-cac6081d-dcbb-4024-bf06-8c5d7fd6c54c tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2182.259072] env[62824]: DEBUG nova.compute.manager [-] [instance: da1ad739-b252-4e29-a22a-ab3bcab173ec] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2182.259225] env[62824]: DEBUG nova.network.neutron [-] [instance: da1ad739-b252-4e29-a22a-ab3bcab173ec] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2182.327258] env[62824]: DEBUG oslo_vmware.api [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]525f46f0-346f-45e5-09ad-ac7432fce2c1, 'name': SearchDatastore_Task, 'duration_secs': 0.021895} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2182.327617] env[62824]: DEBUG oslo_concurrency.lockutils [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2182.328121] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 58243465-ba24-4b37-a640-04ab24963230/58243465-ba24-4b37-a640-04ab24963230.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2182.328465] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-536b6924-c425-4a4b-a354-f8d9e0fa87f4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2182.336569] env[62824]: DEBUG oslo_vmware.api [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 2182.336569] env[62824]: value = "task-2146524" [ 2182.336569] env[62824]: _type = "Task" [ 2182.336569] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2182.345713] env[62824]: DEBUG oslo_vmware.api [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146524, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2182.490705] env[62824]: DEBUG nova.network.neutron [-] [instance: 4c2c2068-664d-404f-a99e-8fc7719f43e5] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2182.497108] env[62824]: DEBUG oslo_vmware.api [None req-0c0f7a2a-4ac1-4123-9e01-b25a8857edf5 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Task: {'id': task-2146520, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2182.501510] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e772be76-8e5e-4ac4-82d4-aa1f8c545fa6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2182.508914] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f65a6d11-e484-4950-bc44-aa90f1d52c12 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2182.547395] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89b728b1-0eea-4635-85a7-91f0e2a9850e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2182.557416] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfd6a11f-91b0-4caf-99e7-65ace0baab0b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2182.573543] env[62824]: DEBUG nova.compute.provider_tree [None req-aefa2097-549e-43b5-9152-0e8ce3b5da7a tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2182.589340] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-db4ddbc9-60ed-4d47-8906-4df0458af54a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2182.598279] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df1b42ca-db5a-4df0-87d4-c10dddfb34c0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2182.646808] env[62824]: DEBUG nova.compute.manager [req-fb0f114b-7718-4642-bbce-5d3b0009ba46 req-f1c3db5f-a304-45f0-a87d-69350309516f service nova] [instance: 4c2c2068-664d-404f-a99e-8fc7719f43e5] Detach interface failed, port_id=96a12f9d-8a1e-4df8-b5fa-24c8b92c413f, reason: Instance 4c2c2068-664d-404f-a99e-8fc7719f43e5 could not be found. {{(pid=62824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11482}} [ 2182.664757] env[62824]: DEBUG oslo_concurrency.lockutils [None req-32333306-ced2-48b5-bba6-b899d39c55fe tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Acquiring lock "f3740d9a-b21a-4a79-9e28-2a89ecd08bb6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2182.665700] env[62824]: DEBUG oslo_concurrency.lockutils [None req-32333306-ced2-48b5-bba6-b899d39c55fe tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Lock "f3740d9a-b21a-4a79-9e28-2a89ecd08bb6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2182.665700] env[62824]: DEBUG oslo_concurrency.lockutils [None req-32333306-ced2-48b5-bba6-b899d39c55fe tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Acquiring lock "f3740d9a-b21a-4a79-9e28-2a89ecd08bb6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2182.665700] env[62824]: DEBUG oslo_concurrency.lockutils [None req-32333306-ced2-48b5-bba6-b899d39c55fe tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Lock "f3740d9a-b21a-4a79-9e28-2a89ecd08bb6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2182.665873] env[62824]: DEBUG oslo_concurrency.lockutils [None req-32333306-ced2-48b5-bba6-b899d39c55fe tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Lock "f3740d9a-b21a-4a79-9e28-2a89ecd08bb6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2182.668551] env[62824]: INFO nova.compute.manager [None req-32333306-ced2-48b5-bba6-b899d39c55fe tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: f3740d9a-b21a-4a79-9e28-2a89ecd08bb6] Terminating instance [ 2182.697436] env[62824]: DEBUG nova.compute.manager [req-eb29ec0a-8104-4ba2-b466-94603bff9280 req-5867f7b8-7272-47e1-8009-b63633166a05 service nova] [instance: da1ad739-b252-4e29-a22a-ab3bcab173ec] Received event network-vif-deleted-734a463a-d647-4c05-b8a9-29a3dd47e376 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2182.697436] env[62824]: INFO nova.compute.manager [req-eb29ec0a-8104-4ba2-b466-94603bff9280 req-5867f7b8-7272-47e1-8009-b63633166a05 service nova] [instance: da1ad739-b252-4e29-a22a-ab3bcab173ec] Neutron deleted interface 734a463a-d647-4c05-b8a9-29a3dd47e376; detaching it from the instance and deleting it from the info cache [ 2182.697662] env[62824]: DEBUG nova.network.neutron [req-eb29ec0a-8104-4ba2-b466-94603bff9280 req-5867f7b8-7272-47e1-8009-b63633166a05 service nova] [instance: da1ad739-b252-4e29-a22a-ab3bcab173ec] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2182.849345] env[62824]: DEBUG oslo_vmware.api [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146524, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.469122} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2182.849880] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 58243465-ba24-4b37-a640-04ab24963230/58243465-ba24-4b37-a640-04ab24963230.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2182.850194] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 58243465-ba24-4b37-a640-04ab24963230] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2182.850402] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b2906b16-eb58-4868-853b-f905f145a606 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2182.858027] env[62824]: DEBUG oslo_vmware.api [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 2182.858027] env[62824]: value = "task-2146525" [ 2182.858027] env[62824]: _type = "Task" [ 2182.858027] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2182.868035] env[62824]: DEBUG oslo_vmware.api [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146525, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2182.870460] env[62824]: DEBUG nova.network.neutron [-] [instance: d067fa97-cedc-4e3d-9be4-d860a79a7723] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2182.995018] env[62824]: INFO nova.compute.manager [-] [instance: 4c2c2068-664d-404f-a99e-8fc7719f43e5] Took 1.42 seconds to deallocate network for instance. [ 2182.995388] env[62824]: DEBUG oslo_vmware.api [None req-0c0f7a2a-4ac1-4123-9e01-b25a8857edf5 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Task: {'id': task-2146520, 'name': ReconfigVM_Task, 'duration_secs': 1.439828} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2182.997050] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c0f7a2a-4ac1-4123-9e01-b25a8857edf5 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2182.999852] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8aa93c2f-81d4-4a80-bf98-400564424268 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2183.014897] env[62824]: DEBUG oslo_vmware.api [None req-0c0f7a2a-4ac1-4123-9e01-b25a8857edf5 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Waiting for the task: (returnval){ [ 2183.014897] env[62824]: value = "task-2146526" [ 2183.014897] env[62824]: _type = "Task" [ 2183.014897] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2183.024699] env[62824]: DEBUG oslo_vmware.api [None req-0c0f7a2a-4ac1-4123-9e01-b25a8857edf5 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Task: {'id': task-2146526, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2183.076773] env[62824]: DEBUG nova.scheduler.client.report [None req-aefa2097-549e-43b5-9152-0e8ce3b5da7a tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2183.152217] env[62824]: DEBUG nova.network.neutron [-] [instance: da1ad739-b252-4e29-a22a-ab3bcab173ec] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2183.173944] env[62824]: DEBUG nova.compute.manager [None req-32333306-ced2-48b5-bba6-b899d39c55fe tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: f3740d9a-b21a-4a79-9e28-2a89ecd08bb6] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2183.174191] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-32333306-ced2-48b5-bba6-b899d39c55fe tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: f3740d9a-b21a-4a79-9e28-2a89ecd08bb6] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2183.175066] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c97336ef-f88b-4945-8d90-93430c62eb25 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2183.183103] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-32333306-ced2-48b5-bba6-b899d39c55fe tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: f3740d9a-b21a-4a79-9e28-2a89ecd08bb6] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2183.183349] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-651cb754-4104-4a32-aace-b1354cca6bd7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2183.189465] env[62824]: DEBUG oslo_vmware.api [None req-32333306-ced2-48b5-bba6-b899d39c55fe tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for the task: (returnval){ [ 2183.189465] env[62824]: value = "task-2146527" [ 2183.189465] env[62824]: _type = "Task" [ 2183.189465] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2183.198670] env[62824]: DEBUG oslo_vmware.api [None req-32333306-ced2-48b5-bba6-b899d39c55fe tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146527, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2183.201351] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5fbb2d7d-3c37-487e-a37f-c1c5ee9d9eb1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2183.213232] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-197e4fb5-5146-4cd4-8786-120a0859a28f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2183.249705] env[62824]: DEBUG nova.compute.manager [req-eb29ec0a-8104-4ba2-b466-94603bff9280 req-5867f7b8-7272-47e1-8009-b63633166a05 service nova] [instance: da1ad739-b252-4e29-a22a-ab3bcab173ec] Detach interface failed, port_id=734a463a-d647-4c05-b8a9-29a3dd47e376, reason: Instance da1ad739-b252-4e29-a22a-ab3bcab173ec could not be found. {{(pid=62824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11482}} [ 2183.368225] env[62824]: DEBUG oslo_vmware.api [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146525, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072877} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2183.368518] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 58243465-ba24-4b37-a640-04ab24963230] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2183.369806] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e46cec61-6853-4b03-acb6-b3891f42e9e6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2183.372818] env[62824]: INFO nova.compute.manager [-] [instance: d067fa97-cedc-4e3d-9be4-d860a79a7723] Took 1.31 seconds to deallocate network for instance. [ 2183.403875] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 58243465-ba24-4b37-a640-04ab24963230] Reconfiguring VM instance instance-00000053 to attach disk [datastore2] 58243465-ba24-4b37-a640-04ab24963230/58243465-ba24-4b37-a640-04ab24963230.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2183.404270] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0fa93518-6579-4dfc-84a2-55922e9d3f60 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2183.429533] env[62824]: DEBUG oslo_vmware.api [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 2183.429533] env[62824]: value = "task-2146528" [ 2183.429533] env[62824]: _type = "Task" [ 2183.429533] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2183.439743] env[62824]: DEBUG oslo_vmware.api [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146528, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2183.504387] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0235e5cb-af08-4951-9315-366312405d6f tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2183.525246] env[62824]: DEBUG oslo_vmware.api [None req-0c0f7a2a-4ac1-4123-9e01-b25a8857edf5 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Task: {'id': task-2146526, 'name': PowerOnVM_Task, 'duration_secs': 0.420336} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2183.525524] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c0f7a2a-4ac1-4123-9e01-b25a8857edf5 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2183.528367] env[62824]: DEBUG nova.compute.manager [None req-0c0f7a2a-4ac1-4123-9e01-b25a8857edf5 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2183.529187] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c812a7a-709e-4c64-a58d-22ca6484d385 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2183.582333] env[62824]: DEBUG oslo_concurrency.lockutils [None req-aefa2097-549e-43b5-9152-0e8ce3b5da7a tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.017s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2183.585046] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5fb295c4-ec07-4626-81a5-775a8bd81d25 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.184s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2183.585046] env[62824]: DEBUG nova.objects.instance [None req-5fb295c4-ec07-4626-81a5-775a8bd81d25 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Lazy-loading 'resources' on Instance uuid 20f93b46-5e7e-4256-8cc1-e0a1b16740d5 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2183.607847] env[62824]: INFO nova.scheduler.client.report [None req-aefa2097-549e-43b5-9152-0e8ce3b5da7a tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Deleted allocations for instance 7a2a51c4-e558-4f5f-b82c-718bc12c1df5 [ 2183.654249] env[62824]: INFO nova.compute.manager [-] [instance: da1ad739-b252-4e29-a22a-ab3bcab173ec] Took 1.39 seconds to deallocate network for instance. [ 2183.699043] env[62824]: DEBUG oslo_vmware.api [None req-32333306-ced2-48b5-bba6-b899d39c55fe tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146527, 'name': PowerOffVM_Task, 'duration_secs': 0.173011} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2183.699325] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-32333306-ced2-48b5-bba6-b899d39c55fe tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: f3740d9a-b21a-4a79-9e28-2a89ecd08bb6] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2183.699513] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-32333306-ced2-48b5-bba6-b899d39c55fe tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: f3740d9a-b21a-4a79-9e28-2a89ecd08bb6] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2183.699800] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-308d99cd-dae1-4d64-ac8e-9de6cef868e4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2183.879516] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0c3b6169-674a-4946-8b27-bab89f3484d0 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2183.938803] env[62824]: DEBUG oslo_vmware.api [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146528, 'name': ReconfigVM_Task, 'duration_secs': 0.374244} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2183.939110] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 58243465-ba24-4b37-a640-04ab24963230] Reconfigured VM instance instance-00000053 to attach disk [datastore2] 58243465-ba24-4b37-a640-04ab24963230/58243465-ba24-4b37-a640-04ab24963230.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2183.939757] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-aa4e34de-bccf-46a4-adc2-13f0660f28d4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2183.945931] env[62824]: DEBUG oslo_vmware.api [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 2183.945931] env[62824]: value = "task-2146530" [ 2183.945931] env[62824]: _type = "Task" [ 2183.945931] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2183.953777] env[62824]: DEBUG oslo_vmware.api [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146530, 'name': Rename_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2184.059589] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-32333306-ced2-48b5-bba6-b899d39c55fe tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: f3740d9a-b21a-4a79-9e28-2a89ecd08bb6] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2184.059843] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-32333306-ced2-48b5-bba6-b899d39c55fe tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: f3740d9a-b21a-4a79-9e28-2a89ecd08bb6] Deleting contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2184.060992] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-32333306-ced2-48b5-bba6-b899d39c55fe tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Deleting the datastore file [datastore1] f3740d9a-b21a-4a79-9e28-2a89ecd08bb6 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2184.060992] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0703bfc5-0bcd-46b4-ba77-5473fb66a52c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2184.067407] env[62824]: DEBUG oslo_vmware.api [None req-32333306-ced2-48b5-bba6-b899d39c55fe tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for the task: (returnval){ [ 2184.067407] env[62824]: value = "task-2146531" [ 2184.067407] env[62824]: _type = "Task" [ 2184.067407] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2184.075228] env[62824]: DEBUG oslo_vmware.api [None req-32333306-ced2-48b5-bba6-b899d39c55fe tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146531, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2184.115024] env[62824]: DEBUG oslo_concurrency.lockutils [None req-aefa2097-549e-43b5-9152-0e8ce3b5da7a tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Lock "7a2a51c4-e558-4f5f-b82c-718bc12c1df5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.578s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2184.137756] env[62824]: DEBUG nova.compute.manager [req-295b14b9-ca2a-4a8f-98bd-872855688bdb req-cd8547d0-b3e0-49c0-ac5e-3e653e32d080 service nova] [instance: d067fa97-cedc-4e3d-9be4-d860a79a7723] Received event network-vif-deleted-2b52b217-d85d-4239-a02a-d97e3c82d93d {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2184.160187] env[62824]: DEBUG oslo_concurrency.lockutils [None req-cac6081d-dcbb-4024-bf06-8c5d7fd6c54c tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2184.411206] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1a3d6be-bb71-44d2-b627-91ebc2e6c5ff {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2184.419576] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c64045e3-6e52-4e07-830d-6e7a3abdef14 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2184.454162] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-607b4225-953f-4cbe-8989-7a84c20782f9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2184.464832] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-543b72e6-860f-4a6c-a2ca-c9a1b77f87fc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2184.468490] env[62824]: DEBUG oslo_vmware.api [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146530, 'name': Rename_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2184.478295] env[62824]: DEBUG nova.compute.provider_tree [None req-5fb295c4-ec07-4626-81a5-775a8bd81d25 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2184.578552] env[62824]: DEBUG oslo_vmware.api [None req-32333306-ced2-48b5-bba6-b899d39c55fe tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146531, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.145055} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2184.579057] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-32333306-ced2-48b5-bba6-b899d39c55fe tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2184.579436] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-32333306-ced2-48b5-bba6-b899d39c55fe tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: f3740d9a-b21a-4a79-9e28-2a89ecd08bb6] Deleted contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2184.579752] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-32333306-ced2-48b5-bba6-b899d39c55fe tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: f3740d9a-b21a-4a79-9e28-2a89ecd08bb6] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2184.580266] env[62824]: INFO nova.compute.manager [None req-32333306-ced2-48b5-bba6-b899d39c55fe tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: f3740d9a-b21a-4a79-9e28-2a89ecd08bb6] Took 1.41 seconds to destroy the instance on the hypervisor. [ 2184.580832] env[62824]: DEBUG oslo.service.loopingcall [None req-32333306-ced2-48b5-bba6-b899d39c55fe tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2184.581162] env[62824]: DEBUG nova.compute.manager [-] [instance: f3740d9a-b21a-4a79-9e28-2a89ecd08bb6] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2184.581371] env[62824]: DEBUG nova.network.neutron [-] [instance: f3740d9a-b21a-4a79-9e28-2a89ecd08bb6] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2184.845376] env[62824]: DEBUG nova.compute.manager [req-320f29f0-3481-4e96-aab2-68d28fad4adb req-0daed896-6cbd-4ad8-bfbd-c014a4012d3e service nova] [instance: f3740d9a-b21a-4a79-9e28-2a89ecd08bb6] Received event network-vif-deleted-f2fb594d-5cb0-4c90-9787-3a362864f81a {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2184.846033] env[62824]: INFO nova.compute.manager [req-320f29f0-3481-4e96-aab2-68d28fad4adb req-0daed896-6cbd-4ad8-bfbd-c014a4012d3e service nova] [instance: f3740d9a-b21a-4a79-9e28-2a89ecd08bb6] Neutron deleted interface f2fb594d-5cb0-4c90-9787-3a362864f81a; detaching it from the instance and deleting it from the info cache [ 2184.846033] env[62824]: DEBUG nova.network.neutron [req-320f29f0-3481-4e96-aab2-68d28fad4adb req-0daed896-6cbd-4ad8-bfbd-c014a4012d3e service nova] [instance: f3740d9a-b21a-4a79-9e28-2a89ecd08bb6] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2184.960540] env[62824]: DEBUG oslo_vmware.api [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146530, 'name': Rename_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2184.982165] env[62824]: DEBUG nova.scheduler.client.report [None req-5fb295c4-ec07-4626-81a5-775a8bd81d25 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2185.321333] env[62824]: DEBUG nova.network.neutron [-] [instance: f3740d9a-b21a-4a79-9e28-2a89ecd08bb6] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2185.350076] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-905e9483-7514-45d4-8120-d10502bfc806 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2185.361700] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35e962ec-73e8-4e28-b928-335874093587 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2185.394954] env[62824]: DEBUG nova.compute.manager [req-320f29f0-3481-4e96-aab2-68d28fad4adb req-0daed896-6cbd-4ad8-bfbd-c014a4012d3e service nova] [instance: f3740d9a-b21a-4a79-9e28-2a89ecd08bb6] Detach interface failed, port_id=f2fb594d-5cb0-4c90-9787-3a362864f81a, reason: Instance f3740d9a-b21a-4a79-9e28-2a89ecd08bb6 could not be found. {{(pid=62824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11482}} [ 2185.461582] env[62824]: DEBUG oslo_vmware.api [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146530, 'name': Rename_Task, 'duration_secs': 1.13553} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2185.461795] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 58243465-ba24-4b37-a640-04ab24963230] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2185.461979] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ce2bc543-0639-4b2d-bc4d-28ae59047496 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2185.468829] env[62824]: DEBUG oslo_vmware.api [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 2185.468829] env[62824]: value = "task-2146532" [ 2185.468829] env[62824]: _type = "Task" [ 2185.468829] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2185.477167] env[62824]: DEBUG oslo_vmware.api [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146532, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2185.491250] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5fb295c4-ec07-4626-81a5-775a8bd81d25 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.907s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2185.494163] env[62824]: DEBUG oslo_concurrency.lockutils [None req-07015f60-1e88-49e8-8469-e207c9f08bc2 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.635s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2185.494163] env[62824]: DEBUG nova.objects.instance [None req-07015f60-1e88-49e8-8469-e207c9f08bc2 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Lazy-loading 'resources' on Instance uuid b0456b8a-348a-4503-a92c-58e5ab455d1c {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2185.522708] env[62824]: INFO nova.scheduler.client.report [None req-5fb295c4-ec07-4626-81a5-775a8bd81d25 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Deleted allocations for instance 20f93b46-5e7e-4256-8cc1-e0a1b16740d5 [ 2185.823671] env[62824]: INFO nova.compute.manager [-] [instance: f3740d9a-b21a-4a79-9e28-2a89ecd08bb6] Took 1.24 seconds to deallocate network for instance. [ 2185.979523] env[62824]: DEBUG oslo_vmware.api [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146532, 'name': PowerOnVM_Task, 'duration_secs': 0.471593} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2185.979827] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 58243465-ba24-4b37-a640-04ab24963230] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2185.980049] env[62824]: INFO nova.compute.manager [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 58243465-ba24-4b37-a640-04ab24963230] Took 8.68 seconds to spawn the instance on the hypervisor. [ 2185.980255] env[62824]: DEBUG nova.compute.manager [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 58243465-ba24-4b37-a640-04ab24963230] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2185.981073] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2748931c-47c4-4184-bd13-09ca559a497e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2186.033213] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5fb295c4-ec07-4626-81a5-775a8bd81d25 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Lock "20f93b46-5e7e-4256-8cc1-e0a1b16740d5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.363s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2186.305458] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2143cca-3a30-44e6-8b20-5109d0156f6c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2186.314106] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1366904d-9564-4201-8a66-a92a9f9c7aab {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2186.346211] env[62824]: DEBUG oslo_concurrency.lockutils [None req-32333306-ced2-48b5-bba6-b899d39c55fe tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2186.347236] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93faee9c-963d-438d-93d3-d9433544a36d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2186.354644] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6576273-e909-4192-87a4-3b811e1af923 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2186.368212] env[62824]: DEBUG nova.compute.provider_tree [None req-07015f60-1e88-49e8-8469-e207c9f08bc2 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2186.498975] env[62824]: INFO nova.compute.manager [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 58243465-ba24-4b37-a640-04ab24963230] Took 38.73 seconds to build instance. [ 2186.711329] env[62824]: DEBUG nova.compute.manager [req-69b2dfed-8207-44bd-b4a6-baa41f4f94e1 req-0182ac1b-5648-4dd4-9282-6cb7fbdaf3d0 service nova] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Received event network-changed-0cefe7bc-5dd6-4f4c-872b-ef1b658a4d60 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2186.711564] env[62824]: DEBUG nova.compute.manager [req-69b2dfed-8207-44bd-b4a6-baa41f4f94e1 req-0182ac1b-5648-4dd4-9282-6cb7fbdaf3d0 service nova] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Refreshing instance network info cache due to event network-changed-0cefe7bc-5dd6-4f4c-872b-ef1b658a4d60. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2186.711766] env[62824]: DEBUG oslo_concurrency.lockutils [req-69b2dfed-8207-44bd-b4a6-baa41f4f94e1 req-0182ac1b-5648-4dd4-9282-6cb7fbdaf3d0 service nova] Acquiring lock "refresh_cache-da46d1d5-f75f-4ef1-b571-fbebab89c2a6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2186.712027] env[62824]: DEBUG oslo_concurrency.lockutils [req-69b2dfed-8207-44bd-b4a6-baa41f4f94e1 req-0182ac1b-5648-4dd4-9282-6cb7fbdaf3d0 service nova] Acquired lock "refresh_cache-da46d1d5-f75f-4ef1-b571-fbebab89c2a6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2186.712098] env[62824]: DEBUG nova.network.neutron [req-69b2dfed-8207-44bd-b4a6-baa41f4f94e1 req-0182ac1b-5648-4dd4-9282-6cb7fbdaf3d0 service nova] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Refreshing network info cache for port 0cefe7bc-5dd6-4f4c-872b-ef1b658a4d60 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2186.871665] env[62824]: DEBUG nova.scheduler.client.report [None req-07015f60-1e88-49e8-8469-e207c9f08bc2 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2187.000803] env[62824]: DEBUG oslo_concurrency.lockutils [None req-368e2e6c-2998-4fdb-9ac4-4e0a819b1139 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "58243465-ba24-4b37-a640-04ab24963230" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 51.778s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2187.119696] env[62824]: DEBUG oslo_concurrency.lockutils [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Acquiring lock "1f162a63-8773-4c34-9c84-b2de1d5e9f50" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2187.119934] env[62824]: DEBUG oslo_concurrency.lockutils [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Lock "1f162a63-8773-4c34-9c84-b2de1d5e9f50" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2187.149483] env[62824]: DEBUG oslo_concurrency.lockutils [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Acquiring lock "7c44d956-53d1-4a5f-9016-c0f3f0f348b4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2187.149781] env[62824]: DEBUG oslo_concurrency.lockutils [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Lock "7c44d956-53d1-4a5f-9016-c0f3f0f348b4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2187.377199] env[62824]: DEBUG oslo_concurrency.lockutils [None req-07015f60-1e88-49e8-8469-e207c9f08bc2 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.883s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2187.379669] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3a1b2b23-923e-4532-b132-fcd23342ce08 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.802s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2187.380531] env[62824]: DEBUG nova.objects.instance [None req-3a1b2b23-923e-4532-b132-fcd23342ce08 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Lazy-loading 'resources' on Instance uuid bd7f2ed5-fa08-4a48-9755-60b61791ea0b {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2187.401373] env[62824]: INFO nova.scheduler.client.report [None req-07015f60-1e88-49e8-8469-e207c9f08bc2 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Deleted allocations for instance b0456b8a-348a-4503-a92c-58e5ab455d1c [ 2187.447526] env[62824]: DEBUG nova.network.neutron [req-69b2dfed-8207-44bd-b4a6-baa41f4f94e1 req-0182ac1b-5648-4dd4-9282-6cb7fbdaf3d0 service nova] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Updated VIF entry in instance network info cache for port 0cefe7bc-5dd6-4f4c-872b-ef1b658a4d60. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2187.447944] env[62824]: DEBUG nova.network.neutron [req-69b2dfed-8207-44bd-b4a6-baa41f4f94e1 req-0182ac1b-5648-4dd4-9282-6cb7fbdaf3d0 service nova] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Updating instance_info_cache with network_info: [{"id": "0cefe7bc-5dd6-4f4c-872b-ef1b658a4d60", "address": "fa:16:3e:e0:51:b5", "network": {"id": "3210f2eb-2afb-4bce-bb59-f66e85888333", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-216881410-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.192", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "c8133feb4bde41e9898b3f7019c24d30", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d2bf584a-b4a3-4e7a-b0b7-eb8a2bc5a11d", "external-id": "nsx-vlan-transportzone-286", "segmentation_id": 286, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0cefe7bc-5d", "ovs_interfaceid": "0cefe7bc-5dd6-4f4c-872b-ef1b658a4d60", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2187.622504] env[62824]: DEBUG nova.compute.manager [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 1f162a63-8773-4c34-9c84-b2de1d5e9f50] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2187.652996] env[62824]: DEBUG nova.compute.manager [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7c44d956-53d1-4a5f-9016-c0f3f0f348b4] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2187.909737] env[62824]: DEBUG oslo_concurrency.lockutils [None req-07015f60-1e88-49e8-8469-e207c9f08bc2 tempest-ServerRescueTestJSON-1204463678 tempest-ServerRescueTestJSON-1204463678-project-member] Lock "b0456b8a-348a-4503-a92c-58e5ab455d1c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.717s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2187.950656] env[62824]: DEBUG oslo_concurrency.lockutils [req-69b2dfed-8207-44bd-b4a6-baa41f4f94e1 req-0182ac1b-5648-4dd4-9282-6cb7fbdaf3d0 service nova] Releasing lock "refresh_cache-da46d1d5-f75f-4ef1-b571-fbebab89c2a6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2187.950976] env[62824]: DEBUG nova.compute.manager [req-69b2dfed-8207-44bd-b4a6-baa41f4f94e1 req-0182ac1b-5648-4dd4-9282-6cb7fbdaf3d0 service nova] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Received event network-changed-0cefe7bc-5dd6-4f4c-872b-ef1b658a4d60 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2187.951249] env[62824]: DEBUG nova.compute.manager [req-69b2dfed-8207-44bd-b4a6-baa41f4f94e1 req-0182ac1b-5648-4dd4-9282-6cb7fbdaf3d0 service nova] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Refreshing instance network info cache due to event network-changed-0cefe7bc-5dd6-4f4c-872b-ef1b658a4d60. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2187.951502] env[62824]: DEBUG oslo_concurrency.lockutils [req-69b2dfed-8207-44bd-b4a6-baa41f4f94e1 req-0182ac1b-5648-4dd4-9282-6cb7fbdaf3d0 service nova] Acquiring lock "refresh_cache-da46d1d5-f75f-4ef1-b571-fbebab89c2a6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2187.951659] env[62824]: DEBUG oslo_concurrency.lockutils [req-69b2dfed-8207-44bd-b4a6-baa41f4f94e1 req-0182ac1b-5648-4dd4-9282-6cb7fbdaf3d0 service nova] Acquired lock "refresh_cache-da46d1d5-f75f-4ef1-b571-fbebab89c2a6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2187.951849] env[62824]: DEBUG nova.network.neutron [req-69b2dfed-8207-44bd-b4a6-baa41f4f94e1 req-0182ac1b-5648-4dd4-9282-6cb7fbdaf3d0 service nova] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Refreshing network info cache for port 0cefe7bc-5dd6-4f4c-872b-ef1b658a4d60 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2188.045941] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquiring lock "58243465-ba24-4b37-a640-04ab24963230" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2188.046220] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "58243465-ba24-4b37-a640-04ab24963230" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2188.046403] env[62824]: INFO nova.compute.manager [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 58243465-ba24-4b37-a640-04ab24963230] Shelving [ 2188.141520] env[62824]: DEBUG oslo_concurrency.lockutils [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2188.170779] env[62824]: DEBUG oslo_concurrency.lockutils [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2188.256202] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04c18998-54f8-4244-b312-c21131428ff6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2188.263778] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e6b9d5c-0565-48ab-b142-3d967148ca96 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2188.299408] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84e1809f-c8e7-4d87-8b33-bef6285b49bc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2188.307897] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f96f1643-2363-489e-8b8d-7c2af2e36273 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2188.324701] env[62824]: DEBUG nova.compute.provider_tree [None req-3a1b2b23-923e-4532-b132-fcd23342ce08 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2188.754013] env[62824]: DEBUG nova.compute.manager [req-8c9f8e1c-f5de-40d8-b6de-430cb2a7bbcc req-c6f2dcd9-0f18-4f8e-9734-5e3683ed908e service nova] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Received event network-changed-0cefe7bc-5dd6-4f4c-872b-ef1b658a4d60 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2188.754268] env[62824]: DEBUG nova.compute.manager [req-8c9f8e1c-f5de-40d8-b6de-430cb2a7bbcc req-c6f2dcd9-0f18-4f8e-9734-5e3683ed908e service nova] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Refreshing instance network info cache due to event network-changed-0cefe7bc-5dd6-4f4c-872b-ef1b658a4d60. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2188.754435] env[62824]: DEBUG oslo_concurrency.lockutils [req-8c9f8e1c-f5de-40d8-b6de-430cb2a7bbcc req-c6f2dcd9-0f18-4f8e-9734-5e3683ed908e service nova] Acquiring lock "refresh_cache-da46d1d5-f75f-4ef1-b571-fbebab89c2a6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2188.819519] env[62824]: DEBUG nova.network.neutron [req-69b2dfed-8207-44bd-b4a6-baa41f4f94e1 req-0182ac1b-5648-4dd4-9282-6cb7fbdaf3d0 service nova] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Updated VIF entry in instance network info cache for port 0cefe7bc-5dd6-4f4c-872b-ef1b658a4d60. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2188.819937] env[62824]: DEBUG nova.network.neutron [req-69b2dfed-8207-44bd-b4a6-baa41f4f94e1 req-0182ac1b-5648-4dd4-9282-6cb7fbdaf3d0 service nova] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Updating instance_info_cache with network_info: [{"id": "0cefe7bc-5dd6-4f4c-872b-ef1b658a4d60", "address": "fa:16:3e:e0:51:b5", "network": {"id": "3210f2eb-2afb-4bce-bb59-f66e85888333", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-216881410-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "c8133feb4bde41e9898b3f7019c24d30", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d2bf584a-b4a3-4e7a-b0b7-eb8a2bc5a11d", "external-id": "nsx-vlan-transportzone-286", "segmentation_id": 286, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0cefe7bc-5d", "ovs_interfaceid": "0cefe7bc-5dd6-4f4c-872b-ef1b658a4d60", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2188.828628] env[62824]: DEBUG nova.scheduler.client.report [None req-3a1b2b23-923e-4532-b132-fcd23342ce08 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2189.055684] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 58243465-ba24-4b37-a640-04ab24963230] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2189.056033] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c9f700de-02f4-46f7-8ee1-83557e191181 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2189.065255] env[62824]: DEBUG oslo_vmware.api [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 2189.065255] env[62824]: value = "task-2146533" [ 2189.065255] env[62824]: _type = "Task" [ 2189.065255] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2189.077395] env[62824]: DEBUG oslo_vmware.api [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146533, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2189.323465] env[62824]: DEBUG oslo_concurrency.lockutils [req-69b2dfed-8207-44bd-b4a6-baa41f4f94e1 req-0182ac1b-5648-4dd4-9282-6cb7fbdaf3d0 service nova] Releasing lock "refresh_cache-da46d1d5-f75f-4ef1-b571-fbebab89c2a6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2189.324596] env[62824]: DEBUG oslo_concurrency.lockutils [req-8c9f8e1c-f5de-40d8-b6de-430cb2a7bbcc req-c6f2dcd9-0f18-4f8e-9734-5e3683ed908e service nova] Acquired lock "refresh_cache-da46d1d5-f75f-4ef1-b571-fbebab89c2a6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2189.324596] env[62824]: DEBUG nova.network.neutron [req-8c9f8e1c-f5de-40d8-b6de-430cb2a7bbcc req-c6f2dcd9-0f18-4f8e-9734-5e3683ed908e service nova] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Refreshing network info cache for port 0cefe7bc-5dd6-4f4c-872b-ef1b658a4d60 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2189.340118] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3a1b2b23-923e-4532-b132-fcd23342ce08 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.959s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2189.341160] env[62824]: DEBUG oslo_concurrency.lockutils [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.273s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2189.345836] env[62824]: INFO nova.compute.claims [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 8c2ac679-6915-4149-aa51-99170b0df685] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2189.351087] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7e6f95cb-45a7-48de-8220-fb8355af6eb3 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Acquiring lock "da46d1d5-f75f-4ef1-b571-fbebab89c2a6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2189.351413] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7e6f95cb-45a7-48de-8220-fb8355af6eb3 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Lock "da46d1d5-f75f-4ef1-b571-fbebab89c2a6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2189.351740] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7e6f95cb-45a7-48de-8220-fb8355af6eb3 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Acquiring lock "da46d1d5-f75f-4ef1-b571-fbebab89c2a6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2189.351837] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7e6f95cb-45a7-48de-8220-fb8355af6eb3 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Lock "da46d1d5-f75f-4ef1-b571-fbebab89c2a6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2189.352092] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7e6f95cb-45a7-48de-8220-fb8355af6eb3 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Lock "da46d1d5-f75f-4ef1-b571-fbebab89c2a6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2189.354370] env[62824]: INFO nova.compute.manager [None req-7e6f95cb-45a7-48de-8220-fb8355af6eb3 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Terminating instance [ 2189.378173] env[62824]: INFO nova.scheduler.client.report [None req-3a1b2b23-923e-4532-b132-fcd23342ce08 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Deleted allocations for instance bd7f2ed5-fa08-4a48-9755-60b61791ea0b [ 2189.578474] env[62824]: DEBUG oslo_vmware.api [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146533, 'name': PowerOffVM_Task, 'duration_secs': 0.162308} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2189.578744] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 58243465-ba24-4b37-a640-04ab24963230] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2189.579621] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fdc84d5-0685-4c49-9b00-4287366b0e6e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2189.599097] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd58a381-7493-4896-892d-a56f19b269c8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2189.858847] env[62824]: DEBUG nova.compute.manager [None req-7e6f95cb-45a7-48de-8220-fb8355af6eb3 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2189.859229] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-7e6f95cb-45a7-48de-8220-fb8355af6eb3 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2189.860625] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91d8b5fb-b1fa-4adf-8f31-8c466048919c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2189.869699] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e6f95cb-45a7-48de-8220-fb8355af6eb3 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2189.869935] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9cb0fd24-be1c-4aca-a8e9-e70453978352 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2189.876422] env[62824]: DEBUG oslo_vmware.api [None req-7e6f95cb-45a7-48de-8220-fb8355af6eb3 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Waiting for the task: (returnval){ [ 2189.876422] env[62824]: value = "task-2146534" [ 2189.876422] env[62824]: _type = "Task" [ 2189.876422] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2189.884128] env[62824]: DEBUG oslo_vmware.api [None req-7e6f95cb-45a7-48de-8220-fb8355af6eb3 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Task: {'id': task-2146534, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2189.888385] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3a1b2b23-923e-4532-b132-fcd23342ce08 tempest-ServerPasswordTestJSON-1515218080 tempest-ServerPasswordTestJSON-1515218080-project-member] Lock "bd7f2ed5-fa08-4a48-9755-60b61791ea0b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.941s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2190.053226] env[62824]: DEBUG nova.network.neutron [req-8c9f8e1c-f5de-40d8-b6de-430cb2a7bbcc req-c6f2dcd9-0f18-4f8e-9734-5e3683ed908e service nova] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Updated VIF entry in instance network info cache for port 0cefe7bc-5dd6-4f4c-872b-ef1b658a4d60. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2190.053604] env[62824]: DEBUG nova.network.neutron [req-8c9f8e1c-f5de-40d8-b6de-430cb2a7bbcc req-c6f2dcd9-0f18-4f8e-9734-5e3683ed908e service nova] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Updating instance_info_cache with network_info: [{"id": "0cefe7bc-5dd6-4f4c-872b-ef1b658a4d60", "address": "fa:16:3e:e0:51:b5", "network": {"id": "3210f2eb-2afb-4bce-bb59-f66e85888333", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-216881410-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "c8133feb4bde41e9898b3f7019c24d30", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d2bf584a-b4a3-4e7a-b0b7-eb8a2bc5a11d", "external-id": "nsx-vlan-transportzone-286", "segmentation_id": 286, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0cefe7bc-5d", "ovs_interfaceid": "0cefe7bc-5dd6-4f4c-872b-ef1b658a4d60", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2190.114027] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 58243465-ba24-4b37-a640-04ab24963230] Creating Snapshot of the VM instance {{(pid=62824) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 2190.114435] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-f12708bb-c241-4e17-a2cc-f3567d052177 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2190.122195] env[62824]: DEBUG oslo_vmware.api [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 2190.122195] env[62824]: value = "task-2146535" [ 2190.122195] env[62824]: _type = "Task" [ 2190.122195] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2190.132590] env[62824]: DEBUG oslo_vmware.api [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146535, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2190.388016] env[62824]: DEBUG oslo_vmware.api [None req-7e6f95cb-45a7-48de-8220-fb8355af6eb3 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Task: {'id': task-2146534, 'name': PowerOffVM_Task, 'duration_secs': 0.203854} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2190.388309] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e6f95cb-45a7-48de-8220-fb8355af6eb3 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2190.388478] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-7e6f95cb-45a7-48de-8220-fb8355af6eb3 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2190.388728] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-22c0ca22-7410-4701-ad78-6aa633c7a913 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2190.556728] env[62824]: DEBUG oslo_concurrency.lockutils [req-8c9f8e1c-f5de-40d8-b6de-430cb2a7bbcc req-c6f2dcd9-0f18-4f8e-9734-5e3683ed908e service nova] Releasing lock "refresh_cache-da46d1d5-f75f-4ef1-b571-fbebab89c2a6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2190.557520] env[62824]: DEBUG nova.compute.manager [req-8c9f8e1c-f5de-40d8-b6de-430cb2a7bbcc req-c6f2dcd9-0f18-4f8e-9734-5e3683ed908e service nova] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Received event network-changed-0cefe7bc-5dd6-4f4c-872b-ef1b658a4d60 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2190.557520] env[62824]: DEBUG nova.compute.manager [req-8c9f8e1c-f5de-40d8-b6de-430cb2a7bbcc req-c6f2dcd9-0f18-4f8e-9734-5e3683ed908e service nova] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Refreshing instance network info cache due to event network-changed-0cefe7bc-5dd6-4f4c-872b-ef1b658a4d60. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2190.557520] env[62824]: DEBUG oslo_concurrency.lockutils [req-8c9f8e1c-f5de-40d8-b6de-430cb2a7bbcc req-c6f2dcd9-0f18-4f8e-9734-5e3683ed908e service nova] Acquiring lock "refresh_cache-da46d1d5-f75f-4ef1-b571-fbebab89c2a6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2190.557520] env[62824]: DEBUG oslo_concurrency.lockutils [req-8c9f8e1c-f5de-40d8-b6de-430cb2a7bbcc req-c6f2dcd9-0f18-4f8e-9734-5e3683ed908e service nova] Acquired lock "refresh_cache-da46d1d5-f75f-4ef1-b571-fbebab89c2a6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2190.557704] env[62824]: DEBUG nova.network.neutron [req-8c9f8e1c-f5de-40d8-b6de-430cb2a7bbcc req-c6f2dcd9-0f18-4f8e-9734-5e3683ed908e service nova] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Refreshing network info cache for port 0cefe7bc-5dd6-4f4c-872b-ef1b658a4d60 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2190.634182] env[62824]: DEBUG oslo_vmware.api [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146535, 'name': CreateSnapshot_Task, 'duration_secs': 0.413139} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2190.634464] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 58243465-ba24-4b37-a640-04ab24963230] Created Snapshot of the VM instance {{(pid=62824) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 2190.635220] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d50a46d-2693-45f5-a68f-ff8a5fba64a7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2190.681370] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f6e4b4b-a83f-4a92-a9c2-03b0e62ecbcd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2190.689153] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-054ea9bb-5744-4275-92f8-390745ffb527 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2190.725113] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e178521-546c-4457-be77-b3b93c82c44e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2190.728097] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-7e6f95cb-45a7-48de-8220-fb8355af6eb3 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2190.728355] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-7e6f95cb-45a7-48de-8220-fb8355af6eb3 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Deleting contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2190.728582] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e6f95cb-45a7-48de-8220-fb8355af6eb3 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Deleting the datastore file [datastore1] da46d1d5-f75f-4ef1-b571-fbebab89c2a6 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2190.728855] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ecfc695a-7d30-4eaf-b30e-848410b4ea9f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2190.738128] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72348e84-fdaf-438b-9f37-233db2437ed1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2190.741379] env[62824]: DEBUG oslo_vmware.api [None req-7e6f95cb-45a7-48de-8220-fb8355af6eb3 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Waiting for the task: (returnval){ [ 2190.741379] env[62824]: value = "task-2146537" [ 2190.741379] env[62824]: _type = "Task" [ 2190.741379] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2190.754454] env[62824]: DEBUG nova.compute.provider_tree [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2190.760361] env[62824]: DEBUG oslo_vmware.api [None req-7e6f95cb-45a7-48de-8220-fb8355af6eb3 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Task: {'id': task-2146537, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2191.163033] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 58243465-ba24-4b37-a640-04ab24963230] Creating linked-clone VM from snapshot {{(pid=62824) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 2191.163033] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-a29e47bd-f2b1-4986-b2f4-acfecbbb4bd3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2191.173191] env[62824]: DEBUG oslo_vmware.api [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 2191.173191] env[62824]: value = "task-2146538" [ 2191.173191] env[62824]: _type = "Task" [ 2191.173191] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2191.181608] env[62824]: DEBUG oslo_vmware.api [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146538, 'name': CloneVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2191.258387] env[62824]: DEBUG nova.scheduler.client.report [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2191.262074] env[62824]: DEBUG oslo_vmware.api [None req-7e6f95cb-45a7-48de-8220-fb8355af6eb3 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Task: {'id': task-2146537, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.179287} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2191.262443] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e6f95cb-45a7-48de-8220-fb8355af6eb3 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2191.262674] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-7e6f95cb-45a7-48de-8220-fb8355af6eb3 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Deleted contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2191.262893] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-7e6f95cb-45a7-48de-8220-fb8355af6eb3 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2191.263137] env[62824]: INFO nova.compute.manager [None req-7e6f95cb-45a7-48de-8220-fb8355af6eb3 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Took 1.40 seconds to destroy the instance on the hypervisor. [ 2191.263457] env[62824]: DEBUG oslo.service.loopingcall [None req-7e6f95cb-45a7-48de-8220-fb8355af6eb3 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2191.263932] env[62824]: DEBUG nova.compute.manager [-] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2191.264049] env[62824]: DEBUG nova.network.neutron [-] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2191.288441] env[62824]: DEBUG nova.network.neutron [req-8c9f8e1c-f5de-40d8-b6de-430cb2a7bbcc req-c6f2dcd9-0f18-4f8e-9734-5e3683ed908e service nova] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Updated VIF entry in instance network info cache for port 0cefe7bc-5dd6-4f4c-872b-ef1b658a4d60. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2191.288829] env[62824]: DEBUG nova.network.neutron [req-8c9f8e1c-f5de-40d8-b6de-430cb2a7bbcc req-c6f2dcd9-0f18-4f8e-9734-5e3683ed908e service nova] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Updating instance_info_cache with network_info: [{"id": "0cefe7bc-5dd6-4f4c-872b-ef1b658a4d60", "address": "fa:16:3e:e0:51:b5", "network": {"id": "3210f2eb-2afb-4bce-bb59-f66e85888333", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-216881410-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "c8133feb4bde41e9898b3f7019c24d30", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d2bf584a-b4a3-4e7a-b0b7-eb8a2bc5a11d", "external-id": "nsx-vlan-transportzone-286", "segmentation_id": 286, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0cefe7bc-5d", "ovs_interfaceid": "0cefe7bc-5dd6-4f4c-872b-ef1b658a4d60", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2191.687012] env[62824]: DEBUG oslo_vmware.api [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146538, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2191.698138] env[62824]: DEBUG nova.compute.manager [req-d7cc8afb-551c-4b02-b8b7-5b392f123386 req-dada0ba2-5757-4597-9b1d-9e593ee4ca63 service nova] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Received event network-vif-deleted-0cefe7bc-5dd6-4f4c-872b-ef1b658a4d60 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2191.698361] env[62824]: INFO nova.compute.manager [req-d7cc8afb-551c-4b02-b8b7-5b392f123386 req-dada0ba2-5757-4597-9b1d-9e593ee4ca63 service nova] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Neutron deleted interface 0cefe7bc-5dd6-4f4c-872b-ef1b658a4d60; detaching it from the instance and deleting it from the info cache [ 2191.698596] env[62824]: DEBUG nova.network.neutron [req-d7cc8afb-551c-4b02-b8b7-5b392f123386 req-dada0ba2-5757-4597-9b1d-9e593ee4ca63 service nova] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2192.428280] env[62824]: DEBUG oslo_concurrency.lockutils [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.087s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2192.428760] env[62824]: DEBUG nova.compute.manager [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 8c2ac679-6915-4149-aa51-99170b0df685] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2192.431545] env[62824]: DEBUG oslo_concurrency.lockutils [req-8c9f8e1c-f5de-40d8-b6de-430cb2a7bbcc req-c6f2dcd9-0f18-4f8e-9734-5e3683ed908e service nova] Releasing lock "refresh_cache-da46d1d5-f75f-4ef1-b571-fbebab89c2a6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2192.432331] env[62824]: DEBUG nova.network.neutron [-] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2192.433133] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.465s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2192.436228] env[62824]: INFO nova.compute.claims [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 0c9fb5cc-de69-46d8-9962-98e0a84e33c3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2192.440065] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5e4db5f2-d15d-47b7-b701-b18d18c41e92 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2192.442293] env[62824]: INFO nova.compute.manager [-] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Took 1.18 seconds to deallocate network for instance. [ 2192.450673] env[62824]: DEBUG oslo_vmware.api [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146538, 'name': CloneVM_Task} progress is 95%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2192.455026] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cdf60e0-2406-4994-b43b-ab20a0a77679 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2192.488307] env[62824]: DEBUG nova.compute.manager [req-d7cc8afb-551c-4b02-b8b7-5b392f123386 req-dada0ba2-5757-4597-9b1d-9e593ee4ca63 service nova] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Detach interface failed, port_id=0cefe7bc-5dd6-4f4c-872b-ef1b658a4d60, reason: Instance da46d1d5-f75f-4ef1-b571-fbebab89c2a6 could not be found. {{(pid=62824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11482}} [ 2192.937121] env[62824]: DEBUG nova.compute.utils [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2192.938526] env[62824]: DEBUG nova.compute.manager [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 8c2ac679-6915-4149-aa51-99170b0df685] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2192.938696] env[62824]: DEBUG nova.network.neutron [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 8c2ac679-6915-4149-aa51-99170b0df685] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2192.948481] env[62824]: DEBUG oslo_vmware.api [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146538, 'name': CloneVM_Task, 'duration_secs': 1.722385} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2192.949703] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7e6f95cb-45a7-48de-8220-fb8355af6eb3 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2192.950229] env[62824]: INFO nova.virt.vmwareapi.vmops [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 58243465-ba24-4b37-a640-04ab24963230] Created linked-clone VM from snapshot [ 2192.951346] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47e3d80c-c88b-4588-aa4b-757eb43e0a5d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2192.960360] env[62824]: DEBUG nova.virt.vmwareapi.images [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 58243465-ba24-4b37-a640-04ab24963230] Uploading image 5277bf68-46cb-496f-878c-a6d6908d561b {{(pid=62824) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 2192.986038] env[62824]: DEBUG nova.policy [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '959d127a7d144b33a0cae94db5c11846', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dfff08982dad4790bf4d555e2b4db5e4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2192.989964] env[62824]: DEBUG oslo_vmware.rw_handles [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 2192.989964] env[62824]: value = "vm-438714" [ 2192.989964] env[62824]: _type = "VirtualMachine" [ 2192.989964] env[62824]: }. {{(pid=62824) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 2192.990572] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-1b6cdef8-a183-40c4-823d-7b2fe8878e6a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2192.998948] env[62824]: DEBUG oslo_vmware.rw_handles [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lease: (returnval){ [ 2192.998948] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52a18f5c-4841-2574-a792-69e47e596898" [ 2192.998948] env[62824]: _type = "HttpNfcLease" [ 2192.998948] env[62824]: } obtained for exporting VM: (result){ [ 2192.998948] env[62824]: value = "vm-438714" [ 2192.998948] env[62824]: _type = "VirtualMachine" [ 2192.998948] env[62824]: }. {{(pid=62824) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 2192.999477] env[62824]: DEBUG oslo_vmware.api [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the lease: (returnval){ [ 2192.999477] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52a18f5c-4841-2574-a792-69e47e596898" [ 2192.999477] env[62824]: _type = "HttpNfcLease" [ 2192.999477] env[62824]: } to be ready. {{(pid=62824) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 2193.013219] env[62824]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 2193.013219] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52a18f5c-4841-2574-a792-69e47e596898" [ 2193.013219] env[62824]: _type = "HttpNfcLease" [ 2193.013219] env[62824]: } is initializing. {{(pid=62824) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 2193.237143] env[62824]: DEBUG nova.network.neutron [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 8c2ac679-6915-4149-aa51-99170b0df685] Successfully created port: 891b53e5-1497-406d-8b11-2db60c8c4e1e {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2193.442417] env[62824]: DEBUG nova.compute.manager [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 8c2ac679-6915-4149-aa51-99170b0df685] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2193.509312] env[62824]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 2193.509312] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52a18f5c-4841-2574-a792-69e47e596898" [ 2193.509312] env[62824]: _type = "HttpNfcLease" [ 2193.509312] env[62824]: } is ready. {{(pid=62824) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 2193.511806] env[62824]: DEBUG oslo_vmware.rw_handles [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 2193.511806] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52a18f5c-4841-2574-a792-69e47e596898" [ 2193.511806] env[62824]: _type = "HttpNfcLease" [ 2193.511806] env[62824]: }. {{(pid=62824) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 2193.512806] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54cd7b94-1530-466f-8b6e-3955c2d20bdd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2193.520658] env[62824]: DEBUG oslo_vmware.rw_handles [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52eed9d3-7100-8b68-b837-4f50f98ea027/disk-0.vmdk from lease info. {{(pid=62824) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 2193.520886] env[62824]: DEBUG oslo_vmware.rw_handles [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52eed9d3-7100-8b68-b837-4f50f98ea027/disk-0.vmdk for reading. {{(pid=62824) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 2193.619826] env[62824]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-83167da9-809a-4554-b076-eeccb8837f94 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2193.884979] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de9b182b-de7a-4cdd-8931-8ede453adcb9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2193.892889] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33ee351c-bcf6-41ad-bcb1-b34af771b344 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2193.923418] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b467c17-556f-4a25-b29e-24a092416557 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2193.931825] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4e55b87-78d1-42d6-b62f-5e53fbe91ae8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2193.946226] env[62824]: DEBUG nova.compute.provider_tree [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2194.449690] env[62824]: DEBUG nova.scheduler.client.report [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2194.454220] env[62824]: DEBUG nova.compute.manager [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 8c2ac679-6915-4149-aa51-99170b0df685] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2194.476856] env[62824]: DEBUG nova.virt.hardware [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2194.477398] env[62824]: DEBUG nova.virt.hardware [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2194.477825] env[62824]: DEBUG nova.virt.hardware [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2194.478076] env[62824]: DEBUG nova.virt.hardware [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2194.478422] env[62824]: DEBUG nova.virt.hardware [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2194.478663] env[62824]: DEBUG nova.virt.hardware [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2194.478930] env[62824]: DEBUG nova.virt.hardware [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2194.479167] env[62824]: DEBUG nova.virt.hardware [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2194.479405] env[62824]: DEBUG nova.virt.hardware [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2194.479663] env[62824]: DEBUG nova.virt.hardware [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2194.479800] env[62824]: DEBUG nova.virt.hardware [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2194.481030] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51c869d1-0b37-42f9-9b37-c0cb12b65526 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2194.491668] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6119de33-4d14-4635-b1a7-d8ce1908863a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2194.799123] env[62824]: DEBUG nova.compute.manager [req-5f14750e-e5cd-4827-a35e-452a227a09d3 req-0575b550-df4e-43b8-a236-8dea199804f4 service nova] [instance: 8c2ac679-6915-4149-aa51-99170b0df685] Received event network-vif-plugged-891b53e5-1497-406d-8b11-2db60c8c4e1e {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2194.799302] env[62824]: DEBUG oslo_concurrency.lockutils [req-5f14750e-e5cd-4827-a35e-452a227a09d3 req-0575b550-df4e-43b8-a236-8dea199804f4 service nova] Acquiring lock "8c2ac679-6915-4149-aa51-99170b0df685-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2194.799516] env[62824]: DEBUG oslo_concurrency.lockutils [req-5f14750e-e5cd-4827-a35e-452a227a09d3 req-0575b550-df4e-43b8-a236-8dea199804f4 service nova] Lock "8c2ac679-6915-4149-aa51-99170b0df685-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2194.799684] env[62824]: DEBUG oslo_concurrency.lockutils [req-5f14750e-e5cd-4827-a35e-452a227a09d3 req-0575b550-df4e-43b8-a236-8dea199804f4 service nova] Lock "8c2ac679-6915-4149-aa51-99170b0df685-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2194.799910] env[62824]: DEBUG nova.compute.manager [req-5f14750e-e5cd-4827-a35e-452a227a09d3 req-0575b550-df4e-43b8-a236-8dea199804f4 service nova] [instance: 8c2ac679-6915-4149-aa51-99170b0df685] No waiting events found dispatching network-vif-plugged-891b53e5-1497-406d-8b11-2db60c8c4e1e {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2194.800192] env[62824]: WARNING nova.compute.manager [req-5f14750e-e5cd-4827-a35e-452a227a09d3 req-0575b550-df4e-43b8-a236-8dea199804f4 service nova] [instance: 8c2ac679-6915-4149-aa51-99170b0df685] Received unexpected event network-vif-plugged-891b53e5-1497-406d-8b11-2db60c8c4e1e for instance with vm_state building and task_state spawning. [ 2194.935585] env[62824]: DEBUG nova.network.neutron [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 8c2ac679-6915-4149-aa51-99170b0df685] Successfully updated port: 891b53e5-1497-406d-8b11-2db60c8c4e1e {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2194.957795] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.524s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2194.958353] env[62824]: DEBUG nova.compute.manager [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 0c9fb5cc-de69-46d8-9962-98e0a84e33c3] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2194.960999] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.598s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2194.962454] env[62824]: INFO nova.compute.claims [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2195.438882] env[62824]: DEBUG oslo_concurrency.lockutils [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquiring lock "refresh_cache-8c2ac679-6915-4149-aa51-99170b0df685" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2195.439065] env[62824]: DEBUG oslo_concurrency.lockutils [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquired lock "refresh_cache-8c2ac679-6915-4149-aa51-99170b0df685" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2195.439244] env[62824]: DEBUG nova.network.neutron [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 8c2ac679-6915-4149-aa51-99170b0df685] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2195.467477] env[62824]: DEBUG nova.compute.utils [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2195.471689] env[62824]: DEBUG nova.compute.manager [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 0c9fb5cc-de69-46d8-9962-98e0a84e33c3] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2195.471917] env[62824]: DEBUG nova.network.neutron [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 0c9fb5cc-de69-46d8-9962-98e0a84e33c3] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2195.522431] env[62824]: DEBUG nova.policy [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '28495352107b45cd9cbd746c6affc4fe', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '64042a790d6c459186c68d73da32c019', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2195.974965] env[62824]: DEBUG nova.compute.manager [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 0c9fb5cc-de69-46d8-9962-98e0a84e33c3] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2195.985938] env[62824]: DEBUG nova.network.neutron [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 8c2ac679-6915-4149-aa51-99170b0df685] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2196.174173] env[62824]: DEBUG nova.network.neutron [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 8c2ac679-6915-4149-aa51-99170b0df685] Updating instance_info_cache with network_info: [{"id": "891b53e5-1497-406d-8b11-2db60c8c4e1e", "address": "fa:16:3e:d8:d6:01", "network": {"id": "0f512108-670f-4a85-9381-108e18b3b5ed", "bridge": "br-int", "label": "tempest-ServersTestJSON-766397782-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dfff08982dad4790bf4d555e2b4db5e4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f3a2eb5-353f-45c5-a73b-869626f4bb13", "external-id": "nsx-vlan-transportzone-411", "segmentation_id": 411, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap891b53e5-14", "ovs_interfaceid": "891b53e5-1497-406d-8b11-2db60c8c4e1e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2196.221465] env[62824]: DEBUG nova.network.neutron [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 0c9fb5cc-de69-46d8-9962-98e0a84e33c3] Successfully created port: b96f49a1-2e1e-495b-9809-a7338087a686 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2196.383254] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18a1b21a-1403-48aa-b17a-f34546477f59 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2196.390053] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b581f837-2b08-40f0-bcc5-45127b6231e0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2196.422557] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed398e95-d5b9-4c89-8c84-7735f41392b7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2196.431201] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96fd2f30-eaf6-442c-9476-df42f4dcf216 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2196.445539] env[62824]: DEBUG nova.compute.provider_tree [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2196.678742] env[62824]: DEBUG oslo_concurrency.lockutils [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Releasing lock "refresh_cache-8c2ac679-6915-4149-aa51-99170b0df685" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2196.679154] env[62824]: DEBUG nova.compute.manager [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 8c2ac679-6915-4149-aa51-99170b0df685] Instance network_info: |[{"id": "891b53e5-1497-406d-8b11-2db60c8c4e1e", "address": "fa:16:3e:d8:d6:01", "network": {"id": "0f512108-670f-4a85-9381-108e18b3b5ed", "bridge": "br-int", "label": "tempest-ServersTestJSON-766397782-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dfff08982dad4790bf4d555e2b4db5e4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f3a2eb5-353f-45c5-a73b-869626f4bb13", "external-id": "nsx-vlan-transportzone-411", "segmentation_id": 411, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap891b53e5-14", "ovs_interfaceid": "891b53e5-1497-406d-8b11-2db60c8c4e1e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2196.679735] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 8c2ac679-6915-4149-aa51-99170b0df685] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d8:d6:01', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9f3a2eb5-353f-45c5-a73b-869626f4bb13', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '891b53e5-1497-406d-8b11-2db60c8c4e1e', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2196.687367] env[62824]: DEBUG oslo.service.loopingcall [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2196.687549] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8c2ac679-6915-4149-aa51-99170b0df685] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2196.687780] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b5553e1f-4d92-4230-8808-b39b0b963287 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2196.708768] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2196.708768] env[62824]: value = "task-2146540" [ 2196.708768] env[62824]: _type = "Task" [ 2196.708768] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2196.717107] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146540, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2196.826764] env[62824]: DEBUG nova.compute.manager [req-e97ba051-b7aa-4084-9907-d5a10cf4b2a0 req-430e03ce-2d04-4024-8b16-1e41056f4de2 service nova] [instance: 8c2ac679-6915-4149-aa51-99170b0df685] Received event network-changed-891b53e5-1497-406d-8b11-2db60c8c4e1e {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2196.826825] env[62824]: DEBUG nova.compute.manager [req-e97ba051-b7aa-4084-9907-d5a10cf4b2a0 req-430e03ce-2d04-4024-8b16-1e41056f4de2 service nova] [instance: 8c2ac679-6915-4149-aa51-99170b0df685] Refreshing instance network info cache due to event network-changed-891b53e5-1497-406d-8b11-2db60c8c4e1e. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2196.827059] env[62824]: DEBUG oslo_concurrency.lockutils [req-e97ba051-b7aa-4084-9907-d5a10cf4b2a0 req-430e03ce-2d04-4024-8b16-1e41056f4de2 service nova] Acquiring lock "refresh_cache-8c2ac679-6915-4149-aa51-99170b0df685" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2196.827205] env[62824]: DEBUG oslo_concurrency.lockutils [req-e97ba051-b7aa-4084-9907-d5a10cf4b2a0 req-430e03ce-2d04-4024-8b16-1e41056f4de2 service nova] Acquired lock "refresh_cache-8c2ac679-6915-4149-aa51-99170b0df685" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2196.827369] env[62824]: DEBUG nova.network.neutron [req-e97ba051-b7aa-4084-9907-d5a10cf4b2a0 req-430e03ce-2d04-4024-8b16-1e41056f4de2 service nova] [instance: 8c2ac679-6915-4149-aa51-99170b0df685] Refreshing network info cache for port 891b53e5-1497-406d-8b11-2db60c8c4e1e {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2196.949869] env[62824]: DEBUG nova.scheduler.client.report [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2196.988632] env[62824]: DEBUG nova.compute.manager [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 0c9fb5cc-de69-46d8-9962-98e0a84e33c3] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2197.022271] env[62824]: DEBUG nova.virt.hardware [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2197.022472] env[62824]: DEBUG nova.virt.hardware [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2197.022632] env[62824]: DEBUG nova.virt.hardware [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2197.022815] env[62824]: DEBUG nova.virt.hardware [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2197.022960] env[62824]: DEBUG nova.virt.hardware [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2197.023122] env[62824]: DEBUG nova.virt.hardware [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2197.023332] env[62824]: DEBUG nova.virt.hardware [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2197.023494] env[62824]: DEBUG nova.virt.hardware [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2197.023661] env[62824]: DEBUG nova.virt.hardware [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2197.023822] env[62824]: DEBUG nova.virt.hardware [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2197.024046] env[62824]: DEBUG nova.virt.hardware [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2197.024980] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd4d7327-1fa3-436f-a53c-9d6c49e7b69c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2197.034202] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36682a15-5d59-4c6e-8cd5-05be1f87734d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2197.219142] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146540, 'name': CreateVM_Task, 'duration_secs': 0.405798} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2197.219358] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8c2ac679-6915-4149-aa51-99170b0df685] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2197.220069] env[62824]: DEBUG oslo_concurrency.lockutils [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2197.220238] env[62824]: DEBUG oslo_concurrency.lockutils [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2197.220930] env[62824]: DEBUG oslo_concurrency.lockutils [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2197.220930] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6129de98-7cc4-46b8-92bc-082fe71af2af {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2197.225586] env[62824]: DEBUG oslo_vmware.api [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2197.225586] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5240d8fe-1f9c-3586-9054-7c106ce769eb" [ 2197.225586] env[62824]: _type = "Task" [ 2197.225586] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2197.233738] env[62824]: DEBUG oslo_vmware.api [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5240d8fe-1f9c-3586-9054-7c106ce769eb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2197.454671] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.494s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2197.455476] env[62824]: DEBUG nova.compute.manager [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2197.458914] env[62824]: DEBUG oslo_concurrency.lockutils [None req-06ef8715-0d4a-45b8-bcde-6870d18da01a tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.305s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2197.459218] env[62824]: DEBUG nova.objects.instance [None req-06ef8715-0d4a-45b8-bcde-6870d18da01a tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Lazy-loading 'resources' on Instance uuid ab37b1a2-0012-40fb-9341-b613525e89cf {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2197.568595] env[62824]: DEBUG nova.network.neutron [req-e97ba051-b7aa-4084-9907-d5a10cf4b2a0 req-430e03ce-2d04-4024-8b16-1e41056f4de2 service nova] [instance: 8c2ac679-6915-4149-aa51-99170b0df685] Updated VIF entry in instance network info cache for port 891b53e5-1497-406d-8b11-2db60c8c4e1e. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2197.568964] env[62824]: DEBUG nova.network.neutron [req-e97ba051-b7aa-4084-9907-d5a10cf4b2a0 req-430e03ce-2d04-4024-8b16-1e41056f4de2 service nova] [instance: 8c2ac679-6915-4149-aa51-99170b0df685] Updating instance_info_cache with network_info: [{"id": "891b53e5-1497-406d-8b11-2db60c8c4e1e", "address": "fa:16:3e:d8:d6:01", "network": {"id": "0f512108-670f-4a85-9381-108e18b3b5ed", "bridge": "br-int", "label": "tempest-ServersTestJSON-766397782-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dfff08982dad4790bf4d555e2b4db5e4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f3a2eb5-353f-45c5-a73b-869626f4bb13", "external-id": "nsx-vlan-transportzone-411", "segmentation_id": 411, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap891b53e5-14", "ovs_interfaceid": "891b53e5-1497-406d-8b11-2db60c8c4e1e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2197.736927] env[62824]: DEBUG oslo_vmware.api [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5240d8fe-1f9c-3586-9054-7c106ce769eb, 'name': SearchDatastore_Task, 'duration_secs': 0.00939} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2197.737296] env[62824]: DEBUG oslo_concurrency.lockutils [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2197.737554] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 8c2ac679-6915-4149-aa51-99170b0df685] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2197.737793] env[62824]: DEBUG oslo_concurrency.lockutils [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2197.737938] env[62824]: DEBUG oslo_concurrency.lockutils [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2197.738157] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2197.738461] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d18d9f48-bf2c-4188-8470-b161c938c4ba {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2197.747891] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2197.748133] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2197.749327] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e1794c44-2898-4cb0-83c6-a11f49167f2f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2197.757042] env[62824]: DEBUG oslo_vmware.api [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2197.757042] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]527e927f-a443-4097-3046-5a6191a7f09c" [ 2197.757042] env[62824]: _type = "Task" [ 2197.757042] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2197.763105] env[62824]: DEBUG oslo_vmware.api [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]527e927f-a443-4097-3046-5a6191a7f09c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2197.807869] env[62824]: DEBUG nova.network.neutron [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 0c9fb5cc-de69-46d8-9962-98e0a84e33c3] Successfully updated port: b96f49a1-2e1e-495b-9809-a7338087a686 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2197.961024] env[62824]: DEBUG nova.compute.utils [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2197.962803] env[62824]: DEBUG nova.compute.manager [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2197.962969] env[62824]: DEBUG nova.network.neutron [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2198.019172] env[62824]: DEBUG nova.policy [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2d3d79fc7c0145c0ba97a846553026c3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '102d2f886fd44e1da6780e48e36f1b62', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2198.071738] env[62824]: DEBUG oslo_concurrency.lockutils [req-e97ba051-b7aa-4084-9907-d5a10cf4b2a0 req-430e03ce-2d04-4024-8b16-1e41056f4de2 service nova] Releasing lock "refresh_cache-8c2ac679-6915-4149-aa51-99170b0df685" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2198.265828] env[62824]: DEBUG oslo_vmware.api [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]527e927f-a443-4097-3046-5a6191a7f09c, 'name': SearchDatastore_Task, 'duration_secs': 0.00943} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2198.269929] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8c31b833-a081-4ea7-b545-ea9b594989c7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2198.275157] env[62824]: DEBUG oslo_vmware.api [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2198.275157] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52e4631a-c03a-b9c6-4bdc-16a42df37183" [ 2198.275157] env[62824]: _type = "Task" [ 2198.275157] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2198.288819] env[62824]: DEBUG oslo_vmware.api [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52e4631a-c03a-b9c6-4bdc-16a42df37183, 'name': SearchDatastore_Task, 'duration_secs': 0.011394} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2198.289121] env[62824]: DEBUG oslo_concurrency.lockutils [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2198.289399] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 8c2ac679-6915-4149-aa51-99170b0df685/8c2ac679-6915-4149-aa51-99170b0df685.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2198.289663] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-25784d67-6d58-43e8-add1-b6aeec070d61 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2198.296624] env[62824]: DEBUG oslo_vmware.api [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2198.296624] env[62824]: value = "task-2146541" [ 2198.296624] env[62824]: _type = "Task" [ 2198.296624] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2198.306737] env[62824]: DEBUG oslo_vmware.api [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146541, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2198.313196] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Acquiring lock "refresh_cache-0c9fb5cc-de69-46d8-9962-98e0a84e33c3" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2198.313196] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Acquired lock "refresh_cache-0c9fb5cc-de69-46d8-9962-98e0a84e33c3" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2198.313196] env[62824]: DEBUG nova.network.neutron [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 0c9fb5cc-de69-46d8-9962-98e0a84e33c3] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2198.361103] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dfe5628-9579-4f86-9b8b-cb2b0a9c900f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2198.375657] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73749b87-d8ef-45af-885c-886d8b2d7f3c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2198.408229] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d007715-efac-45ba-b77a-ee04f417fd09 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2198.416692] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7bb978c-d886-44b9-bb77-68639f478c82 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2198.430857] env[62824]: DEBUG nova.compute.provider_tree [None req-06ef8715-0d4a-45b8-bcde-6870d18da01a tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2198.469484] env[62824]: DEBUG nova.compute.manager [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2198.512744] env[62824]: DEBUG nova.network.neutron [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412] Successfully created port: 0f67546a-a44f-4c20-96e8-e7b09334b221 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2198.807844] env[62824]: DEBUG oslo_vmware.api [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146541, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.501834} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2198.809157] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 8c2ac679-6915-4149-aa51-99170b0df685/8c2ac679-6915-4149-aa51-99170b0df685.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2198.809570] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 8c2ac679-6915-4149-aa51-99170b0df685] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2198.809983] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-22eb99b0-0027-44b6-b5a0-7599414b4e2d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2198.819972] env[62824]: DEBUG oslo_vmware.api [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2198.819972] env[62824]: value = "task-2146542" [ 2198.819972] env[62824]: _type = "Task" [ 2198.819972] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2198.830918] env[62824]: DEBUG oslo_vmware.api [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146542, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2198.850120] env[62824]: DEBUG nova.network.neutron [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 0c9fb5cc-de69-46d8-9962-98e0a84e33c3] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2198.887566] env[62824]: DEBUG nova.compute.manager [req-4324f5c6-c55c-4b51-9f3d-9623a7782cef req-6cdccbe5-79d1-4859-aa34-0388e39d7c57 service nova] [instance: 0c9fb5cc-de69-46d8-9962-98e0a84e33c3] Received event network-vif-plugged-b96f49a1-2e1e-495b-9809-a7338087a686 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2198.887566] env[62824]: DEBUG oslo_concurrency.lockutils [req-4324f5c6-c55c-4b51-9f3d-9623a7782cef req-6cdccbe5-79d1-4859-aa34-0388e39d7c57 service nova] Acquiring lock "0c9fb5cc-de69-46d8-9962-98e0a84e33c3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2198.887566] env[62824]: DEBUG oslo_concurrency.lockutils [req-4324f5c6-c55c-4b51-9f3d-9623a7782cef req-6cdccbe5-79d1-4859-aa34-0388e39d7c57 service nova] Lock "0c9fb5cc-de69-46d8-9962-98e0a84e33c3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2198.887566] env[62824]: DEBUG oslo_concurrency.lockutils [req-4324f5c6-c55c-4b51-9f3d-9623a7782cef req-6cdccbe5-79d1-4859-aa34-0388e39d7c57 service nova] Lock "0c9fb5cc-de69-46d8-9962-98e0a84e33c3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2198.887566] env[62824]: DEBUG nova.compute.manager [req-4324f5c6-c55c-4b51-9f3d-9623a7782cef req-6cdccbe5-79d1-4859-aa34-0388e39d7c57 service nova] [instance: 0c9fb5cc-de69-46d8-9962-98e0a84e33c3] No waiting events found dispatching network-vif-plugged-b96f49a1-2e1e-495b-9809-a7338087a686 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2198.887566] env[62824]: WARNING nova.compute.manager [req-4324f5c6-c55c-4b51-9f3d-9623a7782cef req-6cdccbe5-79d1-4859-aa34-0388e39d7c57 service nova] [instance: 0c9fb5cc-de69-46d8-9962-98e0a84e33c3] Received unexpected event network-vif-plugged-b96f49a1-2e1e-495b-9809-a7338087a686 for instance with vm_state building and task_state spawning. [ 2198.887566] env[62824]: DEBUG nova.compute.manager [req-4324f5c6-c55c-4b51-9f3d-9623a7782cef req-6cdccbe5-79d1-4859-aa34-0388e39d7c57 service nova] [instance: 0c9fb5cc-de69-46d8-9962-98e0a84e33c3] Received event network-changed-b96f49a1-2e1e-495b-9809-a7338087a686 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2198.887566] env[62824]: DEBUG nova.compute.manager [req-4324f5c6-c55c-4b51-9f3d-9623a7782cef req-6cdccbe5-79d1-4859-aa34-0388e39d7c57 service nova] [instance: 0c9fb5cc-de69-46d8-9962-98e0a84e33c3] Refreshing instance network info cache due to event network-changed-b96f49a1-2e1e-495b-9809-a7338087a686. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2198.887566] env[62824]: DEBUG oslo_concurrency.lockutils [req-4324f5c6-c55c-4b51-9f3d-9623a7782cef req-6cdccbe5-79d1-4859-aa34-0388e39d7c57 service nova] Acquiring lock "refresh_cache-0c9fb5cc-de69-46d8-9962-98e0a84e33c3" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2198.936764] env[62824]: DEBUG nova.scheduler.client.report [None req-06ef8715-0d4a-45b8-bcde-6870d18da01a tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2199.023149] env[62824]: DEBUG nova.network.neutron [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 0c9fb5cc-de69-46d8-9962-98e0a84e33c3] Updating instance_info_cache with network_info: [{"id": "b96f49a1-2e1e-495b-9809-a7338087a686", "address": "fa:16:3e:b5:0f:b3", "network": {"id": "9e1885b6-0796-42b9-9cd8-8fa9ce9a3cfd", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-2118936609-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "64042a790d6c459186c68d73da32c019", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "afae6acd-1873-4228-9d5a-1cd5d4efe3e4", "external-id": "nsx-vlan-transportzone-183", "segmentation_id": 183, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb96f49a1-2e", "ovs_interfaceid": "b96f49a1-2e1e-495b-9809-a7338087a686", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2199.113908] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b9a05fc7-ab92-4245-87d2-ec7f1181175f tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Acquiring lock "b9f30f87-3594-4468-9d29-70890d8761e3" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2199.113908] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b9a05fc7-ab92-4245-87d2-ec7f1181175f tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Lock "b9f30f87-3594-4468-9d29-70890d8761e3" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2199.331231] env[62824]: DEBUG oslo_vmware.api [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146542, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.139531} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2199.331544] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 8c2ac679-6915-4149-aa51-99170b0df685] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2199.332413] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5822975f-1ca8-4800-9220-a973563d04df {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2199.355898] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 8c2ac679-6915-4149-aa51-99170b0df685] Reconfiguring VM instance instance-00000054 to attach disk [datastore1] 8c2ac679-6915-4149-aa51-99170b0df685/8c2ac679-6915-4149-aa51-99170b0df685.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2199.356293] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1089d93e-cb90-4fac-aab4-bb08493eff96 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2199.378958] env[62824]: DEBUG oslo_vmware.api [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2199.378958] env[62824]: value = "task-2146543" [ 2199.378958] env[62824]: _type = "Task" [ 2199.378958] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2199.388321] env[62824]: DEBUG oslo_vmware.api [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146543, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2199.440690] env[62824]: DEBUG oslo_concurrency.lockutils [None req-06ef8715-0d4a-45b8-bcde-6870d18da01a tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.982s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2199.443457] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f32f5e12-bb47-427a-aceb-0383bc172546 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.500s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2199.443674] env[62824]: DEBUG nova.objects.instance [None req-f32f5e12-bb47-427a-aceb-0383bc172546 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Lazy-loading 'resources' on Instance uuid 2afd1f18-234c-40b5-9ede-7413ad30dafe {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2199.466314] env[62824]: INFO nova.scheduler.client.report [None req-06ef8715-0d4a-45b8-bcde-6870d18da01a tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Deleted allocations for instance ab37b1a2-0012-40fb-9341-b613525e89cf [ 2199.479567] env[62824]: DEBUG nova.compute.manager [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2199.507032] env[62824]: DEBUG nova.virt.hardware [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2199.507032] env[62824]: DEBUG nova.virt.hardware [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2199.507032] env[62824]: DEBUG nova.virt.hardware [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2199.507832] env[62824]: DEBUG nova.virt.hardware [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2199.507832] env[62824]: DEBUG nova.virt.hardware [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2199.507832] env[62824]: DEBUG nova.virt.hardware [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2199.508017] env[62824]: DEBUG nova.virt.hardware [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2199.508054] env[62824]: DEBUG nova.virt.hardware [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2199.508260] env[62824]: DEBUG nova.virt.hardware [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2199.508442] env[62824]: DEBUG nova.virt.hardware [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2199.508619] env[62824]: DEBUG nova.virt.hardware [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2199.509549] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e77f7b79-7a77-46e7-8342-bbbf2b4aff52 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2199.518059] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbee3b1b-6ce8-42d9-8d15-d4645483c103 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2199.532533] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Releasing lock "refresh_cache-0c9fb5cc-de69-46d8-9962-98e0a84e33c3" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2199.532854] env[62824]: DEBUG nova.compute.manager [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 0c9fb5cc-de69-46d8-9962-98e0a84e33c3] Instance network_info: |[{"id": "b96f49a1-2e1e-495b-9809-a7338087a686", "address": "fa:16:3e:b5:0f:b3", "network": {"id": "9e1885b6-0796-42b9-9cd8-8fa9ce9a3cfd", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-2118936609-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "64042a790d6c459186c68d73da32c019", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "afae6acd-1873-4228-9d5a-1cd5d4efe3e4", "external-id": "nsx-vlan-transportzone-183", "segmentation_id": 183, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb96f49a1-2e", "ovs_interfaceid": "b96f49a1-2e1e-495b-9809-a7338087a686", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2199.533654] env[62824]: DEBUG oslo_concurrency.lockutils [req-4324f5c6-c55c-4b51-9f3d-9623a7782cef req-6cdccbe5-79d1-4859-aa34-0388e39d7c57 service nova] Acquired lock "refresh_cache-0c9fb5cc-de69-46d8-9962-98e0a84e33c3" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2199.533870] env[62824]: DEBUG nova.network.neutron [req-4324f5c6-c55c-4b51-9f3d-9623a7782cef req-6cdccbe5-79d1-4859-aa34-0388e39d7c57 service nova] [instance: 0c9fb5cc-de69-46d8-9962-98e0a84e33c3] Refreshing network info cache for port b96f49a1-2e1e-495b-9809-a7338087a686 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2199.535204] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 0c9fb5cc-de69-46d8-9962-98e0a84e33c3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b5:0f:b3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'afae6acd-1873-4228-9d5a-1cd5d4efe3e4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b96f49a1-2e1e-495b-9809-a7338087a686', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2199.542660] env[62824]: DEBUG oslo.service.loopingcall [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2199.543778] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0c9fb5cc-de69-46d8-9962-98e0a84e33c3] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2199.543995] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e61ad692-12e5-4d21-bd29-f330b7bfaa5f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2199.564827] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2199.564827] env[62824]: value = "task-2146544" [ 2199.564827] env[62824]: _type = "Task" [ 2199.564827] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2199.573281] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146544, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2199.617520] env[62824]: DEBUG nova.compute.utils [None req-b9a05fc7-ab92-4245-87d2-ec7f1181175f tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2199.889891] env[62824]: DEBUG oslo_vmware.api [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146543, 'name': ReconfigVM_Task, 'duration_secs': 0.492753} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2199.890328] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 8c2ac679-6915-4149-aa51-99170b0df685] Reconfigured VM instance instance-00000054 to attach disk [datastore1] 8c2ac679-6915-4149-aa51-99170b0df685/8c2ac679-6915-4149-aa51-99170b0df685.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2199.890843] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ec0273df-0cc7-4636-81e9-834dbd90a057 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2199.897436] env[62824]: DEBUG oslo_vmware.api [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2199.897436] env[62824]: value = "task-2146545" [ 2199.897436] env[62824]: _type = "Task" [ 2199.897436] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2199.905648] env[62824]: DEBUG oslo_vmware.api [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146545, 'name': Rename_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2199.975188] env[62824]: DEBUG oslo_concurrency.lockutils [None req-06ef8715-0d4a-45b8-bcde-6870d18da01a tempest-ServerMetadataTestJSON-290712645 tempest-ServerMetadataTestJSON-290712645-project-member] Lock "ab37b1a2-0012-40fb-9341-b613525e89cf" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.777s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2200.075751] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146544, 'name': CreateVM_Task, 'duration_secs': 0.451522} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2200.079810] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0c9fb5cc-de69-46d8-9962-98e0a84e33c3] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2200.082877] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2200.083064] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2200.083410] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2200.084342] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-de2bd201-7499-4342-86d9-83a46929c2de {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2200.089873] env[62824]: DEBUG oslo_vmware.api [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Waiting for the task: (returnval){ [ 2200.089873] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52934cc0-5a84-cbbb-7d4e-96ea0b331a5d" [ 2200.089873] env[62824]: _type = "Task" [ 2200.089873] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2200.109427] env[62824]: DEBUG oslo_vmware.api [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52934cc0-5a84-cbbb-7d4e-96ea0b331a5d, 'name': SearchDatastore_Task, 'duration_secs': 0.010576} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2200.109764] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2200.110112] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 0c9fb5cc-de69-46d8-9962-98e0a84e33c3] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2200.110295] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2200.110449] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2200.110638] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2200.111039] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ce5a6506-18b6-42e3-87cf-3fed55d59701 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2200.122690] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b9a05fc7-ab92-4245-87d2-ec7f1181175f tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Lock "b9f30f87-3594-4468-9d29-70890d8761e3" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.009s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2200.126134] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2200.126387] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2200.131130] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2629e0c6-a6a7-4fc5-be09-4983621667f3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2200.138436] env[62824]: DEBUG oslo_vmware.api [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Waiting for the task: (returnval){ [ 2200.138436] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]522a2f47-3867-0193-a877-4ef98ffbdc27" [ 2200.138436] env[62824]: _type = "Task" [ 2200.138436] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2200.155715] env[62824]: DEBUG oslo_vmware.api [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]522a2f47-3867-0193-a877-4ef98ffbdc27, 'name': SearchDatastore_Task, 'duration_secs': 0.012317} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2200.157088] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0dfceb94-20f9-4297-9108-52b571398c02 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2200.162479] env[62824]: DEBUG oslo_vmware.api [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Waiting for the task: (returnval){ [ 2200.162479] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52ebac41-6b04-bf40-bac2-fc13326f9eb6" [ 2200.162479] env[62824]: _type = "Task" [ 2200.162479] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2200.176177] env[62824]: DEBUG oslo_vmware.api [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52ebac41-6b04-bf40-bac2-fc13326f9eb6, 'name': SearchDatastore_Task, 'duration_secs': 0.008767} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2200.176473] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2200.176759] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 0c9fb5cc-de69-46d8-9962-98e0a84e33c3/0c9fb5cc-de69-46d8-9962-98e0a84e33c3.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2200.177038] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-810333a1-cf1c-4b6b-90e1-172755eba4b9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2200.188648] env[62824]: DEBUG oslo_vmware.api [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Waiting for the task: (returnval){ [ 2200.188648] env[62824]: value = "task-2146546" [ 2200.188648] env[62824]: _type = "Task" [ 2200.188648] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2200.197105] env[62824]: DEBUG oslo_vmware.api [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146546, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2200.294670] env[62824]: DEBUG nova.network.neutron [req-4324f5c6-c55c-4b51-9f3d-9623a7782cef req-6cdccbe5-79d1-4859-aa34-0388e39d7c57 service nova] [instance: 0c9fb5cc-de69-46d8-9962-98e0a84e33c3] Updated VIF entry in instance network info cache for port b96f49a1-2e1e-495b-9809-a7338087a686. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2200.295111] env[62824]: DEBUG nova.network.neutron [req-4324f5c6-c55c-4b51-9f3d-9623a7782cef req-6cdccbe5-79d1-4859-aa34-0388e39d7c57 service nova] [instance: 0c9fb5cc-de69-46d8-9962-98e0a84e33c3] Updating instance_info_cache with network_info: [{"id": "b96f49a1-2e1e-495b-9809-a7338087a686", "address": "fa:16:3e:b5:0f:b3", "network": {"id": "9e1885b6-0796-42b9-9cd8-8fa9ce9a3cfd", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-2118936609-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "64042a790d6c459186c68d73da32c019", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "afae6acd-1873-4228-9d5a-1cd5d4efe3e4", "external-id": "nsx-vlan-transportzone-183", "segmentation_id": 183, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb96f49a1-2e", "ovs_interfaceid": "b96f49a1-2e1e-495b-9809-a7338087a686", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2200.324589] env[62824]: DEBUG nova.network.neutron [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412] Successfully updated port: 0f67546a-a44f-4c20-96e8-e7b09334b221 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2200.362111] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3150a869-b311-4209-98d3-70bbccfac01e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2200.370427] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be67d261-f65c-4d02-a698-a74e06699fcd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2200.408798] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b764b84f-b913-49f3-9487-651aae04ab91 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2200.420938] env[62824]: DEBUG oslo_vmware.api [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146545, 'name': Rename_Task, 'duration_secs': 0.162587} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2200.423387] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 8c2ac679-6915-4149-aa51-99170b0df685] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2200.423709] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dd7e3a0c-4748-48a0-93fe-a4a236c96633 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2200.426403] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f7ddee9-590a-437b-8180-1a1e90218b10 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2200.443024] env[62824]: DEBUG nova.compute.provider_tree [None req-f32f5e12-bb47-427a-aceb-0383bc172546 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2200.446553] env[62824]: DEBUG oslo_vmware.rw_handles [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52eed9d3-7100-8b68-b837-4f50f98ea027/disk-0.vmdk. {{(pid=62824) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 2200.448637] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a97e0e21-d30d-4e14-aab6-c80cb8ed6a1b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2200.451541] env[62824]: DEBUG oslo_vmware.api [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2200.451541] env[62824]: value = "task-2146547" [ 2200.451541] env[62824]: _type = "Task" [ 2200.451541] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2200.457706] env[62824]: DEBUG oslo_vmware.rw_handles [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52eed9d3-7100-8b68-b837-4f50f98ea027/disk-0.vmdk is in state: ready. {{(pid=62824) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 2200.457869] env[62824]: ERROR oslo_vmware.rw_handles [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52eed9d3-7100-8b68-b837-4f50f98ea027/disk-0.vmdk due to incomplete transfer. [ 2200.458536] env[62824]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-875cbfe2-2b4e-486b-9bb3-69069f9852ec {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2200.463708] env[62824]: DEBUG oslo_vmware.api [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146547, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2200.470029] env[62824]: DEBUG oslo_vmware.rw_handles [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52eed9d3-7100-8b68-b837-4f50f98ea027/disk-0.vmdk. {{(pid=62824) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 2200.470253] env[62824]: DEBUG nova.virt.vmwareapi.images [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 58243465-ba24-4b37-a640-04ab24963230] Uploaded image 5277bf68-46cb-496f-878c-a6d6908d561b to the Glance image server {{(pid=62824) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 2200.472506] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 58243465-ba24-4b37-a640-04ab24963230] Destroying the VM {{(pid=62824) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 2200.473100] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-6082665c-f7dd-42b1-9272-f52e9e7664ed {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2200.479122] env[62824]: DEBUG oslo_vmware.api [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 2200.479122] env[62824]: value = "task-2146548" [ 2200.479122] env[62824]: _type = "Task" [ 2200.479122] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2200.488031] env[62824]: DEBUG oslo_vmware.api [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146548, 'name': Destroy_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2200.705954] env[62824]: DEBUG oslo_vmware.api [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146546, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.491493} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2200.709278] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 0c9fb5cc-de69-46d8-9962-98e0a84e33c3/0c9fb5cc-de69-46d8-9962-98e0a84e33c3.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2200.709520] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 0c9fb5cc-de69-46d8-9962-98e0a84e33c3] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2200.711836] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-39b39031-1df0-4663-8531-8eaf147b66e3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2200.729136] env[62824]: DEBUG oslo_vmware.api [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Waiting for the task: (returnval){ [ 2200.729136] env[62824]: value = "task-2146549" [ 2200.729136] env[62824]: _type = "Task" [ 2200.729136] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2200.739636] env[62824]: DEBUG oslo_vmware.api [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146549, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2200.798332] env[62824]: DEBUG oslo_concurrency.lockutils [req-4324f5c6-c55c-4b51-9f3d-9623a7782cef req-6cdccbe5-79d1-4859-aa34-0388e39d7c57 service nova] Releasing lock "refresh_cache-0c9fb5cc-de69-46d8-9962-98e0a84e33c3" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2200.829495] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Acquiring lock "refresh_cache-c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2200.829770] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Acquired lock "refresh_cache-c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2200.829817] env[62824]: DEBUG nova.network.neutron [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2200.911413] env[62824]: DEBUG nova.compute.manager [req-ea05da4e-461a-4a89-ba72-d0b8a877d855 req-df2a8d45-cfb8-4416-8620-3905afe29262 service nova] [instance: c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412] Received event network-vif-plugged-0f67546a-a44f-4c20-96e8-e7b09334b221 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2200.912478] env[62824]: DEBUG oslo_concurrency.lockutils [req-ea05da4e-461a-4a89-ba72-d0b8a877d855 req-df2a8d45-cfb8-4416-8620-3905afe29262 service nova] Acquiring lock "c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2200.912478] env[62824]: DEBUG oslo_concurrency.lockutils [req-ea05da4e-461a-4a89-ba72-d0b8a877d855 req-df2a8d45-cfb8-4416-8620-3905afe29262 service nova] Lock "c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2200.912478] env[62824]: DEBUG oslo_concurrency.lockutils [req-ea05da4e-461a-4a89-ba72-d0b8a877d855 req-df2a8d45-cfb8-4416-8620-3905afe29262 service nova] Lock "c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2200.912478] env[62824]: DEBUG nova.compute.manager [req-ea05da4e-461a-4a89-ba72-d0b8a877d855 req-df2a8d45-cfb8-4416-8620-3905afe29262 service nova] [instance: c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412] No waiting events found dispatching network-vif-plugged-0f67546a-a44f-4c20-96e8-e7b09334b221 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2200.912478] env[62824]: WARNING nova.compute.manager [req-ea05da4e-461a-4a89-ba72-d0b8a877d855 req-df2a8d45-cfb8-4416-8620-3905afe29262 service nova] [instance: c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412] Received unexpected event network-vif-plugged-0f67546a-a44f-4c20-96e8-e7b09334b221 for instance with vm_state building and task_state spawning. [ 2200.912663] env[62824]: DEBUG nova.compute.manager [req-ea05da4e-461a-4a89-ba72-d0b8a877d855 req-df2a8d45-cfb8-4416-8620-3905afe29262 service nova] [instance: c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412] Received event network-changed-0f67546a-a44f-4c20-96e8-e7b09334b221 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2200.912663] env[62824]: DEBUG nova.compute.manager [req-ea05da4e-461a-4a89-ba72-d0b8a877d855 req-df2a8d45-cfb8-4416-8620-3905afe29262 service nova] [instance: c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412] Refreshing instance network info cache due to event network-changed-0f67546a-a44f-4c20-96e8-e7b09334b221. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2200.912832] env[62824]: DEBUG oslo_concurrency.lockutils [req-ea05da4e-461a-4a89-ba72-d0b8a877d855 req-df2a8d45-cfb8-4416-8620-3905afe29262 service nova] Acquiring lock "refresh_cache-c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2200.953561] env[62824]: DEBUG nova.scheduler.client.report [None req-f32f5e12-bb47-427a-aceb-0383bc172546 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2200.967391] env[62824]: DEBUG oslo_vmware.api [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146547, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2200.994810] env[62824]: DEBUG oslo_vmware.api [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146548, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2201.215008] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b9a05fc7-ab92-4245-87d2-ec7f1181175f tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Acquiring lock "b9f30f87-3594-4468-9d29-70890d8761e3" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2201.215348] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b9a05fc7-ab92-4245-87d2-ec7f1181175f tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Lock "b9f30f87-3594-4468-9d29-70890d8761e3" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2201.215531] env[62824]: INFO nova.compute.manager [None req-b9a05fc7-ab92-4245-87d2-ec7f1181175f tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] Attaching volume a7248bce-cc5d-47e2-b789-e0f635a1ce45 to /dev/sdb [ 2201.242118] env[62824]: DEBUG oslo_vmware.api [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146549, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.100012} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2201.242395] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 0c9fb5cc-de69-46d8-9962-98e0a84e33c3] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2201.243180] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95607fd4-efcb-4560-bf9b-41a706fcfc2b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2201.265331] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 0c9fb5cc-de69-46d8-9962-98e0a84e33c3] Reconfiguring VM instance instance-00000055 to attach disk [datastore1] 0c9fb5cc-de69-46d8-9962-98e0a84e33c3/0c9fb5cc-de69-46d8-9962-98e0a84e33c3.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2201.266546] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1b857c12-f68d-4a71-bf02-53ae5efdda87 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2201.284015] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fd17149-0c46-4acf-9b81-c774a73b9aed {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2201.291362] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6098651a-5e59-44c0-8c02-677b4444bca1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2201.294728] env[62824]: DEBUG oslo_vmware.api [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Waiting for the task: (returnval){ [ 2201.294728] env[62824]: value = "task-2146550" [ 2201.294728] env[62824]: _type = "Task" [ 2201.294728] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2201.305101] env[62824]: DEBUG oslo_vmware.api [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146550, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2201.306748] env[62824]: DEBUG nova.virt.block_device [None req-b9a05fc7-ab92-4245-87d2-ec7f1181175f tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] Updating existing volume attachment record: 3aee304c-51c9-42a2-b15a-d41d989d48f0 {{(pid=62824) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 2201.391100] env[62824]: DEBUG nova.network.neutron [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2201.462597] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f32f5e12-bb47-427a-aceb-0383bc172546 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.019s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2201.465033] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 19.813s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2201.476542] env[62824]: DEBUG oslo_vmware.api [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146547, 'name': PowerOnVM_Task, 'duration_secs': 0.617851} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2201.476542] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 8c2ac679-6915-4149-aa51-99170b0df685] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2201.476890] env[62824]: INFO nova.compute.manager [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 8c2ac679-6915-4149-aa51-99170b0df685] Took 7.02 seconds to spawn the instance on the hypervisor. [ 2201.477896] env[62824]: DEBUG nova.compute.manager [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 8c2ac679-6915-4149-aa51-99170b0df685] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2201.483086] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96edb4d4-896c-48b7-9474-c7e2c9968ba0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2201.503018] env[62824]: DEBUG oslo_vmware.api [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146548, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2201.504307] env[62824]: INFO nova.scheduler.client.report [None req-f32f5e12-bb47-427a-aceb-0383bc172546 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Deleted allocations for instance 2afd1f18-234c-40b5-9ede-7413ad30dafe [ 2201.576094] env[62824]: DEBUG nova.network.neutron [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412] Updating instance_info_cache with network_info: [{"id": "0f67546a-a44f-4c20-96e8-e7b09334b221", "address": "fa:16:3e:ac:62:ba", "network": {"id": "ae1dfabd-223a-4d4c-a8c4-fe65a83dd9c0", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-797102932-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "102d2f886fd44e1da6780e48e36f1b62", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "517421c3-bea0-419c-ab0b-987815e5d160", "external-id": "nsx-vlan-transportzone-68", "segmentation_id": 68, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0f67546a-a4", "ovs_interfaceid": "0f67546a-a44f-4c20-96e8-e7b09334b221", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2201.806902] env[62824]: DEBUG oslo_vmware.api [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146550, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2201.999637] env[62824]: DEBUG oslo_vmware.api [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146548, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2202.013122] env[62824]: INFO nova.compute.manager [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 8c2ac679-6915-4149-aa51-99170b0df685] Took 32.96 seconds to build instance. [ 2202.015674] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f32f5e12-bb47-427a-aceb-0383bc172546 tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Lock "2afd1f18-234c-40b5-9ede-7413ad30dafe" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.728s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2202.078424] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Releasing lock "refresh_cache-c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2202.078838] env[62824]: DEBUG nova.compute.manager [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412] Instance network_info: |[{"id": "0f67546a-a44f-4c20-96e8-e7b09334b221", "address": "fa:16:3e:ac:62:ba", "network": {"id": "ae1dfabd-223a-4d4c-a8c4-fe65a83dd9c0", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-797102932-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "102d2f886fd44e1da6780e48e36f1b62", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "517421c3-bea0-419c-ab0b-987815e5d160", "external-id": "nsx-vlan-transportzone-68", "segmentation_id": 68, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0f67546a-a4", "ovs_interfaceid": "0f67546a-a44f-4c20-96e8-e7b09334b221", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2202.079169] env[62824]: DEBUG oslo_concurrency.lockutils [req-ea05da4e-461a-4a89-ba72-d0b8a877d855 req-df2a8d45-cfb8-4416-8620-3905afe29262 service nova] Acquired lock "refresh_cache-c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2202.079436] env[62824]: DEBUG nova.network.neutron [req-ea05da4e-461a-4a89-ba72-d0b8a877d855 req-df2a8d45-cfb8-4416-8620-3905afe29262 service nova] [instance: c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412] Refreshing network info cache for port 0f67546a-a44f-4c20-96e8-e7b09334b221 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2202.080991] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ac:62:ba', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '517421c3-bea0-419c-ab0b-987815e5d160', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0f67546a-a44f-4c20-96e8-e7b09334b221', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2202.094115] env[62824]: DEBUG oslo.service.loopingcall [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2202.095614] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2202.095926] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6680053a-d3a9-4f55-9692-feebb2041d37 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2202.118622] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2202.118622] env[62824]: value = "task-2146554" [ 2202.118622] env[62824]: _type = "Task" [ 2202.118622] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2202.127531] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146554, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2202.309931] env[62824]: DEBUG oslo_vmware.api [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146550, 'name': ReconfigVM_Task, 'duration_secs': 0.858287} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2202.310334] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 0c9fb5cc-de69-46d8-9962-98e0a84e33c3] Reconfigured VM instance instance-00000055 to attach disk [datastore1] 0c9fb5cc-de69-46d8-9962-98e0a84e33c3/0c9fb5cc-de69-46d8-9962-98e0a84e33c3.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2202.311037] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1f40b595-741c-400b-919e-3251f4b88ea4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2202.317739] env[62824]: DEBUG oslo_vmware.api [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Waiting for the task: (returnval){ [ 2202.317739] env[62824]: value = "task-2146555" [ 2202.317739] env[62824]: _type = "Task" [ 2202.317739] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2202.326245] env[62824]: DEBUG oslo_vmware.api [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146555, 'name': Rename_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2202.498223] env[62824]: DEBUG oslo_vmware.api [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146548, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2202.511047] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 07ce931f-45ef-409b-b714-9f1cd47a3a88 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2202.511493] env[62824]: WARNING nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance f3740d9a-b21a-4a79-9e28-2a89ecd08bb6 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 2202.511702] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance b3003c4b-ae5a-48df-8c12-a915a76253f4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2202.511866] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 6d80ec06-8559-4964-8577-a2512aa366ed actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2202.512073] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance b9f30f87-3594-4468-9d29-70890d8761e3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2202.512247] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance bbfcb2e3-9326-4548-b15b-e054cbfd192e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2202.512416] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 2f315f9e-94e4-47ef-9503-ee92b59e5452 actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2202.512578] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 15479062-af75-4925-99b3-77d6a49751ad actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 2, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2202.512727] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance a6cd2032-de60-4f78-bf1e-79801d049df0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2202.512896] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2202.513065] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 350cfd12-a8d5-4a5f-b3de-d795fa179dfd actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2202.513385] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance ca42fef0-1e90-4ab6-9d60-8ef7e4997884 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2202.513385] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 5b3df31f-15fe-473f-992c-ddb272661c53 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2202.513557] env[62824]: WARNING nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 4c2c2068-664d-404f-a99e-8fc7719f43e5 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 2202.513703] env[62824]: WARNING nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance d067fa97-cedc-4e3d-9be4-d860a79a7723 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 2202.513822] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance fc77c44a-180c-46ff-9690-9072c6213c91 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2202.513948] env[62824]: WARNING nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance da1ad739-b252-4e29-a22a-ab3bcab173ec is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 2202.514080] env[62824]: WARNING nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance da46d1d5-f75f-4ef1-b571-fbebab89c2a6 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 2202.514197] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 58243465-ba24-4b37-a640-04ab24963230 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2202.514309] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 8c2ac679-6915-4149-aa51-99170b0df685 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2202.514421] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 0c9fb5cc-de69-46d8-9962-98e0a84e33c3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2202.514530] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2202.516988] env[62824]: DEBUG oslo_concurrency.lockutils [None req-59998aad-f802-4540-9de9-451add892a2b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "8c2ac679-6915-4149-aa51-99170b0df685" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.474s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2202.628932] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146554, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2202.827634] env[62824]: DEBUG oslo_vmware.api [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146555, 'name': Rename_Task, 'duration_secs': 0.15455} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2202.831020] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 0c9fb5cc-de69-46d8-9962-98e0a84e33c3] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2202.831020] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c2b551ed-c1b6-46fb-bb80-5327b8df7d90 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2202.834091] env[62824]: DEBUG oslo_vmware.api [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Waiting for the task: (returnval){ [ 2202.834091] env[62824]: value = "task-2146556" [ 2202.834091] env[62824]: _type = "Task" [ 2202.834091] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2202.841873] env[62824]: DEBUG oslo_vmware.api [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146556, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2202.936065] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e69cca00-b4f4-4019-be10-e9113c3b4c53 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquiring lock "8c2ac679-6915-4149-aa51-99170b0df685" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2202.936516] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e69cca00-b4f4-4019-be10-e9113c3b4c53 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "8c2ac679-6915-4149-aa51-99170b0df685" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2202.938448] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e69cca00-b4f4-4019-be10-e9113c3b4c53 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquiring lock "8c2ac679-6915-4149-aa51-99170b0df685-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2202.938448] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e69cca00-b4f4-4019-be10-e9113c3b4c53 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "8c2ac679-6915-4149-aa51-99170b0df685-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2202.938448] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e69cca00-b4f4-4019-be10-e9113c3b4c53 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "8c2ac679-6915-4149-aa51-99170b0df685-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2202.939343] env[62824]: INFO nova.compute.manager [None req-e69cca00-b4f4-4019-be10-e9113c3b4c53 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 8c2ac679-6915-4149-aa51-99170b0df685] Terminating instance [ 2202.989574] env[62824]: DEBUG nova.network.neutron [req-ea05da4e-461a-4a89-ba72-d0b8a877d855 req-df2a8d45-cfb8-4416-8620-3905afe29262 service nova] [instance: c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412] Updated VIF entry in instance network info cache for port 0f67546a-a44f-4c20-96e8-e7b09334b221. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2202.989991] env[62824]: DEBUG nova.network.neutron [req-ea05da4e-461a-4a89-ba72-d0b8a877d855 req-df2a8d45-cfb8-4416-8620-3905afe29262 service nova] [instance: c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412] Updating instance_info_cache with network_info: [{"id": "0f67546a-a44f-4c20-96e8-e7b09334b221", "address": "fa:16:3e:ac:62:ba", "network": {"id": "ae1dfabd-223a-4d4c-a8c4-fe65a83dd9c0", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-797102932-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "102d2f886fd44e1da6780e48e36f1b62", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "517421c3-bea0-419c-ab0b-987815e5d160", "external-id": "nsx-vlan-transportzone-68", "segmentation_id": 68, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0f67546a-a4", "ovs_interfaceid": "0f67546a-a44f-4c20-96e8-e7b09334b221", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2203.001434] env[62824]: DEBUG oslo_vmware.api [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146548, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2203.017525] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 1f162a63-8773-4c34-9c84-b2de1d5e9f50 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 2203.129924] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146554, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2203.344018] env[62824]: DEBUG oslo_vmware.api [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146556, 'name': PowerOnVM_Task} progress is 81%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2203.443715] env[62824]: DEBUG nova.compute.manager [None req-e69cca00-b4f4-4019-be10-e9113c3b4c53 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 8c2ac679-6915-4149-aa51-99170b0df685] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2203.443715] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e69cca00-b4f4-4019-be10-e9113c3b4c53 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 8c2ac679-6915-4149-aa51-99170b0df685] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2203.444248] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5c003a5-97ab-46b7-82a8-b35bcb26a959 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2203.452659] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-e69cca00-b4f4-4019-be10-e9113c3b4c53 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 8c2ac679-6915-4149-aa51-99170b0df685] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2203.452799] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a402083d-2699-4883-8d6c-43b94f19a069 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2203.459206] env[62824]: DEBUG oslo_vmware.api [None req-e69cca00-b4f4-4019-be10-e9113c3b4c53 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2203.459206] env[62824]: value = "task-2146557" [ 2203.459206] env[62824]: _type = "Task" [ 2203.459206] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2203.466944] env[62824]: DEBUG oslo_vmware.api [None req-e69cca00-b4f4-4019-be10-e9113c3b4c53 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146557, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2203.496585] env[62824]: DEBUG oslo_concurrency.lockutils [req-ea05da4e-461a-4a89-ba72-d0b8a877d855 req-df2a8d45-cfb8-4416-8620-3905afe29262 service nova] Releasing lock "refresh_cache-c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2203.502296] env[62824]: DEBUG oslo_vmware.api [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146548, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2203.521321] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 7c44d956-53d1-4a5f-9016-c0f3f0f348b4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 2203.521559] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Total usable vcpus: 48, total allocated vcpus: 17 {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 2203.521708] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3776MB phys_disk=200GB used_disk=17GB total_vcpus=48 used_vcpus=17 pci_stats=[] {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 2203.632488] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146554, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2203.635072] env[62824]: DEBUG oslo_concurrency.lockutils [None req-602cdfe0-df66-45cb-bc5a-cc85cb34718d tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Acquiring lock "350cfd12-a8d5-4a5f-b3de-d795fa179dfd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2203.635320] env[62824]: DEBUG oslo_concurrency.lockutils [None req-602cdfe0-df66-45cb-bc5a-cc85cb34718d tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Lock "350cfd12-a8d5-4a5f-b3de-d795fa179dfd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2203.635563] env[62824]: DEBUG oslo_concurrency.lockutils [None req-602cdfe0-df66-45cb-bc5a-cc85cb34718d tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Acquiring lock "350cfd12-a8d5-4a5f-b3de-d795fa179dfd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2203.635781] env[62824]: DEBUG oslo_concurrency.lockutils [None req-602cdfe0-df66-45cb-bc5a-cc85cb34718d tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Lock "350cfd12-a8d5-4a5f-b3de-d795fa179dfd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2203.636109] env[62824]: DEBUG oslo_concurrency.lockutils [None req-602cdfe0-df66-45cb-bc5a-cc85cb34718d tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Lock "350cfd12-a8d5-4a5f-b3de-d795fa179dfd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2203.638608] env[62824]: INFO nova.compute.manager [None req-602cdfe0-df66-45cb-bc5a-cc85cb34718d tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 350cfd12-a8d5-4a5f-b3de-d795fa179dfd] Terminating instance [ 2203.816390] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9db4bbfc-ab2a-49ac-920d-03ed5f9cba82 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2203.823820] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53a69371-048f-48bf-bfe4-c66d9d299599 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2203.855999] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8107c0fb-bfae-43d9-be60-f286d37acedd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2203.863217] env[62824]: DEBUG oslo_vmware.api [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146556, 'name': PowerOnVM_Task, 'duration_secs': 0.856552} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2203.865216] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 0c9fb5cc-de69-46d8-9962-98e0a84e33c3] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2203.865426] env[62824]: INFO nova.compute.manager [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 0c9fb5cc-de69-46d8-9962-98e0a84e33c3] Took 6.88 seconds to spawn the instance on the hypervisor. [ 2203.865608] env[62824]: DEBUG nova.compute.manager [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 0c9fb5cc-de69-46d8-9962-98e0a84e33c3] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2203.866401] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97a53465-e963-4972-b02f-d7f5f3700491 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2203.869657] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1031f0db-0cc7-419a-a8e5-5bb3bcd63466 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2203.888572] env[62824]: DEBUG nova.compute.provider_tree [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2203.968630] env[62824]: DEBUG oslo_vmware.api [None req-e69cca00-b4f4-4019-be10-e9113c3b4c53 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146557, 'name': PowerOffVM_Task, 'duration_secs': 0.200358} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2203.968914] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-e69cca00-b4f4-4019-be10-e9113c3b4c53 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 8c2ac679-6915-4149-aa51-99170b0df685] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2203.969093] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e69cca00-b4f4-4019-be10-e9113c3b4c53 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 8c2ac679-6915-4149-aa51-99170b0df685] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2203.969355] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ad564f82-7584-46a0-98aa-ffd367a5fccb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2204.002011] env[62824]: DEBUG oslo_vmware.api [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146548, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2204.131430] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146554, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2204.142835] env[62824]: DEBUG nova.compute.manager [None req-602cdfe0-df66-45cb-bc5a-cc85cb34718d tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 350cfd12-a8d5-4a5f-b3de-d795fa179dfd] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2204.143061] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-602cdfe0-df66-45cb-bc5a-cc85cb34718d tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 350cfd12-a8d5-4a5f-b3de-d795fa179dfd] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2204.143856] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a2aefcb-4199-4e9d-b10d-7695aeb65f7b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2204.151880] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-602cdfe0-df66-45cb-bc5a-cc85cb34718d tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 350cfd12-a8d5-4a5f-b3de-d795fa179dfd] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2204.152047] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ae79d1d7-65ec-4be6-89d1-683009fcab0e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2204.158196] env[62824]: DEBUG oslo_vmware.api [None req-602cdfe0-df66-45cb-bc5a-cc85cb34718d tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Waiting for the task: (returnval){ [ 2204.158196] env[62824]: value = "task-2146560" [ 2204.158196] env[62824]: _type = "Task" [ 2204.158196] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2204.166079] env[62824]: DEBUG oslo_vmware.api [None req-602cdfe0-df66-45cb-bc5a-cc85cb34718d tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Task: {'id': task-2146560, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2204.395481] env[62824]: DEBUG nova.scheduler.client.report [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2204.400422] env[62824]: INFO nova.compute.manager [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 0c9fb5cc-de69-46d8-9962-98e0a84e33c3] Took 33.45 seconds to build instance. [ 2204.503587] env[62824]: DEBUG oslo_vmware.api [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146548, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2204.616912] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d85dcf93-e801-4553-b0ef-b90b42fed644 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Acquiring lock "0c9fb5cc-de69-46d8-9962-98e0a84e33c3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2204.631749] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146554, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2204.667780] env[62824]: DEBUG oslo_vmware.api [None req-602cdfe0-df66-45cb-bc5a-cc85cb34718d tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Task: {'id': task-2146560, 'name': PowerOffVM_Task, 'duration_secs': 0.217926} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2204.668069] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-602cdfe0-df66-45cb-bc5a-cc85cb34718d tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 350cfd12-a8d5-4a5f-b3de-d795fa179dfd] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2204.668304] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-602cdfe0-df66-45cb-bc5a-cc85cb34718d tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 350cfd12-a8d5-4a5f-b3de-d795fa179dfd] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2204.668568] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-99eba165-184a-4fb1-af7a-bd930ca37e7c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2204.901886] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62824) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 2204.902140] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.437s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2204.902546] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c0eb605e-eace-46aa-b2f3-123c52a995db tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Lock "0c9fb5cc-de69-46d8-9962-98e0a84e33c3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.959s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2204.902791] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0235e5cb-af08-4951-9315-366312405d6f tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.399s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2204.902973] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0235e5cb-af08-4951-9315-366312405d6f tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2204.905033] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0c3b6169-674a-4946-8b27-bab89f3484d0 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.026s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2204.905228] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0c3b6169-674a-4946-8b27-bab89f3484d0 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2204.906953] env[62824]: DEBUG oslo_concurrency.lockutils [None req-cac6081d-dcbb-4024-bf06-8c5d7fd6c54c tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.747s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2204.907155] env[62824]: DEBUG oslo_concurrency.lockutils [None req-cac6081d-dcbb-4024-bf06-8c5d7fd6c54c tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2204.908730] env[62824]: DEBUG oslo_concurrency.lockutils [None req-32333306-ced2-48b5-bba6-b899d39c55fe tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.563s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2204.908874] env[62824]: DEBUG oslo_concurrency.lockutils [None req-32333306-ced2-48b5-bba6-b899d39c55fe tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2204.910468] env[62824]: DEBUG oslo_concurrency.lockutils [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.769s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2204.912248] env[62824]: INFO nova.compute.claims [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 1f162a63-8773-4c34-9c84-b2de1d5e9f50] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2204.916292] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d85dcf93-e801-4553-b0ef-b90b42fed644 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Lock "0c9fb5cc-de69-46d8-9962-98e0a84e33c3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.299s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2204.916292] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d85dcf93-e801-4553-b0ef-b90b42fed644 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Acquiring lock "0c9fb5cc-de69-46d8-9962-98e0a84e33c3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2204.916557] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d85dcf93-e801-4553-b0ef-b90b42fed644 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Lock "0c9fb5cc-de69-46d8-9962-98e0a84e33c3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2204.916557] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d85dcf93-e801-4553-b0ef-b90b42fed644 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Lock "0c9fb5cc-de69-46d8-9962-98e0a84e33c3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2204.918710] env[62824]: INFO nova.compute.manager [None req-d85dcf93-e801-4553-b0ef-b90b42fed644 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 0c9fb5cc-de69-46d8-9962-98e0a84e33c3] Terminating instance [ 2204.935769] env[62824]: INFO nova.scheduler.client.report [None req-32333306-ced2-48b5-bba6-b899d39c55fe tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Deleted allocations for instance f3740d9a-b21a-4a79-9e28-2a89ecd08bb6 [ 2204.937245] env[62824]: INFO nova.scheduler.client.report [None req-cac6081d-dcbb-4024-bf06-8c5d7fd6c54c tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Deleted allocations for instance da1ad739-b252-4e29-a22a-ab3bcab173ec [ 2204.953542] env[62824]: INFO nova.scheduler.client.report [None req-0c3b6169-674a-4946-8b27-bab89f3484d0 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Deleted allocations for instance d067fa97-cedc-4e3d-9be4-d860a79a7723 [ 2204.957853] env[62824]: INFO nova.scheduler.client.report [None req-0235e5cb-af08-4951-9315-366312405d6f tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Deleted allocations for instance 4c2c2068-664d-404f-a99e-8fc7719f43e5 [ 2205.003683] env[62824]: DEBUG oslo_vmware.api [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146548, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2205.131308] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146554, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2205.421732] env[62824]: DEBUG nova.compute.manager [None req-d85dcf93-e801-4553-b0ef-b90b42fed644 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 0c9fb5cc-de69-46d8-9962-98e0a84e33c3] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2205.422368] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-d85dcf93-e801-4553-b0ef-b90b42fed644 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 0c9fb5cc-de69-46d8-9962-98e0a84e33c3] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2205.423329] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-567d5e3e-2131-4d7f-80f1-aa03939ce774 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2205.431367] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d85dcf93-e801-4553-b0ef-b90b42fed644 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 0c9fb5cc-de69-46d8-9962-98e0a84e33c3] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2205.431604] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-98ee353c-aa10-49ce-a627-ad6b0f2ce03d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2205.439701] env[62824]: DEBUG oslo_vmware.api [None req-d85dcf93-e801-4553-b0ef-b90b42fed644 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Waiting for the task: (returnval){ [ 2205.439701] env[62824]: value = "task-2146562" [ 2205.439701] env[62824]: _type = "Task" [ 2205.439701] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2205.451909] env[62824]: DEBUG oslo_vmware.api [None req-d85dcf93-e801-4553-b0ef-b90b42fed644 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146562, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2205.452346] env[62824]: DEBUG oslo_concurrency.lockutils [None req-32333306-ced2-48b5-bba6-b899d39c55fe tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Lock "f3740d9a-b21a-4a79-9e28-2a89ecd08bb6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.787s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2205.453327] env[62824]: DEBUG oslo_concurrency.lockutils [None req-cac6081d-dcbb-4024-bf06-8c5d7fd6c54c tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Lock "da1ad739-b252-4e29-a22a-ab3bcab173ec" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.997s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2205.460019] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0c3b6169-674a-4946-8b27-bab89f3484d0 tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Lock "d067fa97-cedc-4e3d-9be4-d860a79a7723" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.735s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2205.463164] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0235e5cb-af08-4951-9315-366312405d6f tempest-ListServersNegativeTestJSON-1160681400 tempest-ListServersNegativeTestJSON-1160681400-project-member] Lock "4c2c2068-664d-404f-a99e-8fc7719f43e5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.632s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2205.504318] env[62824]: DEBUG oslo_vmware.api [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146548, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2205.633404] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146554, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2205.903501] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f8f74f45-c68b-4875-9f0f-fc836d564d52 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Acquiring lock "1870b82a-9783-44ac-8de2-7b9ffc2a1bc8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2205.903749] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f8f74f45-c68b-4875-9f0f-fc836d564d52 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Lock "1870b82a-9783-44ac-8de2-7b9ffc2a1bc8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2205.903956] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f8f74f45-c68b-4875-9f0f-fc836d564d52 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Acquiring lock "1870b82a-9783-44ac-8de2-7b9ffc2a1bc8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2205.904203] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f8f74f45-c68b-4875-9f0f-fc836d564d52 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Lock "1870b82a-9783-44ac-8de2-7b9ffc2a1bc8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2205.904367] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f8f74f45-c68b-4875-9f0f-fc836d564d52 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Lock "1870b82a-9783-44ac-8de2-7b9ffc2a1bc8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2205.910550] env[62824]: DEBUG oslo_concurrency.lockutils [None req-56a95388-af07-48c0-bd95-c65c098831e2 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Acquiring lock "07ce931f-45ef-409b-b714-9f1cd47a3a88" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2205.910550] env[62824]: DEBUG oslo_concurrency.lockutils [None req-56a95388-af07-48c0-bd95-c65c098831e2 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Lock "07ce931f-45ef-409b-b714-9f1cd47a3a88" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2205.910550] env[62824]: DEBUG oslo_concurrency.lockutils [None req-56a95388-af07-48c0-bd95-c65c098831e2 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Acquiring lock "07ce931f-45ef-409b-b714-9f1cd47a3a88-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2205.910550] env[62824]: DEBUG oslo_concurrency.lockutils [None req-56a95388-af07-48c0-bd95-c65c098831e2 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Lock "07ce931f-45ef-409b-b714-9f1cd47a3a88-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2205.910550] env[62824]: DEBUG oslo_concurrency.lockutils [None req-56a95388-af07-48c0-bd95-c65c098831e2 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Lock "07ce931f-45ef-409b-b714-9f1cd47a3a88-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2205.911896] env[62824]: INFO nova.compute.manager [None req-f8f74f45-c68b-4875-9f0f-fc836d564d52 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8] Terminating instance [ 2205.913776] env[62824]: INFO nova.compute.manager [None req-56a95388-af07-48c0-bd95-c65c098831e2 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Terminating instance [ 2205.951382] env[62824]: DEBUG oslo_vmware.api [None req-d85dcf93-e801-4553-b0ef-b90b42fed644 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146562, 'name': PowerOffVM_Task, 'duration_secs': 0.169295} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2205.954084] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d85dcf93-e801-4553-b0ef-b90b42fed644 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 0c9fb5cc-de69-46d8-9962-98e0a84e33c3] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2205.954276] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-d85dcf93-e801-4553-b0ef-b90b42fed644 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 0c9fb5cc-de69-46d8-9962-98e0a84e33c3] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2205.955126] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-68de7589-5450-467c-b8d5-008537944e57 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2206.008810] env[62824]: DEBUG oslo_vmware.api [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146548, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2206.134328] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146554, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2206.197467] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd9555c2-3fe5-4b9f-a0c6-d24de833f113 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2206.205182] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac80d6e2-c213-45ad-99e7-089fbb00a0c4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2206.236143] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8506c8fa-ad94-4790-a624-f2f2f2e1768e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2206.243914] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c44cae58-bf3a-477a-953d-b127d1103d91 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2206.257982] env[62824]: DEBUG nova.compute.provider_tree [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2206.372276] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-b9a05fc7-ab92-4245-87d2-ec7f1181175f tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] Volume attach. Driver type: vmdk {{(pid=62824) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 2206.372533] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-b9a05fc7-ab92-4245-87d2-ec7f1181175f tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-438718', 'volume_id': 'a7248bce-cc5d-47e2-b789-e0f635a1ce45', 'name': 'volume-a7248bce-cc5d-47e2-b789-e0f635a1ce45', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'b9f30f87-3594-4468-9d29-70890d8761e3', 'attached_at': '', 'detached_at': '', 'volume_id': 'a7248bce-cc5d-47e2-b789-e0f635a1ce45', 'serial': 'a7248bce-cc5d-47e2-b789-e0f635a1ce45'} {{(pid=62824) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 2206.373452] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25af230b-8a68-402c-967f-e652bb04babd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2206.390215] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c4dcf0a-6f6e-4a83-ab2a-5adcc536b21a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2206.416509] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-b9a05fc7-ab92-4245-87d2-ec7f1181175f tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] Reconfiguring VM instance instance-0000002f to attach disk [datastore1] volume-a7248bce-cc5d-47e2-b789-e0f635a1ce45/volume-a7248bce-cc5d-47e2-b789-e0f635a1ce45.vmdk or device None with type thin {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2206.419203] env[62824]: DEBUG nova.compute.manager [None req-f8f74f45-c68b-4875-9f0f-fc836d564d52 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2206.419203] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-f8f74f45-c68b-4875-9f0f-fc836d564d52 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2206.419203] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8359fed2-72b0-4457-be7f-d655ff221155 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2206.430699] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cd16e0f-985b-4817-81da-d05907351c75 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2206.433995] env[62824]: DEBUG nova.compute.manager [None req-56a95388-af07-48c0-bd95-c65c098831e2 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2206.434206] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-56a95388-af07-48c0-bd95-c65c098831e2 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2206.434930] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-116e4e1f-6707-4c5f-be93-548011634d7c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2206.442505] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8f74f45-c68b-4875-9f0f-fc836d564d52 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2206.443592] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6e69c086-1059-4eb9-9e11-1b6a2bb98836 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2206.444964] env[62824]: DEBUG oslo_vmware.api [None req-b9a05fc7-ab92-4245-87d2-ec7f1181175f tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Waiting for the task: (returnval){ [ 2206.444964] env[62824]: value = "task-2146564" [ 2206.444964] env[62824]: _type = "Task" [ 2206.444964] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2206.451033] env[62824]: DEBUG oslo_vmware.api [None req-f8f74f45-c68b-4875-9f0f-fc836d564d52 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the task: (returnval){ [ 2206.451033] env[62824]: value = "task-2146565" [ 2206.451033] env[62824]: _type = "Task" [ 2206.451033] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2206.460635] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-56a95388-af07-48c0-bd95-c65c098831e2 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2206.460635] env[62824]: DEBUG oslo_vmware.api [None req-b9a05fc7-ab92-4245-87d2-ec7f1181175f tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Task: {'id': task-2146564, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2206.460635] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-28b01693-378b-461c-bf33-493f59ba5680 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2206.465347] env[62824]: DEBUG oslo_vmware.api [None req-f8f74f45-c68b-4875-9f0f-fc836d564d52 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146565, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2206.470578] env[62824]: DEBUG oslo_vmware.api [None req-56a95388-af07-48c0-bd95-c65c098831e2 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for the task: (returnval){ [ 2206.470578] env[62824]: value = "task-2146566" [ 2206.470578] env[62824]: _type = "Task" [ 2206.470578] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2206.479207] env[62824]: DEBUG oslo_vmware.api [None req-56a95388-af07-48c0-bd95-c65c098831e2 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146566, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2206.505242] env[62824]: DEBUG oslo_vmware.api [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146548, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2206.580452] env[62824]: DEBUG oslo_concurrency.lockutils [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Acquiring lock "853f3098-881c-402c-8fd7-26540d1ecd88" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2206.580753] env[62824]: DEBUG oslo_concurrency.lockutils [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Lock "853f3098-881c-402c-8fd7-26540d1ecd88" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2206.636279] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146554, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2206.761169] env[62824]: DEBUG nova.scheduler.client.report [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2206.958552] env[62824]: DEBUG oslo_vmware.api [None req-b9a05fc7-ab92-4245-87d2-ec7f1181175f tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Task: {'id': task-2146564, 'name': ReconfigVM_Task, 'duration_secs': 0.365154} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2206.962011] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-b9a05fc7-ab92-4245-87d2-ec7f1181175f tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] Reconfigured VM instance instance-0000002f to attach disk [datastore1] volume-a7248bce-cc5d-47e2-b789-e0f635a1ce45/volume-a7248bce-cc5d-47e2-b789-e0f635a1ce45.vmdk or device None with type thin {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2206.971016] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bc048581-d0ea-43f4-9033-ac77a4d0c93d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2206.985853] env[62824]: DEBUG oslo_vmware.api [None req-f8f74f45-c68b-4875-9f0f-fc836d564d52 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146565, 'name': PowerOffVM_Task, 'duration_secs': 0.216964} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2206.990511] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8f74f45-c68b-4875-9f0f-fc836d564d52 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2206.991584] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-f8f74f45-c68b-4875-9f0f-fc836d564d52 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2206.991665] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-97c44ee2-5906-4e19-b701-7130e7eb6bcf {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2206.997273] env[62824]: DEBUG oslo_vmware.api [None req-b9a05fc7-ab92-4245-87d2-ec7f1181175f tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Waiting for the task: (returnval){ [ 2206.997273] env[62824]: value = "task-2146567" [ 2206.997273] env[62824]: _type = "Task" [ 2206.997273] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2207.004013] env[62824]: DEBUG oslo_vmware.api [None req-56a95388-af07-48c0-bd95-c65c098831e2 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146566, 'name': PowerOffVM_Task, 'duration_secs': 0.17616} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2207.007318] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-56a95388-af07-48c0-bd95-c65c098831e2 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2207.007546] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-56a95388-af07-48c0-bd95-c65c098831e2 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2207.008505] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-335ac732-78ba-4e72-adab-4296d4a5322f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2207.015005] env[62824]: DEBUG oslo_vmware.api [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146548, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2207.019618] env[62824]: DEBUG oslo_vmware.api [None req-b9a05fc7-ab92-4245-87d2-ec7f1181175f tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Task: {'id': task-2146567, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2207.083291] env[62824]: DEBUG nova.compute.manager [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 853f3098-881c-402c-8fd7-26540d1ecd88] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2207.141030] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146554, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2207.268366] env[62824]: DEBUG oslo_concurrency.lockutils [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.358s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2207.269674] env[62824]: DEBUG nova.compute.manager [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 1f162a63-8773-4c34-9c84-b2de1d5e9f50] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2207.272935] env[62824]: DEBUG oslo_concurrency.lockutils [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.102s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2207.274378] env[62824]: INFO nova.compute.claims [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7c44d956-53d1-4a5f-9016-c0f3f0f348b4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2207.507557] env[62824]: DEBUG oslo_vmware.api [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146548, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2207.515271] env[62824]: DEBUG oslo_vmware.api [None req-b9a05fc7-ab92-4245-87d2-ec7f1181175f tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Task: {'id': task-2146567, 'name': ReconfigVM_Task, 'duration_secs': 0.165356} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2207.515654] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-b9a05fc7-ab92-4245-87d2-ec7f1181175f tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-438718', 'volume_id': 'a7248bce-cc5d-47e2-b789-e0f635a1ce45', 'name': 'volume-a7248bce-cc5d-47e2-b789-e0f635a1ce45', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'b9f30f87-3594-4468-9d29-70890d8761e3', 'attached_at': '', 'detached_at': '', 'volume_id': 'a7248bce-cc5d-47e2-b789-e0f635a1ce45', 'serial': 'a7248bce-cc5d-47e2-b789-e0f635a1ce45'} {{(pid=62824) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 2207.609281] env[62824]: DEBUG oslo_concurrency.lockutils [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2207.638070] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146554, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2207.774885] env[62824]: DEBUG nova.compute.utils [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2207.776461] env[62824]: DEBUG nova.compute.manager [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 1f162a63-8773-4c34-9c84-b2de1d5e9f50] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2207.776633] env[62824]: DEBUG nova.network.neutron [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 1f162a63-8773-4c34-9c84-b2de1d5e9f50] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2207.853114] env[62824]: DEBUG nova.policy [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8c6cd8cd9ac848c89218208d1c2290aa', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '461117110de7467fbf8ada5ee9978069', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2208.015547] env[62824]: DEBUG oslo_vmware.api [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146548, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2208.140046] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146554, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2208.224552] env[62824]: DEBUG nova.network.neutron [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 1f162a63-8773-4c34-9c84-b2de1d5e9f50] Successfully created port: f5c2ffd1-3953-40e4-8d96-0542389767da {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2208.279744] env[62824]: DEBUG nova.compute.manager [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 1f162a63-8773-4c34-9c84-b2de1d5e9f50] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2208.514763] env[62824]: DEBUG oslo_vmware.api [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146548, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2208.557504] env[62824]: DEBUG nova.objects.instance [None req-b9a05fc7-ab92-4245-87d2-ec7f1181175f tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Lazy-loading 'flavor' on Instance uuid b9f30f87-3594-4468-9d29-70890d8761e3 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2208.576465] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3caf10a-3161-4ad7-9bc3-43241373c8fb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2208.584959] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b20643bc-13e2-46bb-9fe0-5709c7a81f1f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2208.617822] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64325f46-15b0-4d66-9b69-95fdbcf97950 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2208.625624] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa2c12e4-248d-44f6-9434-10cee5a3cd0e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2208.641184] env[62824]: DEBUG nova.compute.provider_tree [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2208.649859] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146554, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2209.015021] env[62824]: DEBUG oslo_vmware.api [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146548, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2209.063105] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b9a05fc7-ab92-4245-87d2-ec7f1181175f tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Lock "b9f30f87-3594-4468-9d29-70890d8761e3" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.848s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2209.139978] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146554, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2209.144262] env[62824]: DEBUG nova.scheduler.client.report [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2209.289924] env[62824]: DEBUG nova.compute.manager [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 1f162a63-8773-4c34-9c84-b2de1d5e9f50] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2209.314883] env[62824]: DEBUG nova.virt.hardware [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2209.315130] env[62824]: DEBUG nova.virt.hardware [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2209.315298] env[62824]: DEBUG nova.virt.hardware [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2209.315485] env[62824]: DEBUG nova.virt.hardware [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2209.315637] env[62824]: DEBUG nova.virt.hardware [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2209.315788] env[62824]: DEBUG nova.virt.hardware [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2209.315996] env[62824]: DEBUG nova.virt.hardware [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2209.316174] env[62824]: DEBUG nova.virt.hardware [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2209.316345] env[62824]: DEBUG nova.virt.hardware [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2209.316514] env[62824]: DEBUG nova.virt.hardware [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2209.316692] env[62824]: DEBUG nova.virt.hardware [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2209.317580] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f308d43d-be11-4cec-8d9b-31e95c3180fa {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2209.325655] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c54080e-c526-4d16-aa31-cd4230a88e62 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2209.512056] env[62824]: DEBUG oslo_vmware.api [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146548, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2209.641445] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146554, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2209.649360] env[62824]: DEBUG oslo_concurrency.lockutils [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.376s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2209.649893] env[62824]: DEBUG nova.compute.manager [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7c44d956-53d1-4a5f-9016-c0f3f0f348b4] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2209.652415] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7e6f95cb-45a7-48de-8220-fb8355af6eb3 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.703s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2209.652611] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7e6f95cb-45a7-48de-8220-fb8355af6eb3 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2209.654612] env[62824]: DEBUG oslo_concurrency.lockutils [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.046s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2209.656034] env[62824]: INFO nova.compute.claims [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 853f3098-881c-402c-8fd7-26540d1ecd88] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2209.681744] env[62824]: INFO nova.scheduler.client.report [None req-7e6f95cb-45a7-48de-8220-fb8355af6eb3 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Deleted allocations for instance da46d1d5-f75f-4ef1-b571-fbebab89c2a6 [ 2209.915595] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7cf9f153-360a-4432-94f9-b1790d7b2e62 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Acquiring lock "b9f30f87-3594-4468-9d29-70890d8761e3" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2209.915890] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7cf9f153-360a-4432-94f9-b1790d7b2e62 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Lock "b9f30f87-3594-4468-9d29-70890d8761e3" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2210.015398] env[62824]: DEBUG oslo_vmware.api [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146548, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2210.141393] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146554, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2210.142603] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e69cca00-b4f4-4019-be10-e9113c3b4c53 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 8c2ac679-6915-4149-aa51-99170b0df685] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2210.142806] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e69cca00-b4f4-4019-be10-e9113c3b4c53 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 8c2ac679-6915-4149-aa51-99170b0df685] Deleting contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2210.142988] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-e69cca00-b4f4-4019-be10-e9113c3b4c53 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Deleting the datastore file [datastore1] 8c2ac679-6915-4149-aa51-99170b0df685 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2210.143250] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b00db4b7-c993-413f-988a-be01f02dbdc8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2210.149356] env[62824]: DEBUG oslo_vmware.api [None req-e69cca00-b4f4-4019-be10-e9113c3b4c53 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2210.149356] env[62824]: value = "task-2146570" [ 2210.149356] env[62824]: _type = "Task" [ 2210.149356] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2210.156989] env[62824]: DEBUG oslo_vmware.api [None req-e69cca00-b4f4-4019-be10-e9113c3b4c53 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146570, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2210.160334] env[62824]: DEBUG nova.compute.utils [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2210.164128] env[62824]: DEBUG nova.compute.manager [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7c44d956-53d1-4a5f-9016-c0f3f0f348b4] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2210.164318] env[62824]: DEBUG nova.network.neutron [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7c44d956-53d1-4a5f-9016-c0f3f0f348b4] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2210.190144] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7e6f95cb-45a7-48de-8220-fb8355af6eb3 tempest-ServerRescueTestJSONUnderV235-1876719970 tempest-ServerRescueTestJSONUnderV235-1876719970-project-member] Lock "da46d1d5-f75f-4ef1-b571-fbebab89c2a6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.839s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2210.212439] env[62824]: DEBUG nova.policy [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8c6cd8cd9ac848c89218208d1c2290aa', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '461117110de7467fbf8ada5ee9978069', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2210.420112] env[62824]: INFO nova.compute.manager [None req-7cf9f153-360a-4432-94f9-b1790d7b2e62 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] Detaching volume a7248bce-cc5d-47e2-b789-e0f635a1ce45 [ 2210.424540] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-f8f74f45-c68b-4875-9f0f-fc836d564d52 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2210.426971] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-f8f74f45-c68b-4875-9f0f-fc836d564d52 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8] Deleting contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2210.426971] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-f8f74f45-c68b-4875-9f0f-fc836d564d52 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Deleting the datastore file [datastore2] 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2210.427436] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-95e94d8b-7b9a-49da-a9a9-7c870869521c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2210.437406] env[62824]: DEBUG oslo_vmware.api [None req-f8f74f45-c68b-4875-9f0f-fc836d564d52 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the task: (returnval){ [ 2210.437406] env[62824]: value = "task-2146571" [ 2210.437406] env[62824]: _type = "Task" [ 2210.437406] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2210.443768] env[62824]: DEBUG oslo_vmware.api [None req-f8f74f45-c68b-4875-9f0f-fc836d564d52 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146571, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2210.471475] env[62824]: INFO nova.virt.block_device [None req-7cf9f153-360a-4432-94f9-b1790d7b2e62 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] Attempting to driver detach volume a7248bce-cc5d-47e2-b789-e0f635a1ce45 from mountpoint /dev/sdb [ 2210.471813] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-7cf9f153-360a-4432-94f9-b1790d7b2e62 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] Volume detach. Driver type: vmdk {{(pid=62824) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 2210.472284] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-7cf9f153-360a-4432-94f9-b1790d7b2e62 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-438718', 'volume_id': 'a7248bce-cc5d-47e2-b789-e0f635a1ce45', 'name': 'volume-a7248bce-cc5d-47e2-b789-e0f635a1ce45', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'b9f30f87-3594-4468-9d29-70890d8761e3', 'attached_at': '', 'detached_at': '', 'volume_id': 'a7248bce-cc5d-47e2-b789-e0f635a1ce45', 'serial': 'a7248bce-cc5d-47e2-b789-e0f635a1ce45'} {{(pid=62824) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 2210.472950] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66f3816b-0226-4c96-b75c-fe94d2a82fa3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2210.479207] env[62824]: DEBUG nova.network.neutron [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7c44d956-53d1-4a5f-9016-c0f3f0f348b4] Successfully created port: 8f686dbc-67bd-4f14-aec0-a6f9e7d8bfc5 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2210.501681] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42617b8f-a58d-4b83-997e-eeac5b072734 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2210.516302] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39609aee-801b-440c-9182-8de68e4b63bd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2210.519999] env[62824]: DEBUG oslo_vmware.api [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146548, 'name': Destroy_Task, 'duration_secs': 9.717083} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2210.519999] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 58243465-ba24-4b37-a640-04ab24963230] Destroyed the VM [ 2210.519999] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 58243465-ba24-4b37-a640-04ab24963230] Deleting Snapshot of the VM instance {{(pid=62824) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 2210.520492] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-85871bdd-6efb-4f5a-907c-e48f96112a29 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2210.541889] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aac5af50-6ef0-4872-8e6e-390fbfae8ff8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2210.544728] env[62824]: DEBUG oslo_vmware.api [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 2210.544728] env[62824]: value = "task-2146572" [ 2210.544728] env[62824]: _type = "Task" [ 2210.544728] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2210.558633] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-7cf9f153-360a-4432-94f9-b1790d7b2e62 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] The volume has not been displaced from its original location: [datastore1] volume-a7248bce-cc5d-47e2-b789-e0f635a1ce45/volume-a7248bce-cc5d-47e2-b789-e0f635a1ce45.vmdk. No consolidation needed. {{(pid=62824) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 2210.564093] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-7cf9f153-360a-4432-94f9-b1790d7b2e62 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] Reconfiguring VM instance instance-0000002f to detach disk 2001 {{(pid=62824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 2210.565204] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3b0c1c79-b917-4d18-af26-18722ec92584 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2210.582654] env[62824]: DEBUG oslo_vmware.api [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146572, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2210.586975] env[62824]: DEBUG oslo_vmware.api [None req-7cf9f153-360a-4432-94f9-b1790d7b2e62 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Waiting for the task: (returnval){ [ 2210.586975] env[62824]: value = "task-2146573" [ 2210.586975] env[62824]: _type = "Task" [ 2210.586975] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2210.595061] env[62824]: DEBUG oslo_vmware.api [None req-7cf9f153-360a-4432-94f9-b1790d7b2e62 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Task: {'id': task-2146573, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2210.642839] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146554, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2210.658620] env[62824]: DEBUG oslo_vmware.api [None req-e69cca00-b4f4-4019-be10-e9113c3b4c53 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146570, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.27019} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2210.659195] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-e69cca00-b4f4-4019-be10-e9113c3b4c53 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2210.659380] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e69cca00-b4f4-4019-be10-e9113c3b4c53 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 8c2ac679-6915-4149-aa51-99170b0df685] Deleted contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2210.659648] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e69cca00-b4f4-4019-be10-e9113c3b4c53 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 8c2ac679-6915-4149-aa51-99170b0df685] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2210.659855] env[62824]: INFO nova.compute.manager [None req-e69cca00-b4f4-4019-be10-e9113c3b4c53 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 8c2ac679-6915-4149-aa51-99170b0df685] Took 7.22 seconds to destroy the instance on the hypervisor. [ 2210.660131] env[62824]: DEBUG oslo.service.loopingcall [None req-e69cca00-b4f4-4019-be10-e9113c3b4c53 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2210.660342] env[62824]: DEBUG nova.compute.manager [-] [instance: 8c2ac679-6915-4149-aa51-99170b0df685] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2210.660505] env[62824]: DEBUG nova.network.neutron [-] [instance: 8c2ac679-6915-4149-aa51-99170b0df685] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2210.664756] env[62824]: DEBUG nova.compute.manager [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7c44d956-53d1-4a5f-9016-c0f3f0f348b4] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2210.951744] env[62824]: DEBUG oslo_vmware.api [None req-f8f74f45-c68b-4875-9f0f-fc836d564d52 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146571, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.133283} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2210.951992] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-f8f74f45-c68b-4875-9f0f-fc836d564d52 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2210.952927] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-f8f74f45-c68b-4875-9f0f-fc836d564d52 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8] Deleted contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2210.953136] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-f8f74f45-c68b-4875-9f0f-fc836d564d52 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2210.953563] env[62824]: INFO nova.compute.manager [None req-f8f74f45-c68b-4875-9f0f-fc836d564d52 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8] Took 4.54 seconds to destroy the instance on the hypervisor. [ 2210.953563] env[62824]: DEBUG oslo.service.loopingcall [None req-f8f74f45-c68b-4875-9f0f-fc836d564d52 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2210.953763] env[62824]: DEBUG nova.compute.manager [-] [instance: 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2210.953842] env[62824]: DEBUG nova.network.neutron [-] [instance: 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2211.011627] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-628fa055-99d5-4b7e-9a31-c573a1a6bd17 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2211.028628] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9226fddd-5dde-499d-9d2f-bbb17a0be4b8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2211.069274] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-758ae676-e6f8-46a3-8c65-4a947eabaa4a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2211.077497] env[62824]: DEBUG oslo_vmware.api [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146572, 'name': RemoveSnapshot_Task, 'duration_secs': 0.378702} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2211.079788] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 58243465-ba24-4b37-a640-04ab24963230] Deleted Snapshot of the VM instance {{(pid=62824) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 2211.080172] env[62824]: DEBUG nova.compute.manager [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 58243465-ba24-4b37-a640-04ab24963230] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2211.083024] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ece000f5-6f40-426d-8c17-aaa688d4d21f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2211.084429] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a59f84c5-96b5-4957-8501-77998847135b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2211.109019] env[62824]: DEBUG nova.compute.provider_tree [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2211.112504] env[62824]: DEBUG oslo_vmware.api [None req-7cf9f153-360a-4432-94f9-b1790d7b2e62 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Task: {'id': task-2146573, 'name': ReconfigVM_Task, 'duration_secs': 0.262815} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2211.114313] env[62824]: DEBUG nova.compute.manager [req-fcf9947c-8ef5-475e-a2c6-1b02c9a7d925 req-5d860d9e-3e01-4145-88da-8d2c20e29919 service nova] [instance: 8c2ac679-6915-4149-aa51-99170b0df685] Received event network-vif-deleted-891b53e5-1497-406d-8b11-2db60c8c4e1e {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2211.114615] env[62824]: INFO nova.compute.manager [req-fcf9947c-8ef5-475e-a2c6-1b02c9a7d925 req-5d860d9e-3e01-4145-88da-8d2c20e29919 service nova] [instance: 8c2ac679-6915-4149-aa51-99170b0df685] Neutron deleted interface 891b53e5-1497-406d-8b11-2db60c8c4e1e; detaching it from the instance and deleting it from the info cache [ 2211.114733] env[62824]: DEBUG nova.network.neutron [req-fcf9947c-8ef5-475e-a2c6-1b02c9a7d925 req-5d860d9e-3e01-4145-88da-8d2c20e29919 service nova] [instance: 8c2ac679-6915-4149-aa51-99170b0df685] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2211.115973] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-7cf9f153-360a-4432-94f9-b1790d7b2e62 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] Reconfigured VM instance instance-0000002f to detach disk 2001 {{(pid=62824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 2211.122242] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-36b8120e-f0a5-4974-9be3-a0b6b4f188ef {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2211.138827] env[62824]: DEBUG oslo_vmware.api [None req-7cf9f153-360a-4432-94f9-b1790d7b2e62 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Waiting for the task: (returnval){ [ 2211.138827] env[62824]: value = "task-2146574" [ 2211.138827] env[62824]: _type = "Task" [ 2211.138827] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2211.147085] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146554, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2211.151924] env[62824]: DEBUG oslo_vmware.api [None req-7cf9f153-360a-4432-94f9-b1790d7b2e62 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Task: {'id': task-2146574, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2211.299892] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-d85dcf93-e801-4553-b0ef-b90b42fed644 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 0c9fb5cc-de69-46d8-9962-98e0a84e33c3] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2211.299892] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-d85dcf93-e801-4553-b0ef-b90b42fed644 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 0c9fb5cc-de69-46d8-9962-98e0a84e33c3] Deleting contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2211.299892] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-d85dcf93-e801-4553-b0ef-b90b42fed644 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Deleting the datastore file [datastore1] 0c9fb5cc-de69-46d8-9962-98e0a84e33c3 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2211.299892] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3f4cfbb4-a46e-47a8-99c5-521328ff2b50 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2211.308028] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-56a95388-af07-48c0-bd95-c65c098831e2 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2211.308242] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-56a95388-af07-48c0-bd95-c65c098831e2 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Deleting contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2211.308450] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-56a95388-af07-48c0-bd95-c65c098831e2 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Deleting the datastore file [datastore1] 07ce931f-45ef-409b-b714-9f1cd47a3a88 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2211.309258] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f550b173-d069-4a2a-8997-28a44ee2d9be {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2211.311692] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-602cdfe0-df66-45cb-bc5a-cc85cb34718d tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 350cfd12-a8d5-4a5f-b3de-d795fa179dfd] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2211.311692] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-602cdfe0-df66-45cb-bc5a-cc85cb34718d tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 350cfd12-a8d5-4a5f-b3de-d795fa179dfd] Deleting contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2211.311692] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-602cdfe0-df66-45cb-bc5a-cc85cb34718d tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Deleting the datastore file [datastore2] 350cfd12-a8d5-4a5f-b3de-d795fa179dfd {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2211.311872] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-251e71cb-c6c9-40d9-8cfa-911d1c9d75f2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2211.314856] env[62824]: DEBUG oslo_vmware.api [None req-d85dcf93-e801-4553-b0ef-b90b42fed644 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Waiting for the task: (returnval){ [ 2211.314856] env[62824]: value = "task-2146575" [ 2211.314856] env[62824]: _type = "Task" [ 2211.314856] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2211.323018] env[62824]: DEBUG oslo_vmware.api [None req-602cdfe0-df66-45cb-bc5a-cc85cb34718d tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Waiting for the task: (returnval){ [ 2211.323018] env[62824]: value = "task-2146576" [ 2211.323018] env[62824]: _type = "Task" [ 2211.323018] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2211.325337] env[62824]: DEBUG oslo_vmware.api [None req-56a95388-af07-48c0-bd95-c65c098831e2 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for the task: (returnval){ [ 2211.325337] env[62824]: value = "task-2146577" [ 2211.325337] env[62824]: _type = "Task" [ 2211.325337] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2211.325583] env[62824]: DEBUG oslo_vmware.api [None req-d85dcf93-e801-4553-b0ef-b90b42fed644 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146575, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2211.337591] env[62824]: DEBUG oslo_vmware.api [None req-602cdfe0-df66-45cb-bc5a-cc85cb34718d tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Task: {'id': task-2146576, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2211.340930] env[62824]: DEBUG oslo_vmware.api [None req-56a95388-af07-48c0-bd95-c65c098831e2 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146577, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2211.463734] env[62824]: DEBUG nova.network.neutron [-] [instance: 8c2ac679-6915-4149-aa51-99170b0df685] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2211.613550] env[62824]: INFO nova.compute.manager [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 58243465-ba24-4b37-a640-04ab24963230] Shelve offloading [ 2211.617424] env[62824]: DEBUG nova.scheduler.client.report [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2211.622047] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a9c100c6-853e-4d1a-a239-c9c9756f5bed {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2211.631665] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af9b9d65-9b5a-4d25-9009-eeedb8c2b03a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2211.655941] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146554, 'name': CreateVM_Task, 'duration_secs': 9.387582} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2211.658943] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2211.659388] env[62824]: DEBUG oslo_vmware.api [None req-7cf9f153-360a-4432-94f9-b1790d7b2e62 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Task: {'id': task-2146574, 'name': ReconfigVM_Task, 'duration_secs': 0.129197} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2211.659892] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2211.660074] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2211.660483] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2211.660844] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-7cf9f153-360a-4432-94f9-b1790d7b2e62 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-438718', 'volume_id': 'a7248bce-cc5d-47e2-b789-e0f635a1ce45', 'name': 'volume-a7248bce-cc5d-47e2-b789-e0f635a1ce45', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'b9f30f87-3594-4468-9d29-70890d8761e3', 'attached_at': '', 'detached_at': '', 'volume_id': 'a7248bce-cc5d-47e2-b789-e0f635a1ce45', 'serial': 'a7248bce-cc5d-47e2-b789-e0f635a1ce45'} {{(pid=62824) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 2211.674376] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eb70baab-8a48-4b50-9ca3-132ab47a1b18 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2211.676592] env[62824]: DEBUG nova.compute.manager [req-fcf9947c-8ef5-475e-a2c6-1b02c9a7d925 req-5d860d9e-3e01-4145-88da-8d2c20e29919 service nova] [instance: 8c2ac679-6915-4149-aa51-99170b0df685] Detach interface failed, port_id=891b53e5-1497-406d-8b11-2db60c8c4e1e, reason: Instance 8c2ac679-6915-4149-aa51-99170b0df685 could not be found. {{(pid=62824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11482}} [ 2211.677600] env[62824]: DEBUG nova.compute.manager [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7c44d956-53d1-4a5f-9016-c0f3f0f348b4] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2211.684606] env[62824]: DEBUG oslo_vmware.api [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Waiting for the task: (returnval){ [ 2211.684606] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5259d43b-492c-080a-d71a-a4d438998748" [ 2211.684606] env[62824]: _type = "Task" [ 2211.684606] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2211.694976] env[62824]: DEBUG oslo_vmware.api [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5259d43b-492c-080a-d71a-a4d438998748, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2211.710359] env[62824]: DEBUG nova.virt.hardware [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2211.710601] env[62824]: DEBUG nova.virt.hardware [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2211.710759] env[62824]: DEBUG nova.virt.hardware [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2211.710941] env[62824]: DEBUG nova.virt.hardware [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2211.711108] env[62824]: DEBUG nova.virt.hardware [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2211.711264] env[62824]: DEBUG nova.virt.hardware [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2211.711475] env[62824]: DEBUG nova.virt.hardware [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2211.711640] env[62824]: DEBUG nova.virt.hardware [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2211.711797] env[62824]: DEBUG nova.virt.hardware [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2211.711962] env[62824]: DEBUG nova.virt.hardware [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2211.712163] env[62824]: DEBUG nova.virt.hardware [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2211.713148] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c26d6f1-aac3-4415-ac46-47f1afddfd61 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2211.715616] env[62824]: DEBUG nova.network.neutron [-] [instance: 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2211.721950] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1315fdb2-beec-4f25-9c98-11080a6d922b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2211.824508] env[62824]: DEBUG oslo_vmware.api [None req-d85dcf93-e801-4553-b0ef-b90b42fed644 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Task: {'id': task-2146575, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.156723} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2211.824757] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-d85dcf93-e801-4553-b0ef-b90b42fed644 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2211.824939] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-d85dcf93-e801-4553-b0ef-b90b42fed644 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 0c9fb5cc-de69-46d8-9962-98e0a84e33c3] Deleted contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2211.825131] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-d85dcf93-e801-4553-b0ef-b90b42fed644 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 0c9fb5cc-de69-46d8-9962-98e0a84e33c3] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2211.825311] env[62824]: INFO nova.compute.manager [None req-d85dcf93-e801-4553-b0ef-b90b42fed644 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] [instance: 0c9fb5cc-de69-46d8-9962-98e0a84e33c3] Took 6.40 seconds to destroy the instance on the hypervisor. [ 2211.825546] env[62824]: DEBUG oslo.service.loopingcall [None req-d85dcf93-e801-4553-b0ef-b90b42fed644 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2211.825802] env[62824]: DEBUG nova.compute.manager [-] [instance: 0c9fb5cc-de69-46d8-9962-98e0a84e33c3] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2211.825901] env[62824]: DEBUG nova.network.neutron [-] [instance: 0c9fb5cc-de69-46d8-9962-98e0a84e33c3] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2211.840830] env[62824]: DEBUG oslo_vmware.api [None req-56a95388-af07-48c0-bd95-c65c098831e2 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Task: {'id': task-2146577, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.152702} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2211.841076] env[62824]: DEBUG oslo_vmware.api [None req-602cdfe0-df66-45cb-bc5a-cc85cb34718d tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Task: {'id': task-2146576, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.136144} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2211.841299] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-56a95388-af07-48c0-bd95-c65c098831e2 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2211.841477] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-56a95388-af07-48c0-bd95-c65c098831e2 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Deleted contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2211.841658] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-56a95388-af07-48c0-bd95-c65c098831e2 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2211.841825] env[62824]: INFO nova.compute.manager [None req-56a95388-af07-48c0-bd95-c65c098831e2 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Took 5.41 seconds to destroy the instance on the hypervisor. [ 2211.842075] env[62824]: DEBUG oslo.service.loopingcall [None req-56a95388-af07-48c0-bd95-c65c098831e2 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2211.842254] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-602cdfe0-df66-45cb-bc5a-cc85cb34718d tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2211.842414] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-602cdfe0-df66-45cb-bc5a-cc85cb34718d tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 350cfd12-a8d5-4a5f-b3de-d795fa179dfd] Deleted contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2211.842583] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-602cdfe0-df66-45cb-bc5a-cc85cb34718d tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 350cfd12-a8d5-4a5f-b3de-d795fa179dfd] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2211.842744] env[62824]: INFO nova.compute.manager [None req-602cdfe0-df66-45cb-bc5a-cc85cb34718d tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] [instance: 350cfd12-a8d5-4a5f-b3de-d795fa179dfd] Took 7.70 seconds to destroy the instance on the hypervisor. [ 2211.842953] env[62824]: DEBUG oslo.service.loopingcall [None req-602cdfe0-df66-45cb-bc5a-cc85cb34718d tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2211.843150] env[62824]: DEBUG nova.compute.manager [-] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2211.843245] env[62824]: DEBUG nova.network.neutron [-] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2211.844992] env[62824]: DEBUG nova.compute.manager [-] [instance: 350cfd12-a8d5-4a5f-b3de-d795fa179dfd] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2211.845069] env[62824]: DEBUG nova.network.neutron [-] [instance: 350cfd12-a8d5-4a5f-b3de-d795fa179dfd] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2211.966147] env[62824]: INFO nova.compute.manager [-] [instance: 8c2ac679-6915-4149-aa51-99170b0df685] Took 1.31 seconds to deallocate network for instance. [ 2212.121587] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 58243465-ba24-4b37-a640-04ab24963230] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2212.123115] env[62824]: DEBUG oslo_concurrency.lockutils [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.468s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2212.123742] env[62824]: DEBUG nova.compute.manager [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 853f3098-881c-402c-8fd7-26540d1ecd88] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2212.129709] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-85c09429-8de9-4d68-894b-520b0829a19b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2212.137622] env[62824]: DEBUG oslo_vmware.api [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 2212.137622] env[62824]: value = "task-2146578" [ 2212.137622] env[62824]: _type = "Task" [ 2212.137622] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2212.148391] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 58243465-ba24-4b37-a640-04ab24963230] VM already powered off {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 2212.149393] env[62824]: DEBUG nova.compute.manager [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 58243465-ba24-4b37-a640-04ab24963230] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2212.149699] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bec44eb-e58b-4f83-9367-17de1d51cb6d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2212.160133] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquiring lock "refresh_cache-58243465-ba24-4b37-a640-04ab24963230" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2212.160310] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquired lock "refresh_cache-58243465-ba24-4b37-a640-04ab24963230" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2212.161261] env[62824]: DEBUG nova.network.neutron [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 58243465-ba24-4b37-a640-04ab24963230] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2212.194351] env[62824]: DEBUG nova.compute.manager [req-263a9fa9-4507-4ead-a24d-58098f3e8981 req-ec7c2c10-b91d-4bd8-8739-cd9563c5897c service nova] [instance: 350cfd12-a8d5-4a5f-b3de-d795fa179dfd] Received event network-vif-deleted-75b0f475-5e6e-44de-9480-a71716ecb4e0 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2212.194351] env[62824]: INFO nova.compute.manager [req-263a9fa9-4507-4ead-a24d-58098f3e8981 req-ec7c2c10-b91d-4bd8-8739-cd9563c5897c service nova] [instance: 350cfd12-a8d5-4a5f-b3de-d795fa179dfd] Neutron deleted interface 75b0f475-5e6e-44de-9480-a71716ecb4e0; detaching it from the instance and deleting it from the info cache [ 2212.194351] env[62824]: DEBUG nova.network.neutron [req-263a9fa9-4507-4ead-a24d-58098f3e8981 req-ec7c2c10-b91d-4bd8-8739-cd9563c5897c service nova] [instance: 350cfd12-a8d5-4a5f-b3de-d795fa179dfd] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2212.202509] env[62824]: DEBUG oslo_vmware.api [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5259d43b-492c-080a-d71a-a4d438998748, 'name': SearchDatastore_Task, 'duration_secs': 0.019842} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2212.203136] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2212.203136] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2212.203269] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2212.203452] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2212.203632] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2212.203890] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-33bef6ec-be7b-4d6b-ad40-f577afffe4ab {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2212.216981] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2212.217197] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2212.218558] env[62824]: INFO nova.compute.manager [-] [instance: 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8] Took 1.26 seconds to deallocate network for instance. [ 2212.218836] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6f9fd611-1227-4d64-9ef6-b24a6d0fffca {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2212.228400] env[62824]: DEBUG nova.objects.instance [None req-7cf9f153-360a-4432-94f9-b1790d7b2e62 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Lazy-loading 'flavor' on Instance uuid b9f30f87-3594-4468-9d29-70890d8761e3 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2212.234172] env[62824]: DEBUG oslo_vmware.api [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Waiting for the task: (returnval){ [ 2212.234172] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52950af0-efd0-e0b7-64da-cf83eef95d60" [ 2212.234172] env[62824]: _type = "Task" [ 2212.234172] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2212.245403] env[62824]: DEBUG oslo_vmware.api [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52950af0-efd0-e0b7-64da-cf83eef95d60, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2212.475986] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e69cca00-b4f4-4019-be10-e9113c3b4c53 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2212.476288] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e69cca00-b4f4-4019-be10-e9113c3b4c53 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2212.476538] env[62824]: DEBUG nova.objects.instance [None req-e69cca00-b4f4-4019-be10-e9113c3b4c53 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lazy-loading 'resources' on Instance uuid 8c2ac679-6915-4149-aa51-99170b0df685 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2212.553598] env[62824]: DEBUG nova.network.neutron [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7c44d956-53d1-4a5f-9016-c0f3f0f348b4] Successfully updated port: 8f686dbc-67bd-4f14-aec0-a6f9e7d8bfc5 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2212.633208] env[62824]: DEBUG nova.compute.utils [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2212.634755] env[62824]: DEBUG nova.compute.manager [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 853f3098-881c-402c-8fd7-26540d1ecd88] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2212.634930] env[62824]: DEBUG nova.network.neutron [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 853f3098-881c-402c-8fd7-26540d1ecd88] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2212.644784] env[62824]: DEBUG nova.network.neutron [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 1f162a63-8773-4c34-9c84-b2de1d5e9f50] Successfully updated port: f5c2ffd1-3953-40e4-8d96-0542389767da {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2212.658755] env[62824]: DEBUG nova.network.neutron [-] [instance: 350cfd12-a8d5-4a5f-b3de-d795fa179dfd] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2212.677727] env[62824]: DEBUG nova.policy [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4c914343569a462d9b8564f3376c2f54', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '10ee5fe93d584725ace90e76605a1b2e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2212.696103] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1415907e-49bb-4cfb-9b77-4229d657454a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2212.706177] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a30d7b1-6e3b-4503-ad2f-368edcb44324 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2212.728961] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f8f74f45-c68b-4875-9f0f-fc836d564d52 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2212.743284] env[62824]: DEBUG nova.compute.manager [req-263a9fa9-4507-4ead-a24d-58098f3e8981 req-ec7c2c10-b91d-4bd8-8739-cd9563c5897c service nova] [instance: 350cfd12-a8d5-4a5f-b3de-d795fa179dfd] Detach interface failed, port_id=75b0f475-5e6e-44de-9480-a71716ecb4e0, reason: Instance 350cfd12-a8d5-4a5f-b3de-d795fa179dfd could not be found. {{(pid=62824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11482}} [ 2212.755543] env[62824]: DEBUG oslo_vmware.api [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52950af0-efd0-e0b7-64da-cf83eef95d60, 'name': SearchDatastore_Task, 'duration_secs': 0.011127} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2212.756346] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fd172437-cba8-4bbc-8988-daa4a8ff2c9f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2212.764669] env[62824]: DEBUG oslo_vmware.api [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Waiting for the task: (returnval){ [ 2212.764669] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5244a4a9-9b80-3bb0-900e-0c2b1aea7768" [ 2212.764669] env[62824]: _type = "Task" [ 2212.764669] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2212.777023] env[62824]: DEBUG oslo_vmware.api [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5244a4a9-9b80-3bb0-900e-0c2b1aea7768, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2212.793949] env[62824]: DEBUG nova.network.neutron [-] [instance: 0c9fb5cc-de69-46d8-9962-98e0a84e33c3] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2212.797268] env[62824]: DEBUG nova.network.neutron [-] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2213.057492] env[62824]: DEBUG oslo_concurrency.lockutils [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Acquiring lock "refresh_cache-7c44d956-53d1-4a5f-9016-c0f3f0f348b4" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2213.057612] env[62824]: DEBUG oslo_concurrency.lockutils [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Acquired lock "refresh_cache-7c44d956-53d1-4a5f-9016-c0f3f0f348b4" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2213.057684] env[62824]: DEBUG nova.network.neutron [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7c44d956-53d1-4a5f-9016-c0f3f0f348b4] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2213.138422] env[62824]: DEBUG nova.compute.manager [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 853f3098-881c-402c-8fd7-26540d1ecd88] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2213.149448] env[62824]: DEBUG oslo_concurrency.lockutils [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Acquiring lock "refresh_cache-1f162a63-8773-4c34-9c84-b2de1d5e9f50" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2213.150253] env[62824]: DEBUG oslo_concurrency.lockutils [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Acquired lock "refresh_cache-1f162a63-8773-4c34-9c84-b2de1d5e9f50" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2213.150422] env[62824]: DEBUG nova.network.neutron [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 1f162a63-8773-4c34-9c84-b2de1d5e9f50] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2213.162112] env[62824]: INFO nova.compute.manager [-] [instance: 350cfd12-a8d5-4a5f-b3de-d795fa179dfd] Took 1.32 seconds to deallocate network for instance. [ 2213.247239] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7cf9f153-360a-4432-94f9-b1790d7b2e62 tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Lock "b9f30f87-3594-4468-9d29-70890d8761e3" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.331s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2213.275682] env[62824]: DEBUG nova.network.neutron [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 58243465-ba24-4b37-a640-04ab24963230] Updating instance_info_cache with network_info: [{"id": "63af4d3b-ce05-431b-84e1-e3e3fbdf673d", "address": "fa:16:3e:da:11:f0", "network": {"id": "b995934c-4934-4469-884c-73cb4de1796f", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1892307918-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c11aef7fb6247cb9b5272a6a063cd12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15165046-2de9-4ada-9e99-0126e20854a9", "external-id": "nsx-vlan-transportzone-974", "segmentation_id": 974, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap63af4d3b-ce", "ovs_interfaceid": "63af4d3b-ce05-431b-84e1-e3e3fbdf673d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2213.281114] env[62824]: DEBUG oslo_vmware.api [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5244a4a9-9b80-3bb0-900e-0c2b1aea7768, 'name': SearchDatastore_Task, 'duration_secs': 0.010089} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2213.281921] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2213.282404] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412/c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2213.282730] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-15caf12b-71ed-492e-9168-3dcfab747346 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2213.293438] env[62824]: DEBUG oslo_vmware.api [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Waiting for the task: (returnval){ [ 2213.293438] env[62824]: value = "task-2146579" [ 2213.293438] env[62824]: _type = "Task" [ 2213.293438] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2213.299741] env[62824]: INFO nova.compute.manager [-] [instance: 0c9fb5cc-de69-46d8-9962-98e0a84e33c3] Took 1.47 seconds to deallocate network for instance. [ 2213.300149] env[62824]: INFO nova.compute.manager [-] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Took 1.46 seconds to deallocate network for instance. [ 2213.309869] env[62824]: DEBUG oslo_vmware.api [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Task: {'id': task-2146579, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2213.345177] env[62824]: DEBUG nova.compute.manager [req-9e69838b-065e-433f-b77a-b0a409fd857d req-5c291ac1-ce10-4792-8735-e472e5dc6f6b service nova] [instance: 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8] Received event network-vif-deleted-bfd88146-4384-4514-8ceb-46c647c80eee {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2213.345476] env[62824]: DEBUG nova.compute.manager [req-9e69838b-065e-433f-b77a-b0a409fd857d req-5c291ac1-ce10-4792-8735-e472e5dc6f6b service nova] [instance: 0c9fb5cc-de69-46d8-9962-98e0a84e33c3] Received event network-vif-deleted-b96f49a1-2e1e-495b-9809-a7338087a686 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2213.345672] env[62824]: DEBUG nova.compute.manager [req-9e69838b-065e-433f-b77a-b0a409fd857d req-5c291ac1-ce10-4792-8735-e472e5dc6f6b service nova] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Received event network-vif-deleted-309b4346-1935-41f1-a374-d322c4a8c016 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2213.345847] env[62824]: DEBUG nova.compute.manager [req-9e69838b-065e-433f-b77a-b0a409fd857d req-5c291ac1-ce10-4792-8735-e472e5dc6f6b service nova] [instance: 7c44d956-53d1-4a5f-9016-c0f3f0f348b4] Received event network-vif-plugged-8f686dbc-67bd-4f14-aec0-a6f9e7d8bfc5 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2213.346057] env[62824]: DEBUG oslo_concurrency.lockutils [req-9e69838b-065e-433f-b77a-b0a409fd857d req-5c291ac1-ce10-4792-8735-e472e5dc6f6b service nova] Acquiring lock "7c44d956-53d1-4a5f-9016-c0f3f0f348b4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2213.346275] env[62824]: DEBUG oslo_concurrency.lockutils [req-9e69838b-065e-433f-b77a-b0a409fd857d req-5c291ac1-ce10-4792-8735-e472e5dc6f6b service nova] Lock "7c44d956-53d1-4a5f-9016-c0f3f0f348b4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2213.346743] env[62824]: DEBUG oslo_concurrency.lockutils [req-9e69838b-065e-433f-b77a-b0a409fd857d req-5c291ac1-ce10-4792-8735-e472e5dc6f6b service nova] Lock "7c44d956-53d1-4a5f-9016-c0f3f0f348b4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2213.346968] env[62824]: DEBUG nova.compute.manager [req-9e69838b-065e-433f-b77a-b0a409fd857d req-5c291ac1-ce10-4792-8735-e472e5dc6f6b service nova] [instance: 7c44d956-53d1-4a5f-9016-c0f3f0f348b4] No waiting events found dispatching network-vif-plugged-8f686dbc-67bd-4f14-aec0-a6f9e7d8bfc5 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2213.347170] env[62824]: WARNING nova.compute.manager [req-9e69838b-065e-433f-b77a-b0a409fd857d req-5c291ac1-ce10-4792-8735-e472e5dc6f6b service nova] [instance: 7c44d956-53d1-4a5f-9016-c0f3f0f348b4] Received unexpected event network-vif-plugged-8f686dbc-67bd-4f14-aec0-a6f9e7d8bfc5 for instance with vm_state building and task_state spawning. [ 2213.347342] env[62824]: DEBUG nova.compute.manager [req-9e69838b-065e-433f-b77a-b0a409fd857d req-5c291ac1-ce10-4792-8735-e472e5dc6f6b service nova] [instance: 7c44d956-53d1-4a5f-9016-c0f3f0f348b4] Received event network-changed-8f686dbc-67bd-4f14-aec0-a6f9e7d8bfc5 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2213.347502] env[62824]: DEBUG nova.compute.manager [req-9e69838b-065e-433f-b77a-b0a409fd857d req-5c291ac1-ce10-4792-8735-e472e5dc6f6b service nova] [instance: 7c44d956-53d1-4a5f-9016-c0f3f0f348b4] Refreshing instance network info cache due to event network-changed-8f686dbc-67bd-4f14-aec0-a6f9e7d8bfc5. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2213.347679] env[62824]: DEBUG oslo_concurrency.lockutils [req-9e69838b-065e-433f-b77a-b0a409fd857d req-5c291ac1-ce10-4792-8735-e472e5dc6f6b service nova] Acquiring lock "refresh_cache-7c44d956-53d1-4a5f-9016-c0f3f0f348b4" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2213.365560] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16111a6f-3287-46d2-b80e-c2cc7a8ff337 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2213.373831] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5ffedf0-5e3c-4b92-918c-2f05c4f59263 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2213.377533] env[62824]: DEBUG nova.network.neutron [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 853f3098-881c-402c-8fd7-26540d1ecd88] Successfully created port: 03eb70c1-fc19-4e7c-874d-2f4fcd77fe6b {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2213.408720] env[62824]: DEBUG nova.network.neutron [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 1f162a63-8773-4c34-9c84-b2de1d5e9f50] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2213.411791] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-188749f3-a475-4785-9657-ddfa680ab39a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2213.419626] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf696ce4-255f-4797-852c-9599e215a6d5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2213.434105] env[62824]: DEBUG nova.compute.provider_tree [None req-e69cca00-b4f4-4019-be10-e9113c3b4c53 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2213.602271] env[62824]: DEBUG nova.network.neutron [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7c44d956-53d1-4a5f-9016-c0f3f0f348b4] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2213.673736] env[62824]: DEBUG oslo_concurrency.lockutils [None req-602cdfe0-df66-45cb-bc5a-cc85cb34718d tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2213.777734] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Releasing lock "refresh_cache-58243465-ba24-4b37-a640-04ab24963230" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2213.805179] env[62824]: DEBUG oslo_vmware.api [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Task: {'id': task-2146579, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2213.816146] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d85dcf93-e801-4553-b0ef-b90b42fed644 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2213.817173] env[62824]: DEBUG oslo_concurrency.lockutils [None req-56a95388-af07-48c0-bd95-c65c098831e2 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2213.820485] env[62824]: DEBUG nova.network.neutron [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 1f162a63-8773-4c34-9c84-b2de1d5e9f50] Updating instance_info_cache with network_info: [{"id": "f5c2ffd1-3953-40e4-8d96-0542389767da", "address": "fa:16:3e:36:ad:f3", "network": {"id": "91425d93-4190-4e61-ae47-94efd278247d", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1364948114-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "461117110de7467fbf8ada5ee9978069", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c894ab55-c869-4530-9702-cb46d173ce94", "external-id": "nsx-vlan-transportzone-792", "segmentation_id": 792, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf5c2ffd1-39", "ovs_interfaceid": "f5c2ffd1-3953-40e4-8d96-0542389767da", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2213.856756] env[62824]: DEBUG nova.network.neutron [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7c44d956-53d1-4a5f-9016-c0f3f0f348b4] Updating instance_info_cache with network_info: [{"id": "8f686dbc-67bd-4f14-aec0-a6f9e7d8bfc5", "address": "fa:16:3e:71:ae:7b", "network": {"id": "91425d93-4190-4e61-ae47-94efd278247d", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1364948114-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "461117110de7467fbf8ada5ee9978069", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c894ab55-c869-4530-9702-cb46d173ce94", "external-id": "nsx-vlan-transportzone-792", "segmentation_id": 792, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8f686dbc-67", "ovs_interfaceid": "8f686dbc-67bd-4f14-aec0-a6f9e7d8bfc5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2213.937384] env[62824]: DEBUG nova.scheduler.client.report [None req-e69cca00-b4f4-4019-be10-e9113c3b4c53 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2214.153174] env[62824]: DEBUG nova.compute.manager [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 853f3098-881c-402c-8fd7-26540d1ecd88] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2214.185023] env[62824]: DEBUG nova.virt.hardware [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2214.185023] env[62824]: DEBUG nova.virt.hardware [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2214.185023] env[62824]: DEBUG nova.virt.hardware [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2214.185023] env[62824]: DEBUG nova.virt.hardware [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2214.185023] env[62824]: DEBUG nova.virt.hardware [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2214.185023] env[62824]: DEBUG nova.virt.hardware [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2214.185530] env[62824]: DEBUG nova.virt.hardware [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2214.186060] env[62824]: DEBUG nova.virt.hardware [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2214.186384] env[62824]: DEBUG nova.virt.hardware [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2214.186696] env[62824]: DEBUG nova.virt.hardware [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2214.187200] env[62824]: DEBUG nova.virt.hardware [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2214.188399] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a13a7f3-3bce-41f8-a12f-a4f2b7bf5f53 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2214.198354] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b76f6b4-c3e5-44ee-8996-1553c792dc37 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2214.224394] env[62824]: DEBUG nova.compute.manager [req-440b3d6e-8525-4853-9205-7c8d36ba7ca9 req-6212e975-ba2b-42db-9a62-f25fe1cd1c04 service nova] [instance: 1f162a63-8773-4c34-9c84-b2de1d5e9f50] Received event network-vif-plugged-f5c2ffd1-3953-40e4-8d96-0542389767da {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2214.224394] env[62824]: DEBUG oslo_concurrency.lockutils [req-440b3d6e-8525-4853-9205-7c8d36ba7ca9 req-6212e975-ba2b-42db-9a62-f25fe1cd1c04 service nova] Acquiring lock "1f162a63-8773-4c34-9c84-b2de1d5e9f50-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2214.224394] env[62824]: DEBUG oslo_concurrency.lockutils [req-440b3d6e-8525-4853-9205-7c8d36ba7ca9 req-6212e975-ba2b-42db-9a62-f25fe1cd1c04 service nova] Lock "1f162a63-8773-4c34-9c84-b2de1d5e9f50-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2214.224394] env[62824]: DEBUG oslo_concurrency.lockutils [req-440b3d6e-8525-4853-9205-7c8d36ba7ca9 req-6212e975-ba2b-42db-9a62-f25fe1cd1c04 service nova] Lock "1f162a63-8773-4c34-9c84-b2de1d5e9f50-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2214.224394] env[62824]: DEBUG nova.compute.manager [req-440b3d6e-8525-4853-9205-7c8d36ba7ca9 req-6212e975-ba2b-42db-9a62-f25fe1cd1c04 service nova] [instance: 1f162a63-8773-4c34-9c84-b2de1d5e9f50] No waiting events found dispatching network-vif-plugged-f5c2ffd1-3953-40e4-8d96-0542389767da {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2214.224394] env[62824]: WARNING nova.compute.manager [req-440b3d6e-8525-4853-9205-7c8d36ba7ca9 req-6212e975-ba2b-42db-9a62-f25fe1cd1c04 service nova] [instance: 1f162a63-8773-4c34-9c84-b2de1d5e9f50] Received unexpected event network-vif-plugged-f5c2ffd1-3953-40e4-8d96-0542389767da for instance with vm_state building and task_state spawning. [ 2214.224394] env[62824]: DEBUG nova.compute.manager [req-440b3d6e-8525-4853-9205-7c8d36ba7ca9 req-6212e975-ba2b-42db-9a62-f25fe1cd1c04 service nova] [instance: 1f162a63-8773-4c34-9c84-b2de1d5e9f50] Received event network-changed-f5c2ffd1-3953-40e4-8d96-0542389767da {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2214.224940] env[62824]: DEBUG nova.compute.manager [req-440b3d6e-8525-4853-9205-7c8d36ba7ca9 req-6212e975-ba2b-42db-9a62-f25fe1cd1c04 service nova] [instance: 1f162a63-8773-4c34-9c84-b2de1d5e9f50] Refreshing instance network info cache due to event network-changed-f5c2ffd1-3953-40e4-8d96-0542389767da. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2214.225613] env[62824]: DEBUG oslo_concurrency.lockutils [req-440b3d6e-8525-4853-9205-7c8d36ba7ca9 req-6212e975-ba2b-42db-9a62-f25fe1cd1c04 service nova] Acquiring lock "refresh_cache-1f162a63-8773-4c34-9c84-b2de1d5e9f50" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2214.287647] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Acquiring lock "b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2214.287647] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Lock "b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2214.303879] env[62824]: DEBUG oslo_vmware.api [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Task: {'id': task-2146579, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2214.325786] env[62824]: DEBUG oslo_concurrency.lockutils [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Releasing lock "refresh_cache-1f162a63-8773-4c34-9c84-b2de1d5e9f50" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2214.325786] env[62824]: DEBUG nova.compute.manager [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 1f162a63-8773-4c34-9c84-b2de1d5e9f50] Instance network_info: |[{"id": "f5c2ffd1-3953-40e4-8d96-0542389767da", "address": "fa:16:3e:36:ad:f3", "network": {"id": "91425d93-4190-4e61-ae47-94efd278247d", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1364948114-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "461117110de7467fbf8ada5ee9978069", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c894ab55-c869-4530-9702-cb46d173ce94", "external-id": "nsx-vlan-transportzone-792", "segmentation_id": 792, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf5c2ffd1-39", "ovs_interfaceid": "f5c2ffd1-3953-40e4-8d96-0542389767da", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2214.325971] env[62824]: DEBUG oslo_concurrency.lockutils [req-440b3d6e-8525-4853-9205-7c8d36ba7ca9 req-6212e975-ba2b-42db-9a62-f25fe1cd1c04 service nova] Acquired lock "refresh_cache-1f162a63-8773-4c34-9c84-b2de1d5e9f50" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2214.326179] env[62824]: DEBUG nova.network.neutron [req-440b3d6e-8525-4853-9205-7c8d36ba7ca9 req-6212e975-ba2b-42db-9a62-f25fe1cd1c04 service nova] [instance: 1f162a63-8773-4c34-9c84-b2de1d5e9f50] Refreshing network info cache for port f5c2ffd1-3953-40e4-8d96-0542389767da {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2214.327357] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 1f162a63-8773-4c34-9c84-b2de1d5e9f50] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:36:ad:f3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c894ab55-c869-4530-9702-cb46d173ce94', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f5c2ffd1-3953-40e4-8d96-0542389767da', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2214.336106] env[62824]: DEBUG oslo.service.loopingcall [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2214.336955] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1f162a63-8773-4c34-9c84-b2de1d5e9f50] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2214.337192] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-68139061-d47b-4452-915f-28a0dcf66fa1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2214.360257] env[62824]: DEBUG oslo_concurrency.lockutils [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Releasing lock "refresh_cache-7c44d956-53d1-4a5f-9016-c0f3f0f348b4" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2214.360576] env[62824]: DEBUG nova.compute.manager [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7c44d956-53d1-4a5f-9016-c0f3f0f348b4] Instance network_info: |[{"id": "8f686dbc-67bd-4f14-aec0-a6f9e7d8bfc5", "address": "fa:16:3e:71:ae:7b", "network": {"id": "91425d93-4190-4e61-ae47-94efd278247d", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1364948114-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "461117110de7467fbf8ada5ee9978069", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c894ab55-c869-4530-9702-cb46d173ce94", "external-id": "nsx-vlan-transportzone-792", "segmentation_id": 792, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8f686dbc-67", "ovs_interfaceid": "8f686dbc-67bd-4f14-aec0-a6f9e7d8bfc5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2214.364018] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2214.364018] env[62824]: value = "task-2146580" [ 2214.364018] env[62824]: _type = "Task" [ 2214.364018] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2214.364018] env[62824]: DEBUG oslo_concurrency.lockutils [req-9e69838b-065e-433f-b77a-b0a409fd857d req-5c291ac1-ce10-4792-8735-e472e5dc6f6b service nova] Acquired lock "refresh_cache-7c44d956-53d1-4a5f-9016-c0f3f0f348b4" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2214.364018] env[62824]: DEBUG nova.network.neutron [req-9e69838b-065e-433f-b77a-b0a409fd857d req-5c291ac1-ce10-4792-8735-e472e5dc6f6b service nova] [instance: 7c44d956-53d1-4a5f-9016-c0f3f0f348b4] Refreshing network info cache for port 8f686dbc-67bd-4f14-aec0-a6f9e7d8bfc5 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2214.364018] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7c44d956-53d1-4a5f-9016-c0f3f0f348b4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:71:ae:7b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c894ab55-c869-4530-9702-cb46d173ce94', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8f686dbc-67bd-4f14-aec0-a6f9e7d8bfc5', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2214.371471] env[62824]: DEBUG oslo.service.loopingcall [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2214.374513] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7c44d956-53d1-4a5f-9016-c0f3f0f348b4] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2214.378198] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0dffd5ec-780c-4a67-9f7e-1f0962f37cd9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2214.394530] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 58243465-ba24-4b37-a640-04ab24963230] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2214.395513] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc5f87f8-e6ad-46a9-8c6b-a0ed664ee63e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2214.402370] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2214.402370] env[62824]: value = "task-2146581" [ 2214.402370] env[62824]: _type = "Task" [ 2214.402370] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2214.402595] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146580, 'name': CreateVM_Task} progress is 15%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2214.409794] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 58243465-ba24-4b37-a640-04ab24963230] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2214.410381] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-72244d3b-46b4-4f3c-a3a2-25c2d8d832ce {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2214.415081] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146581, 'name': CreateVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2214.441901] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e69cca00-b4f4-4019-be10-e9113c3b4c53 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.966s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2214.444018] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f8f74f45-c68b-4875-9f0f-fc836d564d52 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.715s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2214.444275] env[62824]: DEBUG nova.objects.instance [None req-f8f74f45-c68b-4875-9f0f-fc836d564d52 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Lazy-loading 'resources' on Instance uuid 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2214.465842] env[62824]: INFO nova.scheduler.client.report [None req-e69cca00-b4f4-4019-be10-e9113c3b4c53 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Deleted allocations for instance 8c2ac679-6915-4149-aa51-99170b0df685 [ 2214.510946] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 58243465-ba24-4b37-a640-04ab24963230] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2214.511194] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 58243465-ba24-4b37-a640-04ab24963230] Deleting contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2214.511379] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Deleting the datastore file [datastore2] 58243465-ba24-4b37-a640-04ab24963230 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2214.511882] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ec267df6-f1bc-409c-9135-cb303999c733 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2214.519051] env[62824]: DEBUG oslo_vmware.api [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 2214.519051] env[62824]: value = "task-2146583" [ 2214.519051] env[62824]: _type = "Task" [ 2214.519051] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2214.527842] env[62824]: DEBUG oslo_vmware.api [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146583, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2214.661997] env[62824]: DEBUG nova.network.neutron [req-9e69838b-065e-433f-b77a-b0a409fd857d req-5c291ac1-ce10-4792-8735-e472e5dc6f6b service nova] [instance: 7c44d956-53d1-4a5f-9016-c0f3f0f348b4] Updated VIF entry in instance network info cache for port 8f686dbc-67bd-4f14-aec0-a6f9e7d8bfc5. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2214.662407] env[62824]: DEBUG nova.network.neutron [req-9e69838b-065e-433f-b77a-b0a409fd857d req-5c291ac1-ce10-4792-8735-e472e5dc6f6b service nova] [instance: 7c44d956-53d1-4a5f-9016-c0f3f0f348b4] Updating instance_info_cache with network_info: [{"id": "8f686dbc-67bd-4f14-aec0-a6f9e7d8bfc5", "address": "fa:16:3e:71:ae:7b", "network": {"id": "91425d93-4190-4e61-ae47-94efd278247d", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1364948114-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "461117110de7467fbf8ada5ee9978069", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c894ab55-c869-4530-9702-cb46d173ce94", "external-id": "nsx-vlan-transportzone-792", "segmentation_id": 792, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8f686dbc-67", "ovs_interfaceid": "8f686dbc-67bd-4f14-aec0-a6f9e7d8bfc5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2214.789623] env[62824]: DEBUG nova.compute.manager [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2214.805098] env[62824]: DEBUG oslo_vmware.api [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Task: {'id': task-2146579, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.480666} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2214.805809] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412/c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2214.806113] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2214.806395] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3a147526-2a85-4eb6-9d86-55ec05e1223e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2214.813807] env[62824]: DEBUG oslo_vmware.api [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Waiting for the task: (returnval){ [ 2214.813807] env[62824]: value = "task-2146584" [ 2214.813807] env[62824]: _type = "Task" [ 2214.813807] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2214.823944] env[62824]: DEBUG oslo_vmware.api [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Task: {'id': task-2146584, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2214.882262] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146580, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2214.914198] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146581, 'name': CreateVM_Task, 'duration_secs': 0.360921} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2214.914306] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7c44d956-53d1-4a5f-9016-c0f3f0f348b4] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2214.914954] env[62824]: DEBUG oslo_concurrency.lockutils [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2214.915149] env[62824]: DEBUG oslo_concurrency.lockutils [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2214.915466] env[62824]: DEBUG oslo_concurrency.lockutils [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2214.915739] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eb0c9017-49a8-4cc0-a82d-f75a29e1d29e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2214.920411] env[62824]: DEBUG oslo_vmware.api [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Waiting for the task: (returnval){ [ 2214.920411] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5249bae0-eebb-1089-b69f-9f01151bb4c3" [ 2214.920411] env[62824]: _type = "Task" [ 2214.920411] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2214.929846] env[62824]: DEBUG oslo_vmware.api [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5249bae0-eebb-1089-b69f-9f01151bb4c3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2214.973242] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e69cca00-b4f4-4019-be10-e9113c3b4c53 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "8c2ac679-6915-4149-aa51-99170b0df685" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.037s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2215.031475] env[62824]: DEBUG oslo_vmware.api [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146583, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.140738} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2215.031729] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2215.031887] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 58243465-ba24-4b37-a640-04ab24963230] Deleted contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2215.032303] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 58243465-ba24-4b37-a640-04ab24963230] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2215.064009] env[62824]: INFO nova.scheduler.client.report [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Deleted allocations for instance 58243465-ba24-4b37-a640-04ab24963230 [ 2215.136522] env[62824]: DEBUG nova.network.neutron [req-440b3d6e-8525-4853-9205-7c8d36ba7ca9 req-6212e975-ba2b-42db-9a62-f25fe1cd1c04 service nova] [instance: 1f162a63-8773-4c34-9c84-b2de1d5e9f50] Updated VIF entry in instance network info cache for port f5c2ffd1-3953-40e4-8d96-0542389767da. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2215.136522] env[62824]: DEBUG nova.network.neutron [req-440b3d6e-8525-4853-9205-7c8d36ba7ca9 req-6212e975-ba2b-42db-9a62-f25fe1cd1c04 service nova] [instance: 1f162a63-8773-4c34-9c84-b2de1d5e9f50] Updating instance_info_cache with network_info: [{"id": "f5c2ffd1-3953-40e4-8d96-0542389767da", "address": "fa:16:3e:36:ad:f3", "network": {"id": "91425d93-4190-4e61-ae47-94efd278247d", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1364948114-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "461117110de7467fbf8ada5ee9978069", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c894ab55-c869-4530-9702-cb46d173ce94", "external-id": "nsx-vlan-transportzone-792", "segmentation_id": 792, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf5c2ffd1-39", "ovs_interfaceid": "f5c2ffd1-3953-40e4-8d96-0542389767da", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2215.168169] env[62824]: DEBUG oslo_concurrency.lockutils [req-9e69838b-065e-433f-b77a-b0a409fd857d req-5c291ac1-ce10-4792-8735-e472e5dc6f6b service nova] Releasing lock "refresh_cache-7c44d956-53d1-4a5f-9016-c0f3f0f348b4" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2215.267020] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4d83a16-3f40-4fb7-9556-8678d075edfa {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2215.275035] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7de97e20-2c7b-43f3-b59d-7d778d9d9335 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2215.313377] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6724c9aa-6cc4-4c50-bd7d-76a195fd4caf {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2215.331016] env[62824]: DEBUG oslo_vmware.api [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Task: {'id': task-2146584, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066385} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2215.335953] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2215.335953] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2215.335953] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07dcea8d-bbec-4863-b1b2-fea36dc1832d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2215.338889] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bbc5cc0-e7fe-45ce-868b-92b2d83218fd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2215.353554] env[62824]: DEBUG nova.compute.provider_tree [None req-f8f74f45-c68b-4875-9f0f-fc836d564d52 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2215.374444] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412] Reconfiguring VM instance instance-00000056 to attach disk [datastore1] c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412/c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2215.375816] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9ce1309a-cca3-4fe8-a52d-63eb538fabfa {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2215.397588] env[62824]: DEBUG nova.compute.manager [req-1770164b-6dee-48bc-a0b5-ed41cd90d6bf req-5dab45f0-837b-4eb4-beba-309bb4c2832d service nova] [instance: 58243465-ba24-4b37-a640-04ab24963230] Received event network-vif-unplugged-63af4d3b-ce05-431b-84e1-e3e3fbdf673d {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2215.397993] env[62824]: DEBUG oslo_concurrency.lockutils [req-1770164b-6dee-48bc-a0b5-ed41cd90d6bf req-5dab45f0-837b-4eb4-beba-309bb4c2832d service nova] Acquiring lock "58243465-ba24-4b37-a640-04ab24963230-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2215.398272] env[62824]: DEBUG oslo_concurrency.lockutils [req-1770164b-6dee-48bc-a0b5-ed41cd90d6bf req-5dab45f0-837b-4eb4-beba-309bb4c2832d service nova] Lock "58243465-ba24-4b37-a640-04ab24963230-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2215.398471] env[62824]: DEBUG oslo_concurrency.lockutils [req-1770164b-6dee-48bc-a0b5-ed41cd90d6bf req-5dab45f0-837b-4eb4-beba-309bb4c2832d service nova] Lock "58243465-ba24-4b37-a640-04ab24963230-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2215.398652] env[62824]: DEBUG nova.compute.manager [req-1770164b-6dee-48bc-a0b5-ed41cd90d6bf req-5dab45f0-837b-4eb4-beba-309bb4c2832d service nova] [instance: 58243465-ba24-4b37-a640-04ab24963230] No waiting events found dispatching network-vif-unplugged-63af4d3b-ce05-431b-84e1-e3e3fbdf673d {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2215.398822] env[62824]: WARNING nova.compute.manager [req-1770164b-6dee-48bc-a0b5-ed41cd90d6bf req-5dab45f0-837b-4eb4-beba-309bb4c2832d service nova] [instance: 58243465-ba24-4b37-a640-04ab24963230] Received unexpected event network-vif-unplugged-63af4d3b-ce05-431b-84e1-e3e3fbdf673d for instance with vm_state shelved_offloaded and task_state None. [ 2215.398982] env[62824]: DEBUG nova.compute.manager [req-1770164b-6dee-48bc-a0b5-ed41cd90d6bf req-5dab45f0-837b-4eb4-beba-309bb4c2832d service nova] [instance: 58243465-ba24-4b37-a640-04ab24963230] Received event network-changed-63af4d3b-ce05-431b-84e1-e3e3fbdf673d {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2215.399151] env[62824]: DEBUG nova.compute.manager [req-1770164b-6dee-48bc-a0b5-ed41cd90d6bf req-5dab45f0-837b-4eb4-beba-309bb4c2832d service nova] [instance: 58243465-ba24-4b37-a640-04ab24963230] Refreshing instance network info cache due to event network-changed-63af4d3b-ce05-431b-84e1-e3e3fbdf673d. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2215.399339] env[62824]: DEBUG oslo_concurrency.lockutils [req-1770164b-6dee-48bc-a0b5-ed41cd90d6bf req-5dab45f0-837b-4eb4-beba-309bb4c2832d service nova] Acquiring lock "refresh_cache-58243465-ba24-4b37-a640-04ab24963230" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2215.399491] env[62824]: DEBUG oslo_concurrency.lockutils [req-1770164b-6dee-48bc-a0b5-ed41cd90d6bf req-5dab45f0-837b-4eb4-beba-309bb4c2832d service nova] Acquired lock "refresh_cache-58243465-ba24-4b37-a640-04ab24963230" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2215.399669] env[62824]: DEBUG nova.network.neutron [req-1770164b-6dee-48bc-a0b5-ed41cd90d6bf req-5dab45f0-837b-4eb4-beba-309bb4c2832d service nova] [instance: 58243465-ba24-4b37-a640-04ab24963230] Refreshing network info cache for port 63af4d3b-ce05-431b-84e1-e3e3fbdf673d {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2215.411778] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146580, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2215.412160] env[62824]: DEBUG oslo_vmware.api [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Waiting for the task: (returnval){ [ 2215.412160] env[62824]: value = "task-2146585" [ 2215.412160] env[62824]: _type = "Task" [ 2215.412160] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2215.425589] env[62824]: DEBUG oslo_vmware.api [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Task: {'id': task-2146585, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2215.438440] env[62824]: DEBUG oslo_vmware.api [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5249bae0-eebb-1089-b69f-9f01151bb4c3, 'name': SearchDatastore_Task, 'duration_secs': 0.066724} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2215.439130] env[62824]: DEBUG oslo_concurrency.lockutils [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2215.439130] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7c44d956-53d1-4a5f-9016-c0f3f0f348b4] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2215.439384] env[62824]: DEBUG oslo_concurrency.lockutils [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2215.439548] env[62824]: DEBUG oslo_concurrency.lockutils [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2215.439729] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2215.440056] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f2428dc5-9c47-4041-96bd-d4017818699d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2215.449252] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2215.449479] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2215.450408] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4be34d68-a42b-491b-8594-e2e747ef5055 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2215.456069] env[62824]: DEBUG oslo_vmware.api [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Waiting for the task: (returnval){ [ 2215.456069] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52eff5d0-037e-474b-5f47-f66199b0d231" [ 2215.456069] env[62824]: _type = "Task" [ 2215.456069] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2215.464422] env[62824]: DEBUG oslo_vmware.api [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52eff5d0-037e-474b-5f47-f66199b0d231, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2215.503562] env[62824]: DEBUG nova.network.neutron [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 853f3098-881c-402c-8fd7-26540d1ecd88] Successfully updated port: 03eb70c1-fc19-4e7c-874d-2f4fcd77fe6b {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2215.571464] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2215.638899] env[62824]: DEBUG oslo_concurrency.lockutils [req-440b3d6e-8525-4853-9205-7c8d36ba7ca9 req-6212e975-ba2b-42db-9a62-f25fe1cd1c04 service nova] Releasing lock "refresh_cache-1f162a63-8773-4c34-9c84-b2de1d5e9f50" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2215.881025] env[62824]: DEBUG nova.scheduler.client.report [None req-f8f74f45-c68b-4875-9f0f-fc836d564d52 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2215.892540] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146580, 'name': CreateVM_Task, 'duration_secs': 1.172251} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2215.892719] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1f162a63-8773-4c34-9c84-b2de1d5e9f50] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2215.894157] env[62824]: DEBUG oslo_concurrency.lockutils [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2215.894257] env[62824]: DEBUG oslo_concurrency.lockutils [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2215.894857] env[62824]: DEBUG oslo_concurrency.lockutils [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2215.894956] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-71e3c683-e750-4b5d-8801-08ef8b049076 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2215.901937] env[62824]: DEBUG oslo_vmware.api [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Waiting for the task: (returnval){ [ 2215.901937] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]525f79bc-4298-d1d0-4664-cf133744d4ef" [ 2215.901937] env[62824]: _type = "Task" [ 2215.901937] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2215.914936] env[62824]: DEBUG oslo_vmware.api [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]525f79bc-4298-d1d0-4664-cf133744d4ef, 'name': SearchDatastore_Task, 'duration_secs': 0.010725} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2215.918249] env[62824]: DEBUG oslo_concurrency.lockutils [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2215.918519] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 1f162a63-8773-4c34-9c84-b2de1d5e9f50] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2215.918756] env[62824]: DEBUG oslo_concurrency.lockutils [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2215.919486] env[62824]: DEBUG oslo_concurrency.lockutils [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2215.919486] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2215.919486] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-067e9767-aae7-4af5-b787-3d3a3e1e5ce2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2215.927592] env[62824]: DEBUG oslo_vmware.api [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Task: {'id': task-2146585, 'name': ReconfigVM_Task, 'duration_secs': 0.334562} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2215.928704] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412] Reconfigured VM instance instance-00000056 to attach disk [datastore1] c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412/c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2215.929376] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2215.929555] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2215.930251] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7d3b7e51-ac15-46db-a7dc-74f1bf3035ec {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2215.931993] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a422e2d9-040e-4957-830e-dd027be3dba0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2215.937234] env[62824]: DEBUG oslo_vmware.api [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Waiting for the task: (returnval){ [ 2215.937234] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5215839d-8039-4eb9-8606-a46347e88048" [ 2215.937234] env[62824]: _type = "Task" [ 2215.937234] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2215.943826] env[62824]: DEBUG oslo_vmware.api [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Waiting for the task: (returnval){ [ 2215.943826] env[62824]: value = "task-2146586" [ 2215.943826] env[62824]: _type = "Task" [ 2215.943826] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2215.951066] env[62824]: DEBUG oslo_vmware.api [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5215839d-8039-4eb9-8606-a46347e88048, 'name': SearchDatastore_Task, 'duration_secs': 0.009705} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2215.952108] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f2e7f99a-fbc7-4c19-b4af-0662b3ea4b5c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2215.957784] env[62824]: DEBUG oslo_vmware.api [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Task: {'id': task-2146586, 'name': Rename_Task} progress is 10%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2215.963012] env[62824]: DEBUG oslo_vmware.api [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Waiting for the task: (returnval){ [ 2215.963012] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]524c5e47-eae9-17d2-1037-9b8403716d6f" [ 2215.963012] env[62824]: _type = "Task" [ 2215.963012] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2215.970445] env[62824]: DEBUG oslo_vmware.api [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52eff5d0-037e-474b-5f47-f66199b0d231, 'name': SearchDatastore_Task, 'duration_secs': 0.008989} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2215.972287] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-07d4f28d-0691-43eb-be78-5b229c40ae87 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2215.980073] env[62824]: DEBUG oslo_vmware.api [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]524c5e47-eae9-17d2-1037-9b8403716d6f, 'name': SearchDatastore_Task, 'duration_secs': 0.00924} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2215.980073] env[62824]: DEBUG oslo_concurrency.lockutils [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2215.980554] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 1f162a63-8773-4c34-9c84-b2de1d5e9f50/1f162a63-8773-4c34-9c84-b2de1d5e9f50.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2215.980554] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-08863aa2-1987-4d7d-94aa-8cd1a5b9d2ea {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2215.985627] env[62824]: DEBUG oslo_vmware.api [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Waiting for the task: (returnval){ [ 2215.985627] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]523f14c5-ac6b-cee7-cd36-b4f19757ae7b" [ 2215.985627] env[62824]: _type = "Task" [ 2215.985627] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2215.993361] env[62824]: DEBUG oslo_vmware.api [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Waiting for the task: (returnval){ [ 2215.993361] env[62824]: value = "task-2146587" [ 2215.993361] env[62824]: _type = "Task" [ 2215.993361] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2215.999923] env[62824]: DEBUG oslo_vmware.api [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]523f14c5-ac6b-cee7-cd36-b4f19757ae7b, 'name': SearchDatastore_Task, 'duration_secs': 0.010191} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2216.000632] env[62824]: DEBUG oslo_concurrency.lockutils [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2216.000912] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 7c44d956-53d1-4a5f-9016-c0f3f0f348b4/7c44d956-53d1-4a5f-9016-c0f3f0f348b4.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2216.001584] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9fcf107c-f6b7-4ffd-83a1-9b32955ce4df {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2216.008067] env[62824]: DEBUG oslo_concurrency.lockutils [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Acquiring lock "refresh_cache-853f3098-881c-402c-8fd7-26540d1ecd88" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2216.008067] env[62824]: DEBUG oslo_concurrency.lockutils [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Acquired lock "refresh_cache-853f3098-881c-402c-8fd7-26540d1ecd88" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2216.008067] env[62824]: DEBUG nova.network.neutron [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 853f3098-881c-402c-8fd7-26540d1ecd88] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2216.009617] env[62824]: DEBUG oslo_vmware.api [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': task-2146587, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2216.014286] env[62824]: DEBUG oslo_vmware.api [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Waiting for the task: (returnval){ [ 2216.014286] env[62824]: value = "task-2146588" [ 2216.014286] env[62824]: _type = "Task" [ 2216.014286] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2216.026128] env[62824]: DEBUG oslo_vmware.api [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': task-2146588, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2216.210512] env[62824]: DEBUG nova.network.neutron [req-1770164b-6dee-48bc-a0b5-ed41cd90d6bf req-5dab45f0-837b-4eb4-beba-309bb4c2832d service nova] [instance: 58243465-ba24-4b37-a640-04ab24963230] Updated VIF entry in instance network info cache for port 63af4d3b-ce05-431b-84e1-e3e3fbdf673d. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2216.211127] env[62824]: DEBUG nova.network.neutron [req-1770164b-6dee-48bc-a0b5-ed41cd90d6bf req-5dab45f0-837b-4eb4-beba-309bb4c2832d service nova] [instance: 58243465-ba24-4b37-a640-04ab24963230] Updating instance_info_cache with network_info: [{"id": "63af4d3b-ce05-431b-84e1-e3e3fbdf673d", "address": "fa:16:3e:da:11:f0", "network": {"id": "b995934c-4934-4469-884c-73cb4de1796f", "bridge": null, "label": "tempest-DeleteServersTestJSON-1892307918-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c11aef7fb6247cb9b5272a6a063cd12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap63af4d3b-ce", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2216.388237] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f8f74f45-c68b-4875-9f0f-fc836d564d52 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.944s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2216.390803] env[62824]: DEBUG oslo_concurrency.lockutils [None req-602cdfe0-df66-45cb-bc5a-cc85cb34718d tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.717s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2216.391099] env[62824]: DEBUG nova.objects.instance [None req-602cdfe0-df66-45cb-bc5a-cc85cb34718d tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Lazy-loading 'resources' on Instance uuid 350cfd12-a8d5-4a5f-b3de-d795fa179dfd {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2216.411126] env[62824]: INFO nova.scheduler.client.report [None req-f8f74f45-c68b-4875-9f0f-fc836d564d52 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Deleted allocations for instance 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8 [ 2216.459297] env[62824]: DEBUG oslo_vmware.api [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Task: {'id': task-2146586, 'name': Rename_Task, 'duration_secs': 0.143672} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2216.459460] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2216.459684] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8fd8a213-2b00-4968-afc2-45f3d02617f6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2216.467021] env[62824]: DEBUG oslo_vmware.api [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Waiting for the task: (returnval){ [ 2216.467021] env[62824]: value = "task-2146589" [ 2216.467021] env[62824]: _type = "Task" [ 2216.467021] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2216.480586] env[62824]: DEBUG oslo_vmware.api [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Task: {'id': task-2146589, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2216.503393] env[62824]: DEBUG oslo_vmware.api [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': task-2146587, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2216.535972] env[62824]: DEBUG oslo_vmware.api [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': task-2146588, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2216.562690] env[62824]: DEBUG nova.network.neutron [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 853f3098-881c-402c-8fd7-26540d1ecd88] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2216.621677] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquiring lock "feda7def-7788-4a60-a56a-9353a1475c54" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2216.621950] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "feda7def-7788-4a60-a56a-9353a1475c54" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2216.693870] env[62824]: DEBUG oslo_concurrency.lockutils [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Acquiring lock "71e9733b-c84b-4501-8faf-a487f089b498" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2216.694532] env[62824]: DEBUG oslo_concurrency.lockutils [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Lock "71e9733b-c84b-4501-8faf-a487f089b498" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2216.713918] env[62824]: DEBUG oslo_concurrency.lockutils [req-1770164b-6dee-48bc-a0b5-ed41cd90d6bf req-5dab45f0-837b-4eb4-beba-309bb4c2832d service nova] Releasing lock "refresh_cache-58243465-ba24-4b37-a640-04ab24963230" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2216.736894] env[62824]: DEBUG nova.network.neutron [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 853f3098-881c-402c-8fd7-26540d1ecd88] Updating instance_info_cache with network_info: [{"id": "03eb70c1-fc19-4e7c-874d-2f4fcd77fe6b", "address": "fa:16:3e:8e:18:4c", "network": {"id": "aeacc588-ae73-49ec-9aa1-68f7d92055be", "bridge": "br-int", "label": "tempest-ImagesTestJSON-314094063-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10ee5fe93d584725ace90e76605a1b2e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1e7a4976-597e-4636-990e-6062b5faadee", "external-id": "nsx-vlan-transportzone-847", "segmentation_id": 847, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap03eb70c1-fc", "ovs_interfaceid": "03eb70c1-fc19-4e7c-874d-2f4fcd77fe6b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2216.922448] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f8f74f45-c68b-4875-9f0f-fc836d564d52 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Lock "1870b82a-9783-44ac-8de2-7b9ffc2a1bc8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.017s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2216.980959] env[62824]: DEBUG oslo_vmware.api [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Task: {'id': task-2146589, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2217.005188] env[62824]: DEBUG oslo_vmware.api [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': task-2146587, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.576533} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2217.005479] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 1f162a63-8773-4c34-9c84-b2de1d5e9f50/1f162a63-8773-4c34-9c84-b2de1d5e9f50.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2217.005693] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 1f162a63-8773-4c34-9c84-b2de1d5e9f50] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2217.005957] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b1cbaf05-bc82-4c06-b476-6145a8dd118e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2217.014073] env[62824]: DEBUG oslo_vmware.api [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Waiting for the task: (returnval){ [ 2217.014073] env[62824]: value = "task-2146590" [ 2217.014073] env[62824]: _type = "Task" [ 2217.014073] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2217.032187] env[62824]: DEBUG oslo_vmware.api [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': task-2146588, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.597066} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2217.032431] env[62824]: DEBUG oslo_vmware.api [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': task-2146590, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2217.032643] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 7c44d956-53d1-4a5f-9016-c0f3f0f348b4/7c44d956-53d1-4a5f-9016-c0f3f0f348b4.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2217.032855] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7c44d956-53d1-4a5f-9016-c0f3f0f348b4] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2217.033123] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c8b6bd0e-c306-4a59-b1e1-ea324b70a5ef {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2217.040153] env[62824]: DEBUG oslo_vmware.api [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Waiting for the task: (returnval){ [ 2217.040153] env[62824]: value = "task-2146591" [ 2217.040153] env[62824]: _type = "Task" [ 2217.040153] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2217.051059] env[62824]: DEBUG oslo_vmware.api [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': task-2146591, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2217.121785] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dd993d10-9788-4735-95a7-599a510439c1 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquiring lock "58243465-ba24-4b37-a640-04ab24963230" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2217.129346] env[62824]: DEBUG nova.compute.manager [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: feda7def-7788-4a60-a56a-9353a1475c54] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2217.190644] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c38dcded-8ec9-4fa4-a5cb-169f4641a516 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2217.197748] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ff8a232-2d14-4a2c-b12e-2d36651ab230 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2217.202269] env[62824]: DEBUG nova.compute.manager [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2217.232540] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fda58cd6-db75-41e7-8172-b4392e647c9d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2217.239834] env[62824]: DEBUG oslo_concurrency.lockutils [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Releasing lock "refresh_cache-853f3098-881c-402c-8fd7-26540d1ecd88" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2217.239926] env[62824]: DEBUG nova.compute.manager [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 853f3098-881c-402c-8fd7-26540d1ecd88] Instance network_info: |[{"id": "03eb70c1-fc19-4e7c-874d-2f4fcd77fe6b", "address": "fa:16:3e:8e:18:4c", "network": {"id": "aeacc588-ae73-49ec-9aa1-68f7d92055be", "bridge": "br-int", "label": "tempest-ImagesTestJSON-314094063-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10ee5fe93d584725ace90e76605a1b2e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1e7a4976-597e-4636-990e-6062b5faadee", "external-id": "nsx-vlan-transportzone-847", "segmentation_id": 847, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap03eb70c1-fc", "ovs_interfaceid": "03eb70c1-fc19-4e7c-874d-2f4fcd77fe6b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2217.240874] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 853f3098-881c-402c-8fd7-26540d1ecd88] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8e:18:4c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1e7a4976-597e-4636-990e-6062b5faadee', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '03eb70c1-fc19-4e7c-874d-2f4fcd77fe6b', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2217.250659] env[62824]: DEBUG oslo.service.loopingcall [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2217.252129] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e85939eb-efb8-47ea-b9db-6c32930bedfe {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2217.257155] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 853f3098-881c-402c-8fd7-26540d1ecd88] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2217.257526] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d09678e5-2e68-498c-a0f5-5c3fc51a9c01 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2217.282992] env[62824]: DEBUG nova.compute.provider_tree [None req-602cdfe0-df66-45cb-bc5a-cc85cb34718d tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2217.285417] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2217.285417] env[62824]: value = "task-2146592" [ 2217.285417] env[62824]: _type = "Task" [ 2217.285417] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2217.294272] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146592, 'name': CreateVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2217.410748] env[62824]: DEBUG nova.compute.manager [req-0028b076-be08-4b83-a3a4-064ff013b308 req-4590db6b-ed31-41a1-b8d9-85a3bbf12ce4 service nova] [instance: 853f3098-881c-402c-8fd7-26540d1ecd88] Received event network-vif-plugged-03eb70c1-fc19-4e7c-874d-2f4fcd77fe6b {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2217.411327] env[62824]: DEBUG oslo_concurrency.lockutils [req-0028b076-be08-4b83-a3a4-064ff013b308 req-4590db6b-ed31-41a1-b8d9-85a3bbf12ce4 service nova] Acquiring lock "853f3098-881c-402c-8fd7-26540d1ecd88-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2217.411327] env[62824]: DEBUG oslo_concurrency.lockutils [req-0028b076-be08-4b83-a3a4-064ff013b308 req-4590db6b-ed31-41a1-b8d9-85a3bbf12ce4 service nova] Lock "853f3098-881c-402c-8fd7-26540d1ecd88-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2217.411451] env[62824]: DEBUG oslo_concurrency.lockutils [req-0028b076-be08-4b83-a3a4-064ff013b308 req-4590db6b-ed31-41a1-b8d9-85a3bbf12ce4 service nova] Lock "853f3098-881c-402c-8fd7-26540d1ecd88-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2217.411650] env[62824]: DEBUG nova.compute.manager [req-0028b076-be08-4b83-a3a4-064ff013b308 req-4590db6b-ed31-41a1-b8d9-85a3bbf12ce4 service nova] [instance: 853f3098-881c-402c-8fd7-26540d1ecd88] No waiting events found dispatching network-vif-plugged-03eb70c1-fc19-4e7c-874d-2f4fcd77fe6b {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2217.411698] env[62824]: WARNING nova.compute.manager [req-0028b076-be08-4b83-a3a4-064ff013b308 req-4590db6b-ed31-41a1-b8d9-85a3bbf12ce4 service nova] [instance: 853f3098-881c-402c-8fd7-26540d1ecd88] Received unexpected event network-vif-plugged-03eb70c1-fc19-4e7c-874d-2f4fcd77fe6b for instance with vm_state building and task_state spawning. [ 2217.411864] env[62824]: DEBUG nova.compute.manager [req-0028b076-be08-4b83-a3a4-064ff013b308 req-4590db6b-ed31-41a1-b8d9-85a3bbf12ce4 service nova] [instance: 853f3098-881c-402c-8fd7-26540d1ecd88] Received event network-changed-03eb70c1-fc19-4e7c-874d-2f4fcd77fe6b {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2217.412032] env[62824]: DEBUG nova.compute.manager [req-0028b076-be08-4b83-a3a4-064ff013b308 req-4590db6b-ed31-41a1-b8d9-85a3bbf12ce4 service nova] [instance: 853f3098-881c-402c-8fd7-26540d1ecd88] Refreshing instance network info cache due to event network-changed-03eb70c1-fc19-4e7c-874d-2f4fcd77fe6b. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2217.412224] env[62824]: DEBUG oslo_concurrency.lockutils [req-0028b076-be08-4b83-a3a4-064ff013b308 req-4590db6b-ed31-41a1-b8d9-85a3bbf12ce4 service nova] Acquiring lock "refresh_cache-853f3098-881c-402c-8fd7-26540d1ecd88" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2217.412357] env[62824]: DEBUG oslo_concurrency.lockutils [req-0028b076-be08-4b83-a3a4-064ff013b308 req-4590db6b-ed31-41a1-b8d9-85a3bbf12ce4 service nova] Acquired lock "refresh_cache-853f3098-881c-402c-8fd7-26540d1ecd88" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2217.412517] env[62824]: DEBUG nova.network.neutron [req-0028b076-be08-4b83-a3a4-064ff013b308 req-4590db6b-ed31-41a1-b8d9-85a3bbf12ce4 service nova] [instance: 853f3098-881c-402c-8fd7-26540d1ecd88] Refreshing network info cache for port 03eb70c1-fc19-4e7c-874d-2f4fcd77fe6b {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2217.478587] env[62824]: DEBUG oslo_vmware.api [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Task: {'id': task-2146589, 'name': PowerOnVM_Task, 'duration_secs': 0.520831} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2217.478867] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2217.479088] env[62824]: INFO nova.compute.manager [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412] Took 18.00 seconds to spawn the instance on the hypervisor. [ 2217.479518] env[62824]: DEBUG nova.compute.manager [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2217.480531] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a70e2d6b-a19d-4562-88eb-97a0b90d3156 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2217.527061] env[62824]: DEBUG oslo_vmware.api [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': task-2146590, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.111213} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2217.527061] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 1f162a63-8773-4c34-9c84-b2de1d5e9f50] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2217.527844] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25f2a560-67f9-45c5-a31c-1d18b49dae82 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2217.552839] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 1f162a63-8773-4c34-9c84-b2de1d5e9f50] Reconfiguring VM instance instance-00000057 to attach disk [datastore2] 1f162a63-8773-4c34-9c84-b2de1d5e9f50/1f162a63-8773-4c34-9c84-b2de1d5e9f50.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2217.557725] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d05906c9-a1d6-41c9-84c8-c4a6b17d30fc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2217.577693] env[62824]: DEBUG oslo_vmware.api [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': task-2146591, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065973} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2217.579101] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7c44d956-53d1-4a5f-9016-c0f3f0f348b4] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2217.579490] env[62824]: DEBUG oslo_vmware.api [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Waiting for the task: (returnval){ [ 2217.579490] env[62824]: value = "task-2146593" [ 2217.579490] env[62824]: _type = "Task" [ 2217.579490] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2217.580215] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eef3f671-a19a-48cb-b527-e9e4ccd078a6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2217.610777] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7c44d956-53d1-4a5f-9016-c0f3f0f348b4] Reconfiguring VM instance instance-00000058 to attach disk [datastore1] 7c44d956-53d1-4a5f-9016-c0f3f0f348b4/7c44d956-53d1-4a5f-9016-c0f3f0f348b4.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2217.614471] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e1f367bd-315b-4da6-9762-b1eae82ed54a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2217.632124] env[62824]: DEBUG oslo_vmware.api [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': task-2146593, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2217.637281] env[62824]: DEBUG oslo_vmware.api [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Waiting for the task: (returnval){ [ 2217.637281] env[62824]: value = "task-2146594" [ 2217.637281] env[62824]: _type = "Task" [ 2217.637281] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2217.648507] env[62824]: DEBUG oslo_vmware.api [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': task-2146594, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2217.661952] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2217.727917] env[62824]: DEBUG oslo_concurrency.lockutils [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2217.788570] env[62824]: DEBUG nova.scheduler.client.report [None req-602cdfe0-df66-45cb-bc5a-cc85cb34718d tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2217.806264] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146592, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2218.008553] env[62824]: INFO nova.compute.manager [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412] Took 42.67 seconds to build instance. [ 2218.099610] env[62824]: DEBUG oslo_vmware.api [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': task-2146593, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2218.152876] env[62824]: DEBUG oslo_vmware.api [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': task-2146594, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2218.299604] env[62824]: DEBUG oslo_concurrency.lockutils [None req-602cdfe0-df66-45cb-bc5a-cc85cb34718d tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.909s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2218.303242] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d85dcf93-e801-4553-b0ef-b90b42fed644 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.486s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2218.303242] env[62824]: DEBUG nova.objects.instance [None req-d85dcf93-e801-4553-b0ef-b90b42fed644 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Lazy-loading 'resources' on Instance uuid 0c9fb5cc-de69-46d8-9962-98e0a84e33c3 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2218.312121] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146592, 'name': CreateVM_Task, 'duration_secs': 0.764221} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2218.313170] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 853f3098-881c-402c-8fd7-26540d1ecd88] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2218.313374] env[62824]: DEBUG oslo_concurrency.lockutils [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2218.313733] env[62824]: DEBUG oslo_concurrency.lockutils [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2218.314209] env[62824]: DEBUG oslo_concurrency.lockutils [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2218.315477] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c4b6e758-bedf-4428-a4cc-9db2c6be1688 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2218.322443] env[62824]: DEBUG oslo_vmware.api [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the task: (returnval){ [ 2218.322443] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52832377-3bc6-214e-c99a-5aafd94d8060" [ 2218.322443] env[62824]: _type = "Task" [ 2218.322443] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2218.331633] env[62824]: DEBUG oslo_vmware.api [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52832377-3bc6-214e-c99a-5aafd94d8060, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2218.332997] env[62824]: INFO nova.scheduler.client.report [None req-602cdfe0-df66-45cb-bc5a-cc85cb34718d tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Deleted allocations for instance 350cfd12-a8d5-4a5f-b3de-d795fa179dfd [ 2218.375737] env[62824]: DEBUG nova.network.neutron [req-0028b076-be08-4b83-a3a4-064ff013b308 req-4590db6b-ed31-41a1-b8d9-85a3bbf12ce4 service nova] [instance: 853f3098-881c-402c-8fd7-26540d1ecd88] Updated VIF entry in instance network info cache for port 03eb70c1-fc19-4e7c-874d-2f4fcd77fe6b. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2218.376127] env[62824]: DEBUG nova.network.neutron [req-0028b076-be08-4b83-a3a4-064ff013b308 req-4590db6b-ed31-41a1-b8d9-85a3bbf12ce4 service nova] [instance: 853f3098-881c-402c-8fd7-26540d1ecd88] Updating instance_info_cache with network_info: [{"id": "03eb70c1-fc19-4e7c-874d-2f4fcd77fe6b", "address": "fa:16:3e:8e:18:4c", "network": {"id": "aeacc588-ae73-49ec-9aa1-68f7d92055be", "bridge": "br-int", "label": "tempest-ImagesTestJSON-314094063-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10ee5fe93d584725ace90e76605a1b2e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1e7a4976-597e-4636-990e-6062b5faadee", "external-id": "nsx-vlan-transportzone-847", "segmentation_id": 847, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap03eb70c1-fc", "ovs_interfaceid": "03eb70c1-fc19-4e7c-874d-2f4fcd77fe6b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2218.511324] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ffb4fbb4-dcef-47e3-a9d5-d5223bc55ca2 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Lock "c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 44.177s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2218.597384] env[62824]: DEBUG oslo_vmware.api [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': task-2146593, 'name': ReconfigVM_Task, 'duration_secs': 0.705427} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2218.597384] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 1f162a63-8773-4c34-9c84-b2de1d5e9f50] Reconfigured VM instance instance-00000057 to attach disk [datastore2] 1f162a63-8773-4c34-9c84-b2de1d5e9f50/1f162a63-8773-4c34-9c84-b2de1d5e9f50.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2218.598133] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-96280eed-e1e3-43e4-93a2-e48906241f13 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2218.605859] env[62824]: DEBUG oslo_vmware.api [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Waiting for the task: (returnval){ [ 2218.605859] env[62824]: value = "task-2146595" [ 2218.605859] env[62824]: _type = "Task" [ 2218.605859] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2218.615888] env[62824]: DEBUG oslo_vmware.api [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': task-2146595, 'name': Rename_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2218.652116] env[62824]: DEBUG oslo_vmware.api [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': task-2146594, 'name': ReconfigVM_Task, 'duration_secs': 0.689349} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2218.652404] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7c44d956-53d1-4a5f-9016-c0f3f0f348b4] Reconfigured VM instance instance-00000058 to attach disk [datastore1] 7c44d956-53d1-4a5f-9016-c0f3f0f348b4/7c44d956-53d1-4a5f-9016-c0f3f0f348b4.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2218.653080] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-79a7d232-8dc8-41bb-92ac-ec58368662df {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2218.666729] env[62824]: DEBUG oslo_vmware.api [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Waiting for the task: (returnval){ [ 2218.666729] env[62824]: value = "task-2146596" [ 2218.666729] env[62824]: _type = "Task" [ 2218.666729] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2218.679963] env[62824]: DEBUG oslo_vmware.api [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': task-2146596, 'name': Rename_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2218.839434] env[62824]: DEBUG oslo_vmware.api [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52832377-3bc6-214e-c99a-5aafd94d8060, 'name': SearchDatastore_Task, 'duration_secs': 0.012926} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2218.839945] env[62824]: DEBUG oslo_concurrency.lockutils [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2218.840320] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 853f3098-881c-402c-8fd7-26540d1ecd88] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2218.840742] env[62824]: DEBUG oslo_concurrency.lockutils [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2218.840968] env[62824]: DEBUG oslo_concurrency.lockutils [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2218.841227] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2218.847895] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6d9add78-0f5c-4fd3-b6db-08a83824be26 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2218.851063] env[62824]: DEBUG oslo_concurrency.lockutils [None req-602cdfe0-df66-45cb-bc5a-cc85cb34718d tempest-SecurityGroupsTestJSON-2075934480 tempest-SecurityGroupsTestJSON-2075934480-project-member] Lock "350cfd12-a8d5-4a5f-b3de-d795fa179dfd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.216s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2218.860416] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2218.860964] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2218.861682] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-30b94ad1-f47b-4399-9b79-2202fd8505e1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2218.872298] env[62824]: DEBUG oslo_vmware.api [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the task: (returnval){ [ 2218.872298] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52202b83-743e-00e9-fa8b-f064b251e5cd" [ 2218.872298] env[62824]: _type = "Task" [ 2218.872298] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2218.879620] env[62824]: DEBUG oslo_concurrency.lockutils [req-0028b076-be08-4b83-a3a4-064ff013b308 req-4590db6b-ed31-41a1-b8d9-85a3bbf12ce4 service nova] Releasing lock "refresh_cache-853f3098-881c-402c-8fd7-26540d1ecd88" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2218.887744] env[62824]: DEBUG oslo_vmware.api [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52202b83-743e-00e9-fa8b-f064b251e5cd, 'name': SearchDatastore_Task, 'duration_secs': 0.012877} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2218.888636] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cf9cad19-fe90-4d1f-94da-3c60dec786ac {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2218.894396] env[62824]: DEBUG oslo_vmware.api [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the task: (returnval){ [ 2218.894396] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52b27431-49d8-85ec-e49f-2326a48ebf2e" [ 2218.894396] env[62824]: _type = "Task" [ 2218.894396] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2218.910025] env[62824]: DEBUG oslo_vmware.api [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52b27431-49d8-85ec-e49f-2326a48ebf2e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2219.099172] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c42f9dd-30d9-4ff3-a8da-052462a71501 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2219.109964] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4144f865-3ffc-4dce-bb91-28732679b081 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2219.119687] env[62824]: DEBUG oslo_vmware.api [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': task-2146595, 'name': Rename_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2219.146529] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7069d72-0799-47f4-b6f5-13a5193e536c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2219.154609] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a60f202-1a8a-4a96-a8af-5080bab7cfa7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2219.168010] env[62824]: DEBUG nova.compute.provider_tree [None req-d85dcf93-e801-4553-b0ef-b90b42fed644 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2219.177037] env[62824]: DEBUG oslo_vmware.api [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': task-2146596, 'name': Rename_Task, 'duration_secs': 0.271393} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2219.177864] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7c44d956-53d1-4a5f-9016-c0f3f0f348b4] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2219.178126] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0cce744d-a7e5-4a73-8354-415bd87f461e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2219.184957] env[62824]: DEBUG oslo_vmware.api [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Waiting for the task: (returnval){ [ 2219.184957] env[62824]: value = "task-2146597" [ 2219.184957] env[62824]: _type = "Task" [ 2219.184957] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2219.192702] env[62824]: DEBUG oslo_vmware.api [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': task-2146597, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2219.409166] env[62824]: DEBUG oslo_vmware.api [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52b27431-49d8-85ec-e49f-2326a48ebf2e, 'name': SearchDatastore_Task, 'duration_secs': 0.023077} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2219.409441] env[62824]: DEBUG oslo_concurrency.lockutils [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2219.409696] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 853f3098-881c-402c-8fd7-26540d1ecd88/853f3098-881c-402c-8fd7-26540d1ecd88.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2219.410479] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-49218ed6-b09e-4939-b5cb-7f1979cfb92d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2219.418122] env[62824]: DEBUG oslo_vmware.api [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the task: (returnval){ [ 2219.418122] env[62824]: value = "task-2146598" [ 2219.418122] env[62824]: _type = "Task" [ 2219.418122] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2219.426901] env[62824]: DEBUG oslo_vmware.api [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146598, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2219.484338] env[62824]: DEBUG nova.compute.manager [req-f87956d8-223d-499c-bbf3-83e4b508430c req-86f6de97-8e9d-4383-8421-eb61cb386aad service nova] [instance: c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412] Received event network-changed-0f67546a-a44f-4c20-96e8-e7b09334b221 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2219.484595] env[62824]: DEBUG nova.compute.manager [req-f87956d8-223d-499c-bbf3-83e4b508430c req-86f6de97-8e9d-4383-8421-eb61cb386aad service nova] [instance: c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412] Refreshing instance network info cache due to event network-changed-0f67546a-a44f-4c20-96e8-e7b09334b221. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2219.484802] env[62824]: DEBUG oslo_concurrency.lockutils [req-f87956d8-223d-499c-bbf3-83e4b508430c req-86f6de97-8e9d-4383-8421-eb61cb386aad service nova] Acquiring lock "refresh_cache-c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2219.484966] env[62824]: DEBUG oslo_concurrency.lockutils [req-f87956d8-223d-499c-bbf3-83e4b508430c req-86f6de97-8e9d-4383-8421-eb61cb386aad service nova] Acquired lock "refresh_cache-c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2219.485214] env[62824]: DEBUG nova.network.neutron [req-f87956d8-223d-499c-bbf3-83e4b508430c req-86f6de97-8e9d-4383-8421-eb61cb386aad service nova] [instance: c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412] Refreshing network info cache for port 0f67546a-a44f-4c20-96e8-e7b09334b221 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2219.617677] env[62824]: DEBUG oslo_vmware.api [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': task-2146595, 'name': Rename_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2219.673367] env[62824]: DEBUG nova.scheduler.client.report [None req-d85dcf93-e801-4553-b0ef-b90b42fed644 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2219.698416] env[62824]: DEBUG oslo_vmware.api [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': task-2146597, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2219.929324] env[62824]: DEBUG oslo_vmware.api [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146598, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2220.118535] env[62824]: DEBUG oslo_vmware.api [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': task-2146595, 'name': Rename_Task, 'duration_secs': 1.21704} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2220.118904] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 1f162a63-8773-4c34-9c84-b2de1d5e9f50] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2220.119213] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-24014afb-1b2a-44fc-87a0-19b7be23016f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2220.129479] env[62824]: DEBUG oslo_vmware.api [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Waiting for the task: (returnval){ [ 2220.129479] env[62824]: value = "task-2146599" [ 2220.129479] env[62824]: _type = "Task" [ 2220.129479] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2220.138922] env[62824]: DEBUG oslo_vmware.api [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': task-2146599, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2220.179082] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d85dcf93-e801-4553-b0ef-b90b42fed644 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.876s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2220.181769] env[62824]: DEBUG oslo_concurrency.lockutils [None req-56a95388-af07-48c0-bd95-c65c098831e2 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.365s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2220.182032] env[62824]: DEBUG nova.objects.instance [None req-56a95388-af07-48c0-bd95-c65c098831e2 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Lazy-loading 'resources' on Instance uuid 07ce931f-45ef-409b-b714-9f1cd47a3a88 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2220.196023] env[62824]: DEBUG oslo_vmware.api [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': task-2146597, 'name': PowerOnVM_Task, 'duration_secs': 0.524555} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2220.196023] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7c44d956-53d1-4a5f-9016-c0f3f0f348b4] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2220.196023] env[62824]: INFO nova.compute.manager [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7c44d956-53d1-4a5f-9016-c0f3f0f348b4] Took 8.52 seconds to spawn the instance on the hypervisor. [ 2220.196023] env[62824]: DEBUG nova.compute.manager [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7c44d956-53d1-4a5f-9016-c0f3f0f348b4] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2220.196023] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96f9e181-bb51-406e-9ef7-85f8e257e97f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2220.208273] env[62824]: INFO nova.scheduler.client.report [None req-d85dcf93-e801-4553-b0ef-b90b42fed644 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Deleted allocations for instance 0c9fb5cc-de69-46d8-9962-98e0a84e33c3 [ 2220.250385] env[62824]: DEBUG nova.network.neutron [req-f87956d8-223d-499c-bbf3-83e4b508430c req-86f6de97-8e9d-4383-8421-eb61cb386aad service nova] [instance: c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412] Updated VIF entry in instance network info cache for port 0f67546a-a44f-4c20-96e8-e7b09334b221. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2220.250774] env[62824]: DEBUG nova.network.neutron [req-f87956d8-223d-499c-bbf3-83e4b508430c req-86f6de97-8e9d-4383-8421-eb61cb386aad service nova] [instance: c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412] Updating instance_info_cache with network_info: [{"id": "0f67546a-a44f-4c20-96e8-e7b09334b221", "address": "fa:16:3e:ac:62:ba", "network": {"id": "ae1dfabd-223a-4d4c-a8c4-fe65a83dd9c0", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-797102932-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.145", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "102d2f886fd44e1da6780e48e36f1b62", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "517421c3-bea0-419c-ab0b-987815e5d160", "external-id": "nsx-vlan-transportzone-68", "segmentation_id": 68, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0f67546a-a4", "ovs_interfaceid": "0f67546a-a44f-4c20-96e8-e7b09334b221", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2220.430213] env[62824]: DEBUG oslo_vmware.api [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146598, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.757725} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2220.430898] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 853f3098-881c-402c-8fd7-26540d1ecd88/853f3098-881c-402c-8fd7-26540d1ecd88.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2220.431472] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 853f3098-881c-402c-8fd7-26540d1ecd88] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2220.434240] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e168821c-f1df-4874-84ec-5019999498bc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2220.441018] env[62824]: DEBUG oslo_vmware.api [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the task: (returnval){ [ 2220.441018] env[62824]: value = "task-2146600" [ 2220.441018] env[62824]: _type = "Task" [ 2220.441018] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2220.447387] env[62824]: DEBUG oslo_vmware.api [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146600, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2220.640067] env[62824]: DEBUG oslo_vmware.api [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': task-2146599, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2220.728643] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d85dcf93-e801-4553-b0ef-b90b42fed644 tempest-ImagesOneServerNegativeTestJSON-1714663442 tempest-ImagesOneServerNegativeTestJSON-1714663442-project-member] Lock "0c9fb5cc-de69-46d8-9962-98e0a84e33c3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.811s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2220.728643] env[62824]: INFO nova.compute.manager [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7c44d956-53d1-4a5f-9016-c0f3f0f348b4] Took 32.57 seconds to build instance. [ 2220.754131] env[62824]: DEBUG oslo_concurrency.lockutils [req-f87956d8-223d-499c-bbf3-83e4b508430c req-86f6de97-8e9d-4383-8421-eb61cb386aad service nova] Releasing lock "refresh_cache-c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2220.949690] env[62824]: DEBUG oslo_vmware.api [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146600, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.082558} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2220.950006] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 853f3098-881c-402c-8fd7-26540d1ecd88] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2220.950780] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0fcaa4f-a8cf-4b0e-9987-44b6b62ff9c6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2220.974656] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 853f3098-881c-402c-8fd7-26540d1ecd88] Reconfiguring VM instance instance-00000059 to attach disk [datastore1] 853f3098-881c-402c-8fd7-26540d1ecd88/853f3098-881c-402c-8fd7-26540d1ecd88.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2220.977440] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-11038e23-0862-486d-9404-a473d234aae4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2221.000072] env[62824]: DEBUG oslo_vmware.api [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the task: (returnval){ [ 2221.000072] env[62824]: value = "task-2146601" [ 2221.000072] env[62824]: _type = "Task" [ 2221.000072] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2221.006120] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8133526c-050c-425b-8cd2-1f906ccb32f2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2221.017289] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bd9dd1d-3f29-47ec-aec4-e41c84c8d6e3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2221.021408] env[62824]: DEBUG oslo_vmware.api [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146601, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2221.058659] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44fb072c-b64c-405c-a67e-50e6ea6048fe {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2221.071938] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ed234dd-525a-4582-b538-9da01928d126 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2221.091026] env[62824]: DEBUG nova.compute.provider_tree [None req-56a95388-af07-48c0-bd95-c65c098831e2 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2221.144953] env[62824]: DEBUG oslo_vmware.api [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': task-2146599, 'name': PowerOnVM_Task, 'duration_secs': 0.665877} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2221.145409] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 1f162a63-8773-4c34-9c84-b2de1d5e9f50] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2221.145875] env[62824]: INFO nova.compute.manager [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 1f162a63-8773-4c34-9c84-b2de1d5e9f50] Took 11.86 seconds to spawn the instance on the hypervisor. [ 2221.145875] env[62824]: DEBUG nova.compute.manager [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 1f162a63-8773-4c34-9c84-b2de1d5e9f50] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2221.148383] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7293e377-c49e-4530-a256-916fde7c8c2c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2221.230085] env[62824]: DEBUG oslo_concurrency.lockutils [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Lock "7c44d956-53d1-4a5f-9016-c0f3f0f348b4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.080s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2221.515413] env[62824]: DEBUG oslo_vmware.api [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146601, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2221.594855] env[62824]: DEBUG nova.scheduler.client.report [None req-56a95388-af07-48c0-bd95-c65c098831e2 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2221.670107] env[62824]: INFO nova.compute.manager [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 1f162a63-8773-4c34-9c84-b2de1d5e9f50] Took 33.54 seconds to build instance. [ 2222.011113] env[62824]: DEBUG oslo_vmware.api [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146601, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2222.101400] env[62824]: DEBUG oslo_concurrency.lockutils [None req-56a95388-af07-48c0-bd95-c65c098831e2 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.920s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2222.103593] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.771s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2222.105099] env[62824]: INFO nova.compute.claims [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2222.125384] env[62824]: INFO nova.scheduler.client.report [None req-56a95388-af07-48c0-bd95-c65c098831e2 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Deleted allocations for instance 07ce931f-45ef-409b-b714-9f1cd47a3a88 [ 2222.172223] env[62824]: DEBUG oslo_concurrency.lockutils [None req-170ee4c9-98ec-43da-b445-f1021f8188b6 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Lock "1f162a63-8773-4c34-9c84-b2de1d5e9f50" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.052s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2222.511732] env[62824]: DEBUG oslo_vmware.api [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146601, 'name': ReconfigVM_Task, 'duration_secs': 1.067932} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2222.512063] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 853f3098-881c-402c-8fd7-26540d1ecd88] Reconfigured VM instance instance-00000059 to attach disk [datastore1] 853f3098-881c-402c-8fd7-26540d1ecd88/853f3098-881c-402c-8fd7-26540d1ecd88.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2222.512849] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-30342a55-4be7-45cd-a14a-83b78175ef1f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2222.519519] env[62824]: DEBUG oslo_vmware.api [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the task: (returnval){ [ 2222.519519] env[62824]: value = "task-2146602" [ 2222.519519] env[62824]: _type = "Task" [ 2222.519519] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2222.527697] env[62824]: DEBUG oslo_vmware.api [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146602, 'name': Rename_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2222.634974] env[62824]: DEBUG oslo_concurrency.lockutils [None req-56a95388-af07-48c0-bd95-c65c098831e2 tempest-ServersAdminTestJSON-1079836281 tempest-ServersAdminTestJSON-1079836281-project-member] Lock "07ce931f-45ef-409b-b714-9f1cd47a3a88" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.725s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2222.960281] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c7d29f0c-9cc1-46d5-94e7-68f8528e0f81 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Acquiring lock "1f162a63-8773-4c34-9c84-b2de1d5e9f50" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2222.960541] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c7d29f0c-9cc1-46d5-94e7-68f8528e0f81 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Lock "1f162a63-8773-4c34-9c84-b2de1d5e9f50" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2222.960763] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c7d29f0c-9cc1-46d5-94e7-68f8528e0f81 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Acquiring lock "1f162a63-8773-4c34-9c84-b2de1d5e9f50-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2222.960947] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c7d29f0c-9cc1-46d5-94e7-68f8528e0f81 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Lock "1f162a63-8773-4c34-9c84-b2de1d5e9f50-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2222.961134] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c7d29f0c-9cc1-46d5-94e7-68f8528e0f81 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Lock "1f162a63-8773-4c34-9c84-b2de1d5e9f50-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2222.963407] env[62824]: INFO nova.compute.manager [None req-c7d29f0c-9cc1-46d5-94e7-68f8528e0f81 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 1f162a63-8773-4c34-9c84-b2de1d5e9f50] Terminating instance [ 2223.032199] env[62824]: DEBUG oslo_vmware.api [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146602, 'name': Rename_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2223.043255] env[62824]: DEBUG oslo_concurrency.lockutils [None req-742366bb-e991-49d9-ad5f-cebf8cbed982 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Acquiring lock "7c44d956-53d1-4a5f-9016-c0f3f0f348b4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2223.043522] env[62824]: DEBUG oslo_concurrency.lockutils [None req-742366bb-e991-49d9-ad5f-cebf8cbed982 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Lock "7c44d956-53d1-4a5f-9016-c0f3f0f348b4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2223.044018] env[62824]: DEBUG oslo_concurrency.lockutils [None req-742366bb-e991-49d9-ad5f-cebf8cbed982 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Acquiring lock "7c44d956-53d1-4a5f-9016-c0f3f0f348b4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2223.044018] env[62824]: DEBUG oslo_concurrency.lockutils [None req-742366bb-e991-49d9-ad5f-cebf8cbed982 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Lock "7c44d956-53d1-4a5f-9016-c0f3f0f348b4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2223.044156] env[62824]: DEBUG oslo_concurrency.lockutils [None req-742366bb-e991-49d9-ad5f-cebf8cbed982 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Lock "7c44d956-53d1-4a5f-9016-c0f3f0f348b4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2223.047015] env[62824]: INFO nova.compute.manager [None req-742366bb-e991-49d9-ad5f-cebf8cbed982 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7c44d956-53d1-4a5f-9016-c0f3f0f348b4] Terminating instance [ 2223.375487] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-820d1eac-ca1b-4ab1-8d63-4af0c1641168 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2223.385603] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5878a58-c4e1-4efc-8871-9f4271d0d05a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2223.426040] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f21cae57-7b63-4f31-be45-709c9fa4ba2a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2223.438015] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f179c3d-c23b-48ec-bb3a-8027fa224ba8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2223.449432] env[62824]: DEBUG nova.compute.provider_tree [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2223.466752] env[62824]: DEBUG nova.compute.manager [None req-c7d29f0c-9cc1-46d5-94e7-68f8528e0f81 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 1f162a63-8773-4c34-9c84-b2de1d5e9f50] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2223.466971] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c7d29f0c-9cc1-46d5-94e7-68f8528e0f81 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 1f162a63-8773-4c34-9c84-b2de1d5e9f50] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2223.467889] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b17e114c-4774-4999-bb55-d16180cc8568 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2223.475750] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7d29f0c-9cc1-46d5-94e7-68f8528e0f81 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 1f162a63-8773-4c34-9c84-b2de1d5e9f50] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2223.475988] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-45cf94bb-0757-4ac5-bb48-17005241b808 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2223.482882] env[62824]: DEBUG oslo_vmware.api [None req-c7d29f0c-9cc1-46d5-94e7-68f8528e0f81 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Waiting for the task: (returnval){ [ 2223.482882] env[62824]: value = "task-2146603" [ 2223.482882] env[62824]: _type = "Task" [ 2223.482882] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2223.490622] env[62824]: DEBUG oslo_vmware.api [None req-c7d29f0c-9cc1-46d5-94e7-68f8528e0f81 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': task-2146603, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2223.531313] env[62824]: DEBUG oslo_vmware.api [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146602, 'name': Rename_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2223.552126] env[62824]: DEBUG nova.compute.manager [None req-742366bb-e991-49d9-ad5f-cebf8cbed982 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7c44d956-53d1-4a5f-9016-c0f3f0f348b4] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2223.552380] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-742366bb-e991-49d9-ad5f-cebf8cbed982 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7c44d956-53d1-4a5f-9016-c0f3f0f348b4] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2223.553175] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-293edf7b-78f1-49f4-87c3-bd81b0d581bc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2223.560278] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-742366bb-e991-49d9-ad5f-cebf8cbed982 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7c44d956-53d1-4a5f-9016-c0f3f0f348b4] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2223.560502] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ffcd1e40-068f-4b5f-9d08-aebcb145cf99 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2223.566672] env[62824]: DEBUG oslo_vmware.api [None req-742366bb-e991-49d9-ad5f-cebf8cbed982 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Waiting for the task: (returnval){ [ 2223.566672] env[62824]: value = "task-2146604" [ 2223.566672] env[62824]: _type = "Task" [ 2223.566672] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2223.574482] env[62824]: DEBUG oslo_vmware.api [None req-742366bb-e991-49d9-ad5f-cebf8cbed982 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': task-2146604, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2223.952251] env[62824]: DEBUG nova.scheduler.client.report [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2223.996287] env[62824]: DEBUG oslo_vmware.api [None req-c7d29f0c-9cc1-46d5-94e7-68f8528e0f81 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': task-2146603, 'name': PowerOffVM_Task, 'duration_secs': 0.258865} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2223.996630] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7d29f0c-9cc1-46d5-94e7-68f8528e0f81 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 1f162a63-8773-4c34-9c84-b2de1d5e9f50] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2223.996863] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c7d29f0c-9cc1-46d5-94e7-68f8528e0f81 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 1f162a63-8773-4c34-9c84-b2de1d5e9f50] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2223.997437] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9b633c89-7985-42e4-a208-6bb7a16228a3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2224.031229] env[62824]: DEBUG oslo_vmware.api [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146602, 'name': Rename_Task, 'duration_secs': 1.405839} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2224.031489] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 853f3098-881c-402c-8fd7-26540d1ecd88] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2224.031730] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-817abef6-bf28-447d-b810-39838e79889b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2224.040047] env[62824]: DEBUG oslo_vmware.api [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the task: (returnval){ [ 2224.040047] env[62824]: value = "task-2146606" [ 2224.040047] env[62824]: _type = "Task" [ 2224.040047] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2224.047850] env[62824]: DEBUG oslo_vmware.api [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146606, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2224.076855] env[62824]: DEBUG oslo_vmware.api [None req-742366bb-e991-49d9-ad5f-cebf8cbed982 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': task-2146604, 'name': PowerOffVM_Task, 'duration_secs': 0.175202} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2224.077137] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-742366bb-e991-49d9-ad5f-cebf8cbed982 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7c44d956-53d1-4a5f-9016-c0f3f0f348b4] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2224.077311] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-742366bb-e991-49d9-ad5f-cebf8cbed982 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7c44d956-53d1-4a5f-9016-c0f3f0f348b4] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2224.077935] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cc819a3d-888d-40fb-8d5d-454460231ed0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2224.282579] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c7d29f0c-9cc1-46d5-94e7-68f8528e0f81 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 1f162a63-8773-4c34-9c84-b2de1d5e9f50] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2224.282817] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c7d29f0c-9cc1-46d5-94e7-68f8528e0f81 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 1f162a63-8773-4c34-9c84-b2de1d5e9f50] Deleting contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2224.282999] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7d29f0c-9cc1-46d5-94e7-68f8528e0f81 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Deleting the datastore file [datastore2] 1f162a63-8773-4c34-9c84-b2de1d5e9f50 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2224.283275] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-069714b1-f368-4ebd-91a9-4d71b173a09f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2224.289741] env[62824]: DEBUG oslo_vmware.api [None req-c7d29f0c-9cc1-46d5-94e7-68f8528e0f81 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Waiting for the task: (returnval){ [ 2224.289741] env[62824]: value = "task-2146608" [ 2224.289741] env[62824]: _type = "Task" [ 2224.289741] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2224.300132] env[62824]: DEBUG oslo_vmware.api [None req-c7d29f0c-9cc1-46d5-94e7-68f8528e0f81 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': task-2146608, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2224.452328] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-742366bb-e991-49d9-ad5f-cebf8cbed982 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7c44d956-53d1-4a5f-9016-c0f3f0f348b4] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2224.452328] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-742366bb-e991-49d9-ad5f-cebf8cbed982 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7c44d956-53d1-4a5f-9016-c0f3f0f348b4] Deleting contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2224.452328] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-742366bb-e991-49d9-ad5f-cebf8cbed982 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Deleting the datastore file [datastore1] 7c44d956-53d1-4a5f-9016-c0f3f0f348b4 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2224.452328] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f3ee1db5-8db3-4826-aa51-fa641c5429d0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2224.461022] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.356s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2224.461022] env[62824]: DEBUG nova.compute.manager [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2224.462703] env[62824]: DEBUG oslo_vmware.api [None req-742366bb-e991-49d9-ad5f-cebf8cbed982 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Waiting for the task: (returnval){ [ 2224.462703] env[62824]: value = "task-2146609" [ 2224.462703] env[62824]: _type = "Task" [ 2224.462703] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2224.463249] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.892s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2224.463495] env[62824]: DEBUG nova.objects.instance [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lazy-loading 'resources' on Instance uuid 58243465-ba24-4b37-a640-04ab24963230 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2224.479118] env[62824]: DEBUG oslo_vmware.api [None req-742366bb-e991-49d9-ad5f-cebf8cbed982 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': task-2146609, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2224.549266] env[62824]: DEBUG oslo_vmware.api [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146606, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2224.808148] env[62824]: DEBUG oslo_vmware.api [None req-c7d29f0c-9cc1-46d5-94e7-68f8528e0f81 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': task-2146608, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.135249} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2224.810366] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7d29f0c-9cc1-46d5-94e7-68f8528e0f81 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2224.810600] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c7d29f0c-9cc1-46d5-94e7-68f8528e0f81 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 1f162a63-8773-4c34-9c84-b2de1d5e9f50] Deleted contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2224.810790] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c7d29f0c-9cc1-46d5-94e7-68f8528e0f81 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 1f162a63-8773-4c34-9c84-b2de1d5e9f50] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2224.810970] env[62824]: INFO nova.compute.manager [None req-c7d29f0c-9cc1-46d5-94e7-68f8528e0f81 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 1f162a63-8773-4c34-9c84-b2de1d5e9f50] Took 1.34 seconds to destroy the instance on the hypervisor. [ 2224.811250] env[62824]: DEBUG oslo.service.loopingcall [None req-c7d29f0c-9cc1-46d5-94e7-68f8528e0f81 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2224.811462] env[62824]: DEBUG nova.compute.manager [-] [instance: 1f162a63-8773-4c34-9c84-b2de1d5e9f50] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2224.811562] env[62824]: DEBUG nova.network.neutron [-] [instance: 1f162a63-8773-4c34-9c84-b2de1d5e9f50] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2224.965860] env[62824]: DEBUG nova.compute.utils [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2224.967474] env[62824]: DEBUG nova.compute.manager [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2224.967651] env[62824]: DEBUG nova.network.neutron [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2224.975021] env[62824]: DEBUG nova.objects.instance [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lazy-loading 'numa_topology' on Instance uuid 58243465-ba24-4b37-a640-04ab24963230 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2224.981282] env[62824]: DEBUG oslo_vmware.api [None req-742366bb-e991-49d9-ad5f-cebf8cbed982 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Task: {'id': task-2146609, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.133902} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2224.981486] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-742366bb-e991-49d9-ad5f-cebf8cbed982 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2224.982822] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-742366bb-e991-49d9-ad5f-cebf8cbed982 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7c44d956-53d1-4a5f-9016-c0f3f0f348b4] Deleted contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2224.983097] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-742366bb-e991-49d9-ad5f-cebf8cbed982 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7c44d956-53d1-4a5f-9016-c0f3f0f348b4] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2224.983318] env[62824]: INFO nova.compute.manager [None req-742366bb-e991-49d9-ad5f-cebf8cbed982 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] [instance: 7c44d956-53d1-4a5f-9016-c0f3f0f348b4] Took 1.43 seconds to destroy the instance on the hypervisor. [ 2224.983585] env[62824]: DEBUG oslo.service.loopingcall [None req-742366bb-e991-49d9-ad5f-cebf8cbed982 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2224.983796] env[62824]: DEBUG nova.compute.manager [-] [instance: 7c44d956-53d1-4a5f-9016-c0f3f0f348b4] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2224.983896] env[62824]: DEBUG nova.network.neutron [-] [instance: 7c44d956-53d1-4a5f-9016-c0f3f0f348b4] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2225.053042] env[62824]: DEBUG oslo_vmware.api [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146606, 'name': PowerOnVM_Task, 'duration_secs': 0.620111} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2225.053405] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 853f3098-881c-402c-8fd7-26540d1ecd88] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2225.053613] env[62824]: INFO nova.compute.manager [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 853f3098-881c-402c-8fd7-26540d1ecd88] Took 10.90 seconds to spawn the instance on the hypervisor. [ 2225.053967] env[62824]: DEBUG nova.compute.manager [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 853f3098-881c-402c-8fd7-26540d1ecd88] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2225.056924] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59711c81-0e29-42cc-947c-d2e325ca849f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2225.072471] env[62824]: DEBUG nova.policy [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6287facc8293419c943c0234ed798dac', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd180bbdc8e72414081363fe696ead6fb', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2225.475298] env[62824]: DEBUG nova.compute.manager [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2225.479182] env[62824]: DEBUG nova.objects.base [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Object Instance<58243465-ba24-4b37-a640-04ab24963230> lazy-loaded attributes: resources,numa_topology {{(pid=62824) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 2225.552136] env[62824]: DEBUG nova.network.neutron [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a] Successfully created port: a6e03f6a-b693-48dc-a8a5-b853d424639f {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2225.583341] env[62824]: INFO nova.compute.manager [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 853f3098-881c-402c-8fd7-26540d1ecd88] Took 17.99 seconds to build instance. [ 2225.604630] env[62824]: DEBUG nova.compute.manager [req-aaec6115-5376-4ca6-96a7-9c98627d4701 req-5ef6f577-0204-42b6-99f6-6e3eaba0b6ce service nova] [instance: 1f162a63-8773-4c34-9c84-b2de1d5e9f50] Received event network-vif-deleted-f5c2ffd1-3953-40e4-8d96-0542389767da {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2225.604776] env[62824]: INFO nova.compute.manager [req-aaec6115-5376-4ca6-96a7-9c98627d4701 req-5ef6f577-0204-42b6-99f6-6e3eaba0b6ce service nova] [instance: 1f162a63-8773-4c34-9c84-b2de1d5e9f50] Neutron deleted interface f5c2ffd1-3953-40e4-8d96-0542389767da; detaching it from the instance and deleting it from the info cache [ 2225.604923] env[62824]: DEBUG nova.network.neutron [req-aaec6115-5376-4ca6-96a7-9c98627d4701 req-5ef6f577-0204-42b6-99f6-6e3eaba0b6ce service nova] [instance: 1f162a63-8773-4c34-9c84-b2de1d5e9f50] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2225.765343] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-169c8bfb-5020-48ff-b574-3680d895c5e8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2225.773669] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b163074e-10dc-470d-bbea-836bfe383d99 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2225.817505] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9b78f73-ed08-44d4-93c7-eca7841fa4c5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2225.825670] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23d447ed-24ff-43d9-9ecd-f07fa95f8da2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2225.843511] env[62824]: DEBUG nova.compute.provider_tree [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2225.873710] env[62824]: DEBUG nova.network.neutron [-] [instance: 1f162a63-8773-4c34-9c84-b2de1d5e9f50] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2225.902340] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquiring lock "0394d368-c846-4f3b-bfcf-2bc8e858052a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2225.904061] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "0394d368-c846-4f3b-bfcf-2bc8e858052a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2225.972894] env[62824]: DEBUG nova.network.neutron [-] [instance: 7c44d956-53d1-4a5f-9016-c0f3f0f348b4] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2226.086516] env[62824]: DEBUG oslo_concurrency.lockutils [None req-565694f4-7955-4bcd-b834-1dc21c431590 tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Lock "853f3098-881c-402c-8fd7-26540d1ecd88" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.505s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2226.114163] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c5a877e7-b869-4dfa-9d36-b125deb95016 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2226.122902] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c604238c-002f-4295-a1d1-1cf559f7cad9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2226.138326] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3f312930-e4fe-46f8-b10a-fcb8add8c2e9 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquiring lock "interface-fc77c44a-180c-46ff-9690-9072c6213c91-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2226.138586] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3f312930-e4fe-46f8-b10a-fcb8add8c2e9 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lock "interface-fc77c44a-180c-46ff-9690-9072c6213c91-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2226.138890] env[62824]: DEBUG nova.objects.instance [None req-3f312930-e4fe-46f8-b10a-fcb8add8c2e9 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lazy-loading 'flavor' on Instance uuid fc77c44a-180c-46ff-9690-9072c6213c91 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2226.158302] env[62824]: DEBUG nova.compute.manager [req-aaec6115-5376-4ca6-96a7-9c98627d4701 req-5ef6f577-0204-42b6-99f6-6e3eaba0b6ce service nova] [instance: 1f162a63-8773-4c34-9c84-b2de1d5e9f50] Detach interface failed, port_id=f5c2ffd1-3953-40e4-8d96-0542389767da, reason: Instance 1f162a63-8773-4c34-9c84-b2de1d5e9f50 could not be found. {{(pid=62824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11482}} [ 2226.205521] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fd709298-cce5-4857-be32-91ead529dfbb tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquiring lock "5b3df31f-15fe-473f-992c-ddb272661c53" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2226.205674] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fd709298-cce5-4857-be32-91ead529dfbb tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lock "5b3df31f-15fe-473f-992c-ddb272661c53" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2226.205878] env[62824]: DEBUG nova.compute.manager [None req-fd709298-cce5-4857-be32-91ead529dfbb tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2226.206832] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acea740e-83b9-43e2-b9ac-dfff8bdf0d6b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2226.213937] env[62824]: DEBUG nova.compute.manager [None req-fd709298-cce5-4857-be32-91ead529dfbb tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62824) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3404}} [ 2226.214236] env[62824]: DEBUG nova.objects.instance [None req-fd709298-cce5-4857-be32-91ead529dfbb tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lazy-loading 'flavor' on Instance uuid 5b3df31f-15fe-473f-992c-ddb272661c53 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2226.347032] env[62824]: DEBUG nova.scheduler.client.report [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2226.375703] env[62824]: INFO nova.compute.manager [-] [instance: 1f162a63-8773-4c34-9c84-b2de1d5e9f50] Took 1.56 seconds to deallocate network for instance. [ 2226.410923] env[62824]: DEBUG nova.compute.manager [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2226.477910] env[62824]: INFO nova.compute.manager [-] [instance: 7c44d956-53d1-4a5f-9016-c0f3f0f348b4] Took 1.49 seconds to deallocate network for instance. [ 2226.493210] env[62824]: DEBUG nova.compute.manager [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2226.528616] env[62824]: DEBUG nova.virt.hardware [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2226.528811] env[62824]: DEBUG nova.virt.hardware [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2226.528971] env[62824]: DEBUG nova.virt.hardware [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2226.529243] env[62824]: DEBUG nova.virt.hardware [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2226.529333] env[62824]: DEBUG nova.virt.hardware [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2226.531031] env[62824]: DEBUG nova.virt.hardware [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2226.531031] env[62824]: DEBUG nova.virt.hardware [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2226.531031] env[62824]: DEBUG nova.virt.hardware [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2226.531277] env[62824]: DEBUG nova.virt.hardware [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2226.531277] env[62824]: DEBUG nova.virt.hardware [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2226.531437] env[62824]: DEBUG nova.virt.hardware [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2226.532765] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d031c6f5-422d-425d-ad8e-457c54c101f1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2226.543421] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69007e08-91a2-4d08-950a-c6b1dde6d9d1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2226.644508] env[62824]: DEBUG nova.objects.instance [None req-3f312930-e4fe-46f8-b10a-fcb8add8c2e9 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lazy-loading 'pci_requests' on Instance uuid fc77c44a-180c-46ff-9690-9072c6213c91 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2226.852290] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.389s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2226.854807] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.193s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2226.856861] env[62824]: INFO nova.compute.claims [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: feda7def-7788-4a60-a56a-9353a1475c54] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2226.882754] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c7d29f0c-9cc1-46d5-94e7-68f8528e0f81 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2226.934422] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2226.984534] env[62824]: DEBUG oslo_concurrency.lockutils [None req-742366bb-e991-49d9-ad5f-cebf8cbed982 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2227.146128] env[62824]: DEBUG nova.objects.base [None req-3f312930-e4fe-46f8-b10a-fcb8add8c2e9 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=62824) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 2227.146380] env[62824]: DEBUG nova.network.neutron [None req-3f312930-e4fe-46f8-b10a-fcb8add8c2e9 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2227.223220] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd709298-cce5-4857-be32-91ead529dfbb tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2227.223577] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-45f02f2e-df15-4492-b2cb-decf36b6f638 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2227.235419] env[62824]: DEBUG oslo_vmware.api [None req-fd709298-cce5-4857-be32-91ead529dfbb tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2227.235419] env[62824]: value = "task-2146610" [ 2227.235419] env[62824]: _type = "Task" [ 2227.235419] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2227.242181] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3f312930-e4fe-46f8-b10a-fcb8add8c2e9 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lock "interface-fc77c44a-180c-46ff-9690-9072c6213c91-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.103s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2227.252814] env[62824]: DEBUG oslo_vmware.api [None req-fd709298-cce5-4857-be32-91ead529dfbb tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2146610, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2227.359774] env[62824]: DEBUG nova.compute.manager [req-f6959cd6-f5bc-47fb-872f-e9843c2a3eea req-9ce74dfd-e0e9-4cfe-acdb-6f0a5aaff271 service nova] [instance: b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a] Received event network-vif-plugged-a6e03f6a-b693-48dc-a8a5-b853d424639f {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2227.360075] env[62824]: DEBUG oslo_concurrency.lockutils [req-f6959cd6-f5bc-47fb-872f-e9843c2a3eea req-9ce74dfd-e0e9-4cfe-acdb-6f0a5aaff271 service nova] Acquiring lock "b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2227.360261] env[62824]: DEBUG oslo_concurrency.lockutils [req-f6959cd6-f5bc-47fb-872f-e9843c2a3eea req-9ce74dfd-e0e9-4cfe-acdb-6f0a5aaff271 service nova] Lock "b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2227.360856] env[62824]: DEBUG oslo_concurrency.lockutils [req-f6959cd6-f5bc-47fb-872f-e9843c2a3eea req-9ce74dfd-e0e9-4cfe-acdb-6f0a5aaff271 service nova] Lock "b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2227.361129] env[62824]: DEBUG nova.compute.manager [req-f6959cd6-f5bc-47fb-872f-e9843c2a3eea req-9ce74dfd-e0e9-4cfe-acdb-6f0a5aaff271 service nova] [instance: b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a] No waiting events found dispatching network-vif-plugged-a6e03f6a-b693-48dc-a8a5-b853d424639f {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2227.361386] env[62824]: WARNING nova.compute.manager [req-f6959cd6-f5bc-47fb-872f-e9843c2a3eea req-9ce74dfd-e0e9-4cfe-acdb-6f0a5aaff271 service nova] [instance: b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a] Received unexpected event network-vif-plugged-a6e03f6a-b693-48dc-a8a5-b853d424639f for instance with vm_state building and task_state spawning. [ 2227.377428] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2b0f0f00-78a5-4272-9ccc-d0e5f363adae tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "58243465-ba24-4b37-a640-04ab24963230" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 39.331s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2227.383593] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dd993d10-9788-4735-95a7-599a510439c1 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "58243465-ba24-4b37-a640-04ab24963230" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 10.262s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2227.383593] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dd993d10-9788-4735-95a7-599a510439c1 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquiring lock "58243465-ba24-4b37-a640-04ab24963230-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2227.383814] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dd993d10-9788-4735-95a7-599a510439c1 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "58243465-ba24-4b37-a640-04ab24963230-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2227.383984] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dd993d10-9788-4735-95a7-599a510439c1 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "58243465-ba24-4b37-a640-04ab24963230-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2227.386590] env[62824]: INFO nova.compute.manager [None req-dd993d10-9788-4735-95a7-599a510439c1 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 58243465-ba24-4b37-a640-04ab24963230] Terminating instance [ 2227.464448] env[62824]: DEBUG nova.compute.manager [None req-e2b575de-d6cc-43f9-9d33-cc6625760cab tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 853f3098-881c-402c-8fd7-26540d1ecd88] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2227.465398] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af0d4bbf-6738-4c69-86ae-5e1ee31eb907 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2227.588493] env[62824]: DEBUG nova.network.neutron [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a] Successfully updated port: a6e03f6a-b693-48dc-a8a5-b853d424639f {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2227.690633] env[62824]: DEBUG nova.compute.manager [req-167a636e-a6a8-4060-b8fb-df316471ccca req-599dff95-85cb-430f-86c0-2381adbbb59e service nova] [instance: 7c44d956-53d1-4a5f-9016-c0f3f0f348b4] Received event network-vif-deleted-8f686dbc-67bd-4f14-aec0-a6f9e7d8bfc5 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2227.745972] env[62824]: DEBUG oslo_vmware.api [None req-fd709298-cce5-4857-be32-91ead529dfbb tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2146610, 'name': PowerOffVM_Task, 'duration_secs': 0.201481} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2227.746253] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd709298-cce5-4857-be32-91ead529dfbb tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2227.746483] env[62824]: DEBUG nova.compute.manager [None req-fd709298-cce5-4857-be32-91ead529dfbb tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2227.747231] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba9a5525-8b97-43b4-b744-54b93e9b0c3f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2227.890299] env[62824]: DEBUG nova.compute.manager [None req-dd993d10-9788-4735-95a7-599a510439c1 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 58243465-ba24-4b37-a640-04ab24963230] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2227.890567] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-dd993d10-9788-4735-95a7-599a510439c1 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 58243465-ba24-4b37-a640-04ab24963230] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2227.891130] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8e2f8fac-6103-46ce-8fb2-a13205cad40a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2227.904519] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d994c8e-d958-4a9f-8660-8e6af3baab75 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2227.938216] env[62824]: WARNING nova.virt.vmwareapi.vmops [None req-dd993d10-9788-4735-95a7-599a510439c1 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 58243465-ba24-4b37-a640-04ab24963230] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 58243465-ba24-4b37-a640-04ab24963230 could not be found. [ 2227.938487] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-dd993d10-9788-4735-95a7-599a510439c1 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 58243465-ba24-4b37-a640-04ab24963230] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2227.938763] env[62824]: INFO nova.compute.manager [None req-dd993d10-9788-4735-95a7-599a510439c1 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 58243465-ba24-4b37-a640-04ab24963230] Took 0.05 seconds to destroy the instance on the hypervisor. [ 2227.939082] env[62824]: DEBUG oslo.service.loopingcall [None req-dd993d10-9788-4735-95a7-599a510439c1 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2227.941653] env[62824]: DEBUG nova.compute.manager [-] [instance: 58243465-ba24-4b37-a640-04ab24963230] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2227.942664] env[62824]: DEBUG nova.network.neutron [-] [instance: 58243465-ba24-4b37-a640-04ab24963230] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2227.978816] env[62824]: INFO nova.compute.manager [None req-e2b575de-d6cc-43f9-9d33-cc6625760cab tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 853f3098-881c-402c-8fd7-26540d1ecd88] instance snapshotting [ 2227.984495] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e74df584-3f6a-4a9c-ae84-915352d58018 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2228.005376] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1682b87d-9f0e-4ef8-828a-66d8589700a6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2228.093512] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Acquiring lock "refresh_cache-b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2228.093675] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Acquired lock "refresh_cache-b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2228.093832] env[62824]: DEBUG nova.network.neutron [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2228.149376] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f8e9a56-e2f9-4c7c-83af-cc57395b9083 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2228.157204] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97e9997f-c146-4035-a8f9-fff54983698f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2228.193247] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-255f2fa4-e70c-4b21-81b4-e8322b0d25a1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2228.201675] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71cf540a-db99-4c7d-9929-28d414e070f9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2228.216786] env[62824]: DEBUG nova.compute.provider_tree [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2228.259525] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fd709298-cce5-4857-be32-91ead529dfbb tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lock "5b3df31f-15fe-473f-992c-ddb272661c53" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.053s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2228.517190] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e2b575de-d6cc-43f9-9d33-cc6625760cab tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 853f3098-881c-402c-8fd7-26540d1ecd88] Creating Snapshot of the VM instance {{(pid=62824) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 2228.517877] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-33c5932c-ea30-48b5-b724-7db90d5768ba {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2228.525478] env[62824]: DEBUG oslo_vmware.api [None req-e2b575de-d6cc-43f9-9d33-cc6625760cab tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the task: (returnval){ [ 2228.525478] env[62824]: value = "task-2146611" [ 2228.525478] env[62824]: _type = "Task" [ 2228.525478] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2228.532963] env[62824]: DEBUG oslo_vmware.api [None req-e2b575de-d6cc-43f9-9d33-cc6625760cab tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146611, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2228.627330] env[62824]: DEBUG nova.network.neutron [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2228.719215] env[62824]: DEBUG nova.scheduler.client.report [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2228.723339] env[62824]: DEBUG nova.network.neutron [-] [instance: 58243465-ba24-4b37-a640-04ab24963230] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2228.820980] env[62824]: DEBUG nova.network.neutron [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a] Updating instance_info_cache with network_info: [{"id": "a6e03f6a-b693-48dc-a8a5-b853d424639f", "address": "fa:16:3e:4b:ff:d7", "network": {"id": "d299eecf-1628-4fce-bc67-f9f038d261cd", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1701982809-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d180bbdc8e72414081363fe696ead6fb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae4e3171-21cd-4094-b6cf-81bf366c75bd", "external-id": "nsx-vlan-transportzone-193", "segmentation_id": 193, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa6e03f6a-b6", "ovs_interfaceid": "a6e03f6a-b693-48dc-a8a5-b853d424639f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2228.906999] env[62824]: DEBUG nova.objects.instance [None req-c7558da8-21bc-422c-b39e-49171f99acf8 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lazy-loading 'flavor' on Instance uuid 5b3df31f-15fe-473f-992c-ddb272661c53 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2229.035737] env[62824]: DEBUG oslo_vmware.api [None req-e2b575de-d6cc-43f9-9d33-cc6625760cab tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146611, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2229.182583] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c948012c-66d5-416f-91d2-75a66106d0c6 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquiring lock "interface-fc77c44a-180c-46ff-9690-9072c6213c91-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2229.182865] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c948012c-66d5-416f-91d2-75a66106d0c6 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lock "interface-fc77c44a-180c-46ff-9690-9072c6213c91-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2229.183225] env[62824]: DEBUG nova.objects.instance [None req-c948012c-66d5-416f-91d2-75a66106d0c6 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lazy-loading 'flavor' on Instance uuid fc77c44a-180c-46ff-9690-9072c6213c91 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2229.225046] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.370s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2229.225519] env[62824]: DEBUG nova.compute.manager [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: feda7def-7788-4a60-a56a-9353a1475c54] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2229.228718] env[62824]: DEBUG oslo_concurrency.lockutils [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.501s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2229.230902] env[62824]: INFO nova.compute.claims [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2229.233548] env[62824]: INFO nova.compute.manager [-] [instance: 58243465-ba24-4b37-a640-04ab24963230] Took 1.29 seconds to deallocate network for instance. [ 2229.323662] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Releasing lock "refresh_cache-b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2229.323986] env[62824]: DEBUG nova.compute.manager [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a] Instance network_info: |[{"id": "a6e03f6a-b693-48dc-a8a5-b853d424639f", "address": "fa:16:3e:4b:ff:d7", "network": {"id": "d299eecf-1628-4fce-bc67-f9f038d261cd", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1701982809-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d180bbdc8e72414081363fe696ead6fb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae4e3171-21cd-4094-b6cf-81bf366c75bd", "external-id": "nsx-vlan-transportzone-193", "segmentation_id": 193, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa6e03f6a-b6", "ovs_interfaceid": "a6e03f6a-b693-48dc-a8a5-b853d424639f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2229.324425] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4b:ff:d7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ae4e3171-21cd-4094-b6cf-81bf366c75bd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a6e03f6a-b693-48dc-a8a5-b853d424639f', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2229.332537] env[62824]: DEBUG oslo.service.loopingcall [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2229.332748] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2229.332984] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-73176fef-3cce-44ef-91fe-9f9b4320d8ac {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2229.352643] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2229.352643] env[62824]: value = "task-2146612" [ 2229.352643] env[62824]: _type = "Task" [ 2229.352643] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2229.359794] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146612, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2229.412008] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c7558da8-21bc-422c-b39e-49171f99acf8 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquiring lock "refresh_cache-5b3df31f-15fe-473f-992c-ddb272661c53" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2229.412166] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c7558da8-21bc-422c-b39e-49171f99acf8 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquired lock "refresh_cache-5b3df31f-15fe-473f-992c-ddb272661c53" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2229.412300] env[62824]: DEBUG nova.network.neutron [None req-c7558da8-21bc-422c-b39e-49171f99acf8 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2229.412528] env[62824]: DEBUG nova.objects.instance [None req-c7558da8-21bc-422c-b39e-49171f99acf8 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lazy-loading 'info_cache' on Instance uuid 5b3df31f-15fe-473f-992c-ddb272661c53 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2229.536028] env[62824]: DEBUG oslo_vmware.api [None req-e2b575de-d6cc-43f9-9d33-cc6625760cab tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146611, 'name': CreateSnapshot_Task, 'duration_secs': 0.687391} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2229.536332] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e2b575de-d6cc-43f9-9d33-cc6625760cab tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 853f3098-881c-402c-8fd7-26540d1ecd88] Created Snapshot of the VM instance {{(pid=62824) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 2229.537618] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11321f4c-46a0-4aac-9730-af99d9181a4c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2229.650838] env[62824]: DEBUG nova.compute.manager [req-5b12ae6d-9023-4f12-b127-deac0a021edc req-f7904f9e-1ccf-460e-a944-67afa3aacbe0 service nova] [instance: b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a] Received event network-changed-a6e03f6a-b693-48dc-a8a5-b853d424639f {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2229.651122] env[62824]: DEBUG nova.compute.manager [req-5b12ae6d-9023-4f12-b127-deac0a021edc req-f7904f9e-1ccf-460e-a944-67afa3aacbe0 service nova] [instance: b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a] Refreshing instance network info cache due to event network-changed-a6e03f6a-b693-48dc-a8a5-b853d424639f. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2229.651286] env[62824]: DEBUG oslo_concurrency.lockutils [req-5b12ae6d-9023-4f12-b127-deac0a021edc req-f7904f9e-1ccf-460e-a944-67afa3aacbe0 service nova] Acquiring lock "refresh_cache-b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2229.651430] env[62824]: DEBUG oslo_concurrency.lockutils [req-5b12ae6d-9023-4f12-b127-deac0a021edc req-f7904f9e-1ccf-460e-a944-67afa3aacbe0 service nova] Acquired lock "refresh_cache-b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2229.651591] env[62824]: DEBUG nova.network.neutron [req-5b12ae6d-9023-4f12-b127-deac0a021edc req-f7904f9e-1ccf-460e-a944-67afa3aacbe0 service nova] [instance: b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a] Refreshing network info cache for port a6e03f6a-b693-48dc-a8a5-b853d424639f {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2229.736785] env[62824]: DEBUG nova.compute.utils [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2229.741428] env[62824]: DEBUG nova.compute.manager [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: feda7def-7788-4a60-a56a-9353a1475c54] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2229.741596] env[62824]: DEBUG nova.network.neutron [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: feda7def-7788-4a60-a56a-9353a1475c54] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2229.757138] env[62824]: DEBUG nova.objects.instance [None req-c948012c-66d5-416f-91d2-75a66106d0c6 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lazy-loading 'pci_requests' on Instance uuid fc77c44a-180c-46ff-9690-9072c6213c91 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2229.785229] env[62824]: DEBUG nova.policy [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '959d127a7d144b33a0cae94db5c11846', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dfff08982dad4790bf4d555e2b4db5e4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2229.862703] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146612, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2229.915633] env[62824]: DEBUG nova.objects.base [None req-c7558da8-21bc-422c-b39e-49171f99acf8 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Object Instance<5b3df31f-15fe-473f-992c-ddb272661c53> lazy-loaded attributes: flavor,info_cache {{(pid=62824) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 2230.035390] env[62824]: DEBUG nova.network.neutron [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: feda7def-7788-4a60-a56a-9353a1475c54] Successfully created port: b133d2a7-e5fb-44fd-b716-20381faecc3d {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2230.058020] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e2b575de-d6cc-43f9-9d33-cc6625760cab tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 853f3098-881c-402c-8fd7-26540d1ecd88] Creating linked-clone VM from snapshot {{(pid=62824) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 2230.058020] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-95708a2c-067c-499c-9b3b-c63dce57124c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2230.066552] env[62824]: DEBUG oslo_vmware.api [None req-e2b575de-d6cc-43f9-9d33-cc6625760cab tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the task: (returnval){ [ 2230.066552] env[62824]: value = "task-2146613" [ 2230.066552] env[62824]: _type = "Task" [ 2230.066552] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2230.075051] env[62824]: DEBUG oslo_vmware.api [None req-e2b575de-d6cc-43f9-9d33-cc6625760cab tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146613, 'name': CloneVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2230.245139] env[62824]: DEBUG nova.compute.manager [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: feda7def-7788-4a60-a56a-9353a1475c54] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2230.263446] env[62824]: DEBUG nova.objects.base [None req-c948012c-66d5-416f-91d2-75a66106d0c6 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=62824) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 2230.263446] env[62824]: DEBUG nova.network.neutron [None req-c948012c-66d5-416f-91d2-75a66106d0c6 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2230.270232] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dd993d10-9788-4735-95a7-599a510439c1 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "58243465-ba24-4b37-a640-04ab24963230" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 2.887s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2230.332608] env[62824]: DEBUG nova.policy [None req-c948012c-66d5-416f-91d2-75a66106d0c6 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b989eb89498b45dab6604116870d2fa1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c987c5f844be42c99cc9d57a9ca84c31', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2230.364750] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146612, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2230.489713] env[62824]: DEBUG nova.network.neutron [req-5b12ae6d-9023-4f12-b127-deac0a021edc req-f7904f9e-1ccf-460e-a944-67afa3aacbe0 service nova] [instance: b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a] Updated VIF entry in instance network info cache for port a6e03f6a-b693-48dc-a8a5-b853d424639f. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2230.490098] env[62824]: DEBUG nova.network.neutron [req-5b12ae6d-9023-4f12-b127-deac0a021edc req-f7904f9e-1ccf-460e-a944-67afa3aacbe0 service nova] [instance: b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a] Updating instance_info_cache with network_info: [{"id": "a6e03f6a-b693-48dc-a8a5-b853d424639f", "address": "fa:16:3e:4b:ff:d7", "network": {"id": "d299eecf-1628-4fce-bc67-f9f038d261cd", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1701982809-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d180bbdc8e72414081363fe696ead6fb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae4e3171-21cd-4094-b6cf-81bf366c75bd", "external-id": "nsx-vlan-transportzone-193", "segmentation_id": 193, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa6e03f6a-b6", "ovs_interfaceid": "a6e03f6a-b693-48dc-a8a5-b853d424639f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2230.546643] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93ac3fee-43cd-491f-ac53-1ca7bbc59a03 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2230.554535] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b674cd5-552a-4515-920a-c4285ea9a34b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2230.590590] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-769cfcfa-77e4-4a2b-97d3-ad1ecee92db9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2230.602549] env[62824]: DEBUG oslo_vmware.api [None req-e2b575de-d6cc-43f9-9d33-cc6625760cab tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146613, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2230.603755] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60949512-5653-46ea-a894-b902a0fd3d53 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2230.616785] env[62824]: DEBUG nova.compute.provider_tree [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2230.720162] env[62824]: DEBUG nova.network.neutron [None req-c948012c-66d5-416f-91d2-75a66106d0c6 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Successfully created port: 2b6381b4-f3c7-4e22-b698-de190c8199ee {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2230.796535] env[62824]: DEBUG nova.network.neutron [None req-c7558da8-21bc-422c-b39e-49171f99acf8 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Updating instance_info_cache with network_info: [{"id": "7be95c5e-6e98-4fec-8550-1c22a787b95b", "address": "fa:16:3e:e8:69:3a", "network": {"id": "bba9b98c-c7eb-4890-8de5-eafacfdebd7d", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1928394728-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.160", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7fcdbfc29df64a1ba8d982bdcc667b64", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b49c5024-2ced-42ca-90cc-6066766d43e6", "external-id": "nsx-vlan-transportzone-239", "segmentation_id": 239, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7be95c5e-6e", "ovs_interfaceid": "7be95c5e-6e98-4fec-8550-1c22a787b95b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2230.863356] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146612, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2230.994569] env[62824]: DEBUG oslo_concurrency.lockutils [req-5b12ae6d-9023-4f12-b127-deac0a021edc req-f7904f9e-1ccf-460e-a944-67afa3aacbe0 service nova] Releasing lock "refresh_cache-b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2231.095974] env[62824]: DEBUG oslo_vmware.api [None req-e2b575de-d6cc-43f9-9d33-cc6625760cab tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146613, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2231.122158] env[62824]: DEBUG nova.scheduler.client.report [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2231.259159] env[62824]: DEBUG nova.compute.manager [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: feda7def-7788-4a60-a56a-9353a1475c54] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2231.283544] env[62824]: DEBUG nova.virt.hardware [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2231.283793] env[62824]: DEBUG nova.virt.hardware [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2231.283961] env[62824]: DEBUG nova.virt.hardware [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2231.284166] env[62824]: DEBUG nova.virt.hardware [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2231.284314] env[62824]: DEBUG nova.virt.hardware [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2231.284463] env[62824]: DEBUG nova.virt.hardware [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2231.284672] env[62824]: DEBUG nova.virt.hardware [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2231.284835] env[62824]: DEBUG nova.virt.hardware [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2231.285009] env[62824]: DEBUG nova.virt.hardware [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2231.285186] env[62824]: DEBUG nova.virt.hardware [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2231.285363] env[62824]: DEBUG nova.virt.hardware [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2231.286240] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab7f398b-0813-4b7c-8327-a87f4cb606b1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2231.294205] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97d7e072-5c15-4eb0-abdf-13b8c24dc360 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2231.298376] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c7558da8-21bc-422c-b39e-49171f99acf8 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Releasing lock "refresh_cache-5b3df31f-15fe-473f-992c-ddb272661c53" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2231.363730] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146612, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2231.473250] env[62824]: DEBUG oslo_concurrency.lockutils [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquiring lock "204b866f-e9c5-441d-bf8e-77d324b39ecd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2231.473491] env[62824]: DEBUG oslo_concurrency.lockutils [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "204b866f-e9c5-441d-bf8e-77d324b39ecd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2231.596425] env[62824]: DEBUG oslo_vmware.api [None req-e2b575de-d6cc-43f9-9d33-cc6625760cab tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146613, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2231.627011] env[62824]: DEBUG oslo_concurrency.lockutils [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.398s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2231.627521] env[62824]: DEBUG nova.compute.manager [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2231.630082] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c7d29f0c-9cc1-46d5-94e7-68f8528e0f81 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.747s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2231.630307] env[62824]: DEBUG nova.objects.instance [None req-c7d29f0c-9cc1-46d5-94e7-68f8528e0f81 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Lazy-loading 'resources' on Instance uuid 1f162a63-8773-4c34-9c84-b2de1d5e9f50 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2231.864475] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146612, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2231.976130] env[62824]: DEBUG nova.compute.manager [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 204b866f-e9c5-441d-bf8e-77d324b39ecd] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2232.097321] env[62824]: DEBUG oslo_vmware.api [None req-e2b575de-d6cc-43f9-9d33-cc6625760cab tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146613, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2232.132824] env[62824]: DEBUG nova.compute.utils [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2232.137045] env[62824]: DEBUG nova.compute.manager [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2232.137241] env[62824]: DEBUG nova.network.neutron [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2232.186915] env[62824]: DEBUG nova.policy [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3492c645a51548408a639e3af73c213c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '35b7289f4746463e9b1cc47fe914a835', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2232.304718] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7558da8-21bc-422c-b39e-49171f99acf8 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2232.305051] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-72cc85e1-3d9d-4206-bd74-5b56c542613c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2232.314247] env[62824]: DEBUG oslo_vmware.api [None req-c7558da8-21bc-422c-b39e-49171f99acf8 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2232.314247] env[62824]: value = "task-2146614" [ 2232.314247] env[62824]: _type = "Task" [ 2232.314247] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2232.329078] env[62824]: DEBUG oslo_vmware.api [None req-c7558da8-21bc-422c-b39e-49171f99acf8 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2146614, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2232.365587] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146612, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2232.378768] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08220583-2b58-42ad-a6a6-68d97dc42b87 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2232.385944] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8416ce4-761b-44d4-9646-0ca7eebf6f2c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2232.418223] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc251f6e-582c-4aef-b02e-880fcd5d92c1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2232.425924] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67376292-fa2c-4419-8f87-4a20e5880b0b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2232.440344] env[62824]: DEBUG nova.compute.provider_tree [None req-c7d29f0c-9cc1-46d5-94e7-68f8528e0f81 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2232.452766] env[62824]: DEBUG nova.network.neutron [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Successfully created port: 4214bf6e-29f0-4d92-be44-44fccf7ab149 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2232.495183] env[62824]: DEBUG oslo_concurrency.lockutils [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2232.600086] env[62824]: DEBUG oslo_vmware.api [None req-e2b575de-d6cc-43f9-9d33-cc6625760cab tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146613, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2232.638160] env[62824]: DEBUG nova.compute.manager [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2232.824392] env[62824]: DEBUG oslo_vmware.api [None req-c7558da8-21bc-422c-b39e-49171f99acf8 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2146614, 'name': PowerOnVM_Task, 'duration_secs': 0.363678} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2232.824656] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7558da8-21bc-422c-b39e-49171f99acf8 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2232.824865] env[62824]: DEBUG nova.compute.manager [None req-c7558da8-21bc-422c-b39e-49171f99acf8 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2232.825662] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6af7de22-530c-41dd-a9b7-9ca370feb4a7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2232.866452] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146612, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2232.943659] env[62824]: DEBUG nova.scheduler.client.report [None req-c7d29f0c-9cc1-46d5-94e7-68f8528e0f81 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2233.099519] env[62824]: DEBUG oslo_vmware.api [None req-e2b575de-d6cc-43f9-9d33-cc6625760cab tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146613, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2233.367858] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146612, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2233.448572] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c7d29f0c-9cc1-46d5-94e7-68f8528e0f81 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.818s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2233.450929] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.517s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2233.452391] env[62824]: INFO nova.compute.claims [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2233.471587] env[62824]: INFO nova.scheduler.client.report [None req-c7d29f0c-9cc1-46d5-94e7-68f8528e0f81 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Deleted allocations for instance 1f162a63-8773-4c34-9c84-b2de1d5e9f50 [ 2233.600309] env[62824]: DEBUG oslo_vmware.api [None req-e2b575de-d6cc-43f9-9d33-cc6625760cab tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146613, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2233.650358] env[62824]: DEBUG nova.compute.manager [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2233.675238] env[62824]: DEBUG nova.virt.hardware [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2233.675496] env[62824]: DEBUG nova.virt.hardware [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2233.675661] env[62824]: DEBUG nova.virt.hardware [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2233.675885] env[62824]: DEBUG nova.virt.hardware [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2233.676069] env[62824]: DEBUG nova.virt.hardware [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2233.676230] env[62824]: DEBUG nova.virt.hardware [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2233.676440] env[62824]: DEBUG nova.virt.hardware [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2233.676607] env[62824]: DEBUG nova.virt.hardware [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2233.676771] env[62824]: DEBUG nova.virt.hardware [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2233.676936] env[62824]: DEBUG nova.virt.hardware [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2233.677131] env[62824]: DEBUG nova.virt.hardware [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2233.677999] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-691fe5de-7678-4564-ae2b-e0a34d93e037 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2233.686163] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d02b6768-c44d-4dd0-8046-827e932b21d2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2233.869549] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146612, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2233.982033] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c7d29f0c-9cc1-46d5-94e7-68f8528e0f81 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Lock "1f162a63-8773-4c34-9c84-b2de1d5e9f50" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.021s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2234.115290] env[62824]: DEBUG oslo_vmware.api [None req-e2b575de-d6cc-43f9-9d33-cc6625760cab tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146613, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2234.370820] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146612, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2234.602108] env[62824]: DEBUG oslo_vmware.api [None req-e2b575de-d6cc-43f9-9d33-cc6625760cab tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146613, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2234.683959] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cdb3dd3-8fcb-4679-bbef-77f002abd485 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2234.691281] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9fe9ba7-7c26-4b06-a761-be17b28df333 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2234.721435] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f869ec2b-0e5f-4fab-9bc7-1e2c14da3b1c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2234.728950] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48d8ce7b-acdd-425d-919b-8b1d3621dd38 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2234.741953] env[62824]: DEBUG nova.compute.provider_tree [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2234.871614] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146612, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2235.103830] env[62824]: DEBUG oslo_vmware.api [None req-e2b575de-d6cc-43f9-9d33-cc6625760cab tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146613, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2235.245600] env[62824]: DEBUG nova.scheduler.client.report [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2235.371788] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146612, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2235.604425] env[62824]: DEBUG oslo_vmware.api [None req-e2b575de-d6cc-43f9-9d33-cc6625760cab tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146613, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2235.752768] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.302s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2235.753340] env[62824]: DEBUG nova.compute.manager [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2235.756011] env[62824]: DEBUG oslo_concurrency.lockutils [None req-742366bb-e991-49d9-ad5f-cebf8cbed982 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.772s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2235.756247] env[62824]: DEBUG nova.objects.instance [None req-742366bb-e991-49d9-ad5f-cebf8cbed982 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Lazy-loading 'resources' on Instance uuid 7c44d956-53d1-4a5f-9016-c0f3f0f348b4 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2235.872795] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146612, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2236.104333] env[62824]: DEBUG oslo_vmware.api [None req-e2b575de-d6cc-43f9-9d33-cc6625760cab tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146613, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2236.260246] env[62824]: DEBUG nova.compute.utils [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2236.265306] env[62824]: DEBUG nova.compute.manager [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2236.265694] env[62824]: DEBUG nova.network.neutron [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2236.303302] env[62824]: DEBUG nova.policy [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9cc01135021d4a90a718dc605ce98af2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '815ecf9340934733a83d4dcb26612fd4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2236.376109] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146612, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2236.454555] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5869f2d0-4f3c-4258-9043-9b14630fcc49 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2236.462030] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99492cf5-5f94-42cd-aa6b-b34af9a6d5af {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2236.492279] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8baae4a-6b83-40c5-9a14-913f015717dc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2236.499477] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c03b645-67c0-4e95-ab2b-38bdb098ec85 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2236.512397] env[62824]: DEBUG nova.compute.provider_tree [None req-742366bb-e991-49d9-ad5f-cebf8cbed982 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2236.594798] env[62824]: DEBUG nova.network.neutron [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Successfully created port: 2c7a5b60-534e-457c-a27b-29a489877783 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2236.605447] env[62824]: DEBUG oslo_vmware.api [None req-e2b575de-d6cc-43f9-9d33-cc6625760cab tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146613, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2236.765938] env[62824]: DEBUG nova.compute.manager [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2236.873612] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146612, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2237.015488] env[62824]: DEBUG nova.scheduler.client.report [None req-742366bb-e991-49d9-ad5f-cebf8cbed982 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2237.106627] env[62824]: DEBUG oslo_vmware.api [None req-e2b575de-d6cc-43f9-9d33-cc6625760cab tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146613, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2237.376232] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146612, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2237.520274] env[62824]: DEBUG oslo_concurrency.lockutils [None req-742366bb-e991-49d9-ad5f-cebf8cbed982 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.764s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2237.525012] env[62824]: DEBUG oslo_concurrency.lockutils [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.028s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2237.525012] env[62824]: INFO nova.compute.claims [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 204b866f-e9c5-441d-bf8e-77d324b39ecd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2237.543028] env[62824]: INFO nova.scheduler.client.report [None req-742366bb-e991-49d9-ad5f-cebf8cbed982 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Deleted allocations for instance 7c44d956-53d1-4a5f-9016-c0f3f0f348b4 [ 2237.607549] env[62824]: DEBUG oslo_vmware.api [None req-e2b575de-d6cc-43f9-9d33-cc6625760cab tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146613, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2237.775216] env[62824]: DEBUG nova.compute.manager [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2237.800127] env[62824]: DEBUG nova.virt.hardware [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2237.800380] env[62824]: DEBUG nova.virt.hardware [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2237.800572] env[62824]: DEBUG nova.virt.hardware [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2237.800761] env[62824]: DEBUG nova.virt.hardware [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2237.800914] env[62824]: DEBUG nova.virt.hardware [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2237.801089] env[62824]: DEBUG nova.virt.hardware [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2237.801302] env[62824]: DEBUG nova.virt.hardware [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2237.801463] env[62824]: DEBUG nova.virt.hardware [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2237.801630] env[62824]: DEBUG nova.virt.hardware [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2237.801792] env[62824]: DEBUG nova.virt.hardware [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2237.801963] env[62824]: DEBUG nova.virt.hardware [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2237.802832] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf6b628d-49ea-4b95-b78f-4e945270bccb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2237.810993] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d1ff3dd-7b23-4571-94de-9e3a8b620a11 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2237.874588] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146612, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2238.050747] env[62824]: DEBUG oslo_concurrency.lockutils [None req-742366bb-e991-49d9-ad5f-cebf8cbed982 tempest-MultipleCreateTestJSON-1796864087 tempest-MultipleCreateTestJSON-1796864087-project-member] Lock "7c44d956-53d1-4a5f-9016-c0f3f0f348b4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.007s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2238.108193] env[62824]: DEBUG oslo_vmware.api [None req-e2b575de-d6cc-43f9-9d33-cc6625760cab tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146613, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2238.375834] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146612, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2238.390576] env[62824]: DEBUG oslo_concurrency.lockutils [None req-4304e8a4-2b17-4cea-b613-125b1018f207 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Acquiring lock "bbfcb2e3-9326-4548-b15b-e054cbfd192e" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2238.390890] env[62824]: DEBUG oslo_concurrency.lockutils [None req-4304e8a4-2b17-4cea-b613-125b1018f207 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Lock "bbfcb2e3-9326-4548-b15b-e054cbfd192e" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2238.609737] env[62824]: DEBUG oslo_vmware.api [None req-e2b575de-d6cc-43f9-9d33-cc6625760cab tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146613, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2238.770405] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c445f8a8-8656-4d34-897e-c6fd2b2912b1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2238.777716] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a80a4d38-4232-4f94-ac8a-0b2960ebf5bd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2238.809058] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-072c4d33-b865-4dca-ad45-4b29a6e45f53 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2238.817465] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87df1eec-1531-4389-ae01-a1be67475b7b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2238.831293] env[62824]: DEBUG nova.compute.provider_tree [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2238.876905] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146612, 'name': CreateVM_Task, 'duration_secs': 9.328681} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2238.877251] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2238.877789] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2238.877961] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2238.878339] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2238.878622] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f06d1b74-7690-4ccf-8a1e-4302b65b5739 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2238.883731] env[62824]: DEBUG oslo_vmware.api [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Waiting for the task: (returnval){ [ 2238.883731] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52f76425-bc71-4a7b-36aa-7724530a7916" [ 2238.883731] env[62824]: _type = "Task" [ 2238.883731] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2238.891939] env[62824]: DEBUG oslo_vmware.api [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52f76425-bc71-4a7b-36aa-7724530a7916, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2238.893918] env[62824]: DEBUG nova.compute.utils [None req-4304e8a4-2b17-4cea-b613-125b1018f207 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2239.110875] env[62824]: DEBUG oslo_vmware.api [None req-e2b575de-d6cc-43f9-9d33-cc6625760cab tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146613, 'name': CloneVM_Task} progress is 95%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2239.334218] env[62824]: DEBUG nova.scheduler.client.report [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2239.398695] env[62824]: DEBUG oslo_concurrency.lockutils [None req-4304e8a4-2b17-4cea-b613-125b1018f207 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Lock "bbfcb2e3-9326-4548-b15b-e054cbfd192e" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.008s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2239.399457] env[62824]: DEBUG oslo_vmware.api [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52f76425-bc71-4a7b-36aa-7724530a7916, 'name': SearchDatastore_Task, 'duration_secs': 0.00935} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2239.399875] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2239.400191] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2239.400489] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2239.400672] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2239.401040] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2239.402284] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b8094d4c-3510-4c3f-8cfa-7c286aa3b8f8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2239.416934] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2239.417198] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2239.417879] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ba3c2683-f342-4e92-a31e-a1e29fe4ce3d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2239.423579] env[62824]: DEBUG oslo_vmware.api [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Waiting for the task: (returnval){ [ 2239.423579] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]526e690c-40b1-70cd-84be-652893b04d3a" [ 2239.423579] env[62824]: _type = "Task" [ 2239.423579] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2239.431385] env[62824]: DEBUG oslo_vmware.api [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]526e690c-40b1-70cd-84be-652893b04d3a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2239.620854] env[62824]: DEBUG oslo_vmware.api [None req-e2b575de-d6cc-43f9-9d33-cc6625760cab tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146613, 'name': CloneVM_Task, 'duration_secs': 9.408123} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2239.621658] env[62824]: INFO nova.virt.vmwareapi.vmops [None req-e2b575de-d6cc-43f9-9d33-cc6625760cab tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 853f3098-881c-402c-8fd7-26540d1ecd88] Created linked-clone VM from snapshot [ 2239.622490] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab862cf0-408e-4b37-8091-82488bc2d7c6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2239.631605] env[62824]: DEBUG nova.virt.vmwareapi.images [None req-e2b575de-d6cc-43f9-9d33-cc6625760cab tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 853f3098-881c-402c-8fd7-26540d1ecd88] Uploading image cc58daa7-49cf-44db-bb28-abad81e298fe {{(pid=62824) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 2239.643783] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2b575de-d6cc-43f9-9d33-cc6625760cab tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 853f3098-881c-402c-8fd7-26540d1ecd88] Destroying the VM {{(pid=62824) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 2239.644063] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-c378271d-8cb3-470a-b5eb-a6ea1747a546 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2239.653865] env[62824]: DEBUG oslo_vmware.api [None req-e2b575de-d6cc-43f9-9d33-cc6625760cab tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the task: (returnval){ [ 2239.653865] env[62824]: value = "task-2146615" [ 2239.653865] env[62824]: _type = "Task" [ 2239.653865] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2239.669213] env[62824]: DEBUG oslo_vmware.api [None req-e2b575de-d6cc-43f9-9d33-cc6625760cab tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146615, 'name': Destroy_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2239.839209] env[62824]: DEBUG oslo_concurrency.lockutils [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.316s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2239.839739] env[62824]: DEBUG nova.compute.manager [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 204b866f-e9c5-441d-bf8e-77d324b39ecd] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2239.937078] env[62824]: DEBUG oslo_vmware.api [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]526e690c-40b1-70cd-84be-652893b04d3a, 'name': SearchDatastore_Task, 'duration_secs': 0.008562} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2239.938319] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-00ae075e-df73-4a57-871e-77b50017bc3a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2239.944609] env[62824]: DEBUG oslo_vmware.api [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Waiting for the task: (returnval){ [ 2239.944609] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]521ebc8a-6261-b743-9246-440ec03d4ed6" [ 2239.944609] env[62824]: _type = "Task" [ 2239.944609] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2239.954360] env[62824]: DEBUG oslo_vmware.api [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]521ebc8a-6261-b743-9246-440ec03d4ed6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2240.086569] env[62824]: DEBUG nova.network.neutron [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: feda7def-7788-4a60-a56a-9353a1475c54] Successfully updated port: b133d2a7-e5fb-44fd-b716-20381faecc3d {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2240.107813] env[62824]: DEBUG nova.network.neutron [None req-c948012c-66d5-416f-91d2-75a66106d0c6 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Successfully updated port: 2b6381b4-f3c7-4e22-b698-de190c8199ee {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2240.166241] env[62824]: DEBUG oslo_vmware.api [None req-e2b575de-d6cc-43f9-9d33-cc6625760cab tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146615, 'name': Destroy_Task, 'duration_secs': 0.510632} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2240.166714] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-e2b575de-d6cc-43f9-9d33-cc6625760cab tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 853f3098-881c-402c-8fd7-26540d1ecd88] Destroyed the VM [ 2240.167394] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e2b575de-d6cc-43f9-9d33-cc6625760cab tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 853f3098-881c-402c-8fd7-26540d1ecd88] Deleting Snapshot of the VM instance {{(pid=62824) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 2240.167734] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-3b3bad99-caa1-4d69-a1ab-98f8aa26d746 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2240.175549] env[62824]: DEBUG oslo_vmware.api [None req-e2b575de-d6cc-43f9-9d33-cc6625760cab tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the task: (returnval){ [ 2240.175549] env[62824]: value = "task-2146616" [ 2240.175549] env[62824]: _type = "Task" [ 2240.175549] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2240.184462] env[62824]: DEBUG oslo_vmware.api [None req-e2b575de-d6cc-43f9-9d33-cc6625760cab tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146616, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2240.274409] env[62824]: DEBUG nova.network.neutron [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Successfully updated port: 4214bf6e-29f0-4d92-be44-44fccf7ab149 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2240.293686] env[62824]: DEBUG nova.network.neutron [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Successfully updated port: 2c7a5b60-534e-457c-a27b-29a489877783 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2240.345724] env[62824]: DEBUG nova.compute.utils [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2240.348848] env[62824]: DEBUG nova.compute.manager [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 204b866f-e9c5-441d-bf8e-77d324b39ecd] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2240.348848] env[62824]: DEBUG nova.network.neutron [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 204b866f-e9c5-441d-bf8e-77d324b39ecd] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2240.421781] env[62824]: DEBUG nova.policy [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c379c824a178444e99bb121716fc7887', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0c11aef7fb6247cb9b5272a6a063cd12', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2240.459022] env[62824]: DEBUG oslo_vmware.api [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]521ebc8a-6261-b743-9246-440ec03d4ed6, 'name': SearchDatastore_Task, 'duration_secs': 0.011511} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2240.459427] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2240.459826] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a/b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2240.460213] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d53c2ad3-3ea7-4773-ace3-f58c0f03e925 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2240.469029] env[62824]: DEBUG oslo_vmware.api [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Waiting for the task: (returnval){ [ 2240.469029] env[62824]: value = "task-2146617" [ 2240.469029] env[62824]: _type = "Task" [ 2240.469029] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2240.479286] env[62824]: DEBUG oslo_vmware.api [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Task: {'id': task-2146617, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2240.483070] env[62824]: DEBUG oslo_concurrency.lockutils [None req-4304e8a4-2b17-4cea-b613-125b1018f207 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Acquiring lock "bbfcb2e3-9326-4548-b15b-e054cbfd192e" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2240.483345] env[62824]: DEBUG oslo_concurrency.lockutils [None req-4304e8a4-2b17-4cea-b613-125b1018f207 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Lock "bbfcb2e3-9326-4548-b15b-e054cbfd192e" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2240.483672] env[62824]: INFO nova.compute.manager [None req-4304e8a4-2b17-4cea-b613-125b1018f207 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Attaching volume 10a6ba42-78a0-4e92-b936-f88a95560707 to /dev/sdb [ 2240.527301] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67574ff9-5218-4869-9546-eeed357b32a3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2240.537434] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e94b312d-8463-47dd-b485-8da1988317f1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2240.550533] env[62824]: DEBUG nova.virt.block_device [None req-4304e8a4-2b17-4cea-b613-125b1018f207 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Updating existing volume attachment record: ca707fed-42ac-4167-97c3-f4105ddd7537 {{(pid=62824) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 2240.590469] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquiring lock "refresh_cache-feda7def-7788-4a60-a56a-9353a1475c54" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2240.590646] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquired lock "refresh_cache-feda7def-7788-4a60-a56a-9353a1475c54" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2240.590832] env[62824]: DEBUG nova.network.neutron [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: feda7def-7788-4a60-a56a-9353a1475c54] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2240.610262] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c948012c-66d5-416f-91d2-75a66106d0c6 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquiring lock "refresh_cache-fc77c44a-180c-46ff-9690-9072c6213c91" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2240.610483] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c948012c-66d5-416f-91d2-75a66106d0c6 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquired lock "refresh_cache-fc77c44a-180c-46ff-9690-9072c6213c91" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2240.610614] env[62824]: DEBUG nova.network.neutron [None req-c948012c-66d5-416f-91d2-75a66106d0c6 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2240.692941] env[62824]: DEBUG oslo_vmware.api [None req-e2b575de-d6cc-43f9-9d33-cc6625760cab tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146616, 'name': RemoveSnapshot_Task} progress is 97%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2240.778426] env[62824]: DEBUG oslo_concurrency.lockutils [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Acquiring lock "refresh_cache-71e9733b-c84b-4501-8faf-a487f089b498" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2240.778426] env[62824]: DEBUG oslo_concurrency.lockutils [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Acquired lock "refresh_cache-71e9733b-c84b-4501-8faf-a487f089b498" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2240.778426] env[62824]: DEBUG nova.network.neutron [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2240.795631] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquiring lock "refresh_cache-0394d368-c846-4f3b-bfcf-2bc8e858052a" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2240.795783] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquired lock "refresh_cache-0394d368-c846-4f3b-bfcf-2bc8e858052a" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2240.795933] env[62824]: DEBUG nova.network.neutron [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2240.854751] env[62824]: DEBUG nova.compute.manager [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 204b866f-e9c5-441d-bf8e-77d324b39ecd] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2240.860441] env[62824]: DEBUG nova.network.neutron [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 204b866f-e9c5-441d-bf8e-77d324b39ecd] Successfully created port: e41f4cd6-ab65-4a8d-8514-3be4cde1915b {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2240.980636] env[62824]: DEBUG oslo_vmware.api [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Task: {'id': task-2146617, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.485894} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2240.980636] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a/b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2240.980636] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2240.980636] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-34a5e847-aac1-488d-b3de-13739fe1b047 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2240.986523] env[62824]: DEBUG oslo_vmware.api [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Waiting for the task: (returnval){ [ 2240.986523] env[62824]: value = "task-2146619" [ 2240.986523] env[62824]: _type = "Task" [ 2240.986523] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2240.992390] env[62824]: DEBUG nova.compute.manager [req-186a8313-c7de-447e-9782-de4879e728ca req-7749295f-dafa-49b0-88d6-8e93c4d9955d service nova] [instance: feda7def-7788-4a60-a56a-9353a1475c54] Received event network-vif-plugged-b133d2a7-e5fb-44fd-b716-20381faecc3d {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2240.992609] env[62824]: DEBUG oslo_concurrency.lockutils [req-186a8313-c7de-447e-9782-de4879e728ca req-7749295f-dafa-49b0-88d6-8e93c4d9955d service nova] Acquiring lock "feda7def-7788-4a60-a56a-9353a1475c54-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2240.992816] env[62824]: DEBUG oslo_concurrency.lockutils [req-186a8313-c7de-447e-9782-de4879e728ca req-7749295f-dafa-49b0-88d6-8e93c4d9955d service nova] Lock "feda7def-7788-4a60-a56a-9353a1475c54-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2240.992985] env[62824]: DEBUG oslo_concurrency.lockutils [req-186a8313-c7de-447e-9782-de4879e728ca req-7749295f-dafa-49b0-88d6-8e93c4d9955d service nova] Lock "feda7def-7788-4a60-a56a-9353a1475c54-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2240.993171] env[62824]: DEBUG nova.compute.manager [req-186a8313-c7de-447e-9782-de4879e728ca req-7749295f-dafa-49b0-88d6-8e93c4d9955d service nova] [instance: feda7def-7788-4a60-a56a-9353a1475c54] No waiting events found dispatching network-vif-plugged-b133d2a7-e5fb-44fd-b716-20381faecc3d {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2240.993336] env[62824]: WARNING nova.compute.manager [req-186a8313-c7de-447e-9782-de4879e728ca req-7749295f-dafa-49b0-88d6-8e93c4d9955d service nova] [instance: feda7def-7788-4a60-a56a-9353a1475c54] Received unexpected event network-vif-plugged-b133d2a7-e5fb-44fd-b716-20381faecc3d for instance with vm_state building and task_state spawning. [ 2240.998909] env[62824]: DEBUG oslo_vmware.api [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Task: {'id': task-2146619, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2241.016200] env[62824]: DEBUG nova.compute.manager [req-20a4497f-dbfd-4fcf-a6fe-e5060a471507 req-34720fc4-54b1-4acf-b587-da4bd7fd20e8 service nova] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Received event network-vif-plugged-2c7a5b60-534e-457c-a27b-29a489877783 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2241.016200] env[62824]: DEBUG oslo_concurrency.lockutils [req-20a4497f-dbfd-4fcf-a6fe-e5060a471507 req-34720fc4-54b1-4acf-b587-da4bd7fd20e8 service nova] Acquiring lock "0394d368-c846-4f3b-bfcf-2bc8e858052a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2241.016200] env[62824]: DEBUG oslo_concurrency.lockutils [req-20a4497f-dbfd-4fcf-a6fe-e5060a471507 req-34720fc4-54b1-4acf-b587-da4bd7fd20e8 service nova] Lock "0394d368-c846-4f3b-bfcf-2bc8e858052a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2241.016200] env[62824]: DEBUG oslo_concurrency.lockutils [req-20a4497f-dbfd-4fcf-a6fe-e5060a471507 req-34720fc4-54b1-4acf-b587-da4bd7fd20e8 service nova] Lock "0394d368-c846-4f3b-bfcf-2bc8e858052a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2241.016200] env[62824]: DEBUG nova.compute.manager [req-20a4497f-dbfd-4fcf-a6fe-e5060a471507 req-34720fc4-54b1-4acf-b587-da4bd7fd20e8 service nova] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] No waiting events found dispatching network-vif-plugged-2c7a5b60-534e-457c-a27b-29a489877783 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2241.016200] env[62824]: WARNING nova.compute.manager [req-20a4497f-dbfd-4fcf-a6fe-e5060a471507 req-34720fc4-54b1-4acf-b587-da4bd7fd20e8 service nova] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Received unexpected event network-vif-plugged-2c7a5b60-534e-457c-a27b-29a489877783 for instance with vm_state building and task_state spawning. [ 2241.161808] env[62824]: DEBUG nova.network.neutron [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: feda7def-7788-4a60-a56a-9353a1475c54] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2241.169011] env[62824]: WARNING nova.network.neutron [None req-c948012c-66d5-416f-91d2-75a66106d0c6 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] aa1c09e5-8d72-43ad-a903-a0f5e711da80 already exists in list: networks containing: ['aa1c09e5-8d72-43ad-a903-a0f5e711da80']. ignoring it [ 2241.186776] env[62824]: DEBUG oslo_vmware.api [None req-e2b575de-d6cc-43f9-9d33-cc6625760cab tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146616, 'name': RemoveSnapshot_Task, 'duration_secs': 0.747225} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2241.187205] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e2b575de-d6cc-43f9-9d33-cc6625760cab tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 853f3098-881c-402c-8fd7-26540d1ecd88] Deleted Snapshot of the VM instance {{(pid=62824) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 2241.352763] env[62824]: DEBUG nova.network.neutron [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2241.356619] env[62824]: DEBUG nova.network.neutron [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2241.496561] env[62824]: DEBUG oslo_vmware.api [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Task: {'id': task-2146619, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066375} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2241.497082] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2241.497925] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d4f86f3-422c-42c3-9605-83949bc5547b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2241.522598] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a] Reconfiguring VM instance instance-0000005a to attach disk [datastore2] b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a/b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2241.525383] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1391510a-21b0-426f-812f-116bfb9a4283 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2241.548190] env[62824]: DEBUG oslo_vmware.api [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Waiting for the task: (returnval){ [ 2241.548190] env[62824]: value = "task-2146622" [ 2241.548190] env[62824]: _type = "Task" [ 2241.548190] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2241.556221] env[62824]: DEBUG oslo_vmware.api [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Task: {'id': task-2146622, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2241.561263] env[62824]: DEBUG nova.network.neutron [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: feda7def-7788-4a60-a56a-9353a1475c54] Updating instance_info_cache with network_info: [{"id": "b133d2a7-e5fb-44fd-b716-20381faecc3d", "address": "fa:16:3e:61:ac:de", "network": {"id": "0f512108-670f-4a85-9381-108e18b3b5ed", "bridge": "br-int", "label": "tempest-ServersTestJSON-766397782-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dfff08982dad4790bf4d555e2b4db5e4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f3a2eb5-353f-45c5-a73b-869626f4bb13", "external-id": "nsx-vlan-transportzone-411", "segmentation_id": 411, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb133d2a7-e5", "ovs_interfaceid": "b133d2a7-e5fb-44fd-b716-20381faecc3d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2241.632512] env[62824]: DEBUG nova.network.neutron [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Updating instance_info_cache with network_info: [{"id": "2c7a5b60-534e-457c-a27b-29a489877783", "address": "fa:16:3e:df:88:fa", "network": {"id": "b4ed285c-26a7-461b-8a71-d18e07582ab8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-997150974-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "815ecf9340934733a83d4dcb26612fd4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69f65356-c85e-4b7f-ad28-7c7b5e8cf50c", "external-id": "nsx-vlan-transportzone-281", "segmentation_id": 281, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2c7a5b60-53", "ovs_interfaceid": "2c7a5b60-534e-457c-a27b-29a489877783", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2241.700024] env[62824]: WARNING nova.compute.manager [None req-e2b575de-d6cc-43f9-9d33-cc6625760cab tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 853f3098-881c-402c-8fd7-26540d1ecd88] Image not found during snapshot: nova.exception.ImageNotFound: Image cc58daa7-49cf-44db-bb28-abad81e298fe could not be found. [ 2241.755778] env[62824]: DEBUG nova.network.neutron [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Updating instance_info_cache with network_info: [{"id": "4214bf6e-29f0-4d92-be44-44fccf7ab149", "address": "fa:16:3e:a6:d8:c4", "network": {"id": "fc5d7128-fe9b-44ef-9d91-839190a97226", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-796405330-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35b7289f4746463e9b1cc47fe914a835", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0954fad3-d24d-496c-83e6-a09d3cb556fc", "external-id": "nsx-vlan-transportzone-216", "segmentation_id": 216, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4214bf6e-29", "ovs_interfaceid": "4214bf6e-29f0-4d92-be44-44fccf7ab149", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2241.870096] env[62824]: DEBUG nova.compute.manager [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 204b866f-e9c5-441d-bf8e-77d324b39ecd] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2241.904721] env[62824]: DEBUG nova.virt.hardware [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2241.905174] env[62824]: DEBUG nova.virt.hardware [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2241.905529] env[62824]: DEBUG nova.virt.hardware [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2241.905752] env[62824]: DEBUG nova.virt.hardware [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2241.905905] env[62824]: DEBUG nova.virt.hardware [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2241.906068] env[62824]: DEBUG nova.virt.hardware [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2241.906458] env[62824]: DEBUG nova.virt.hardware [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2241.906832] env[62824]: DEBUG nova.virt.hardware [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2241.907034] env[62824]: DEBUG nova.virt.hardware [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2241.907233] env[62824]: DEBUG nova.virt.hardware [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2241.907532] env[62824]: DEBUG nova.virt.hardware [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2241.908751] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e8f758f-cae8-411e-b19f-9aaa4a8596d2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2241.912826] env[62824]: DEBUG nova.network.neutron [None req-c948012c-66d5-416f-91d2-75a66106d0c6 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Updating instance_info_cache with network_info: [{"id": "e285a842-d54b-4614-84ea-239e34f5b41f", "address": "fa:16:3e:9d:17:cd", "network": {"id": "aa1c09e5-8d72-43ad-a903-a0f5e711da80", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1643616572-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c987c5f844be42c99cc9d57a9ca84c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9bc2632-36f9-4912-8782-8bbb789f909d", "external-id": "nsx-vlan-transportzone-897", "segmentation_id": 897, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape285a842-d5", "ovs_interfaceid": "e285a842-d54b-4614-84ea-239e34f5b41f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "2b6381b4-f3c7-4e22-b698-de190c8199ee", "address": "fa:16:3e:f4:8a:c1", "network": {"id": "aa1c09e5-8d72-43ad-a903-a0f5e711da80", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1643616572-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c987c5f844be42c99cc9d57a9ca84c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9bc2632-36f9-4912-8782-8bbb789f909d", "external-id": "nsx-vlan-transportzone-897", "segmentation_id": 897, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b6381b4-f3", "ovs_interfaceid": "2b6381b4-f3c7-4e22-b698-de190c8199ee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2241.925023] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b73cff62-1620-4431-9d71-f48a73ba9f16 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2242.060467] env[62824]: DEBUG oslo_vmware.api [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Task: {'id': task-2146622, 'name': ReconfigVM_Task, 'duration_secs': 0.278546} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2242.060950] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a] Reconfigured VM instance instance-0000005a to attach disk [datastore2] b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a/b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2242.062229] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-93ce818b-9d8d-4b52-8218-70dd698a750a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2242.066351] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Releasing lock "refresh_cache-feda7def-7788-4a60-a56a-9353a1475c54" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2242.066351] env[62824]: DEBUG nova.compute.manager [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: feda7def-7788-4a60-a56a-9353a1475c54] Instance network_info: |[{"id": "b133d2a7-e5fb-44fd-b716-20381faecc3d", "address": "fa:16:3e:61:ac:de", "network": {"id": "0f512108-670f-4a85-9381-108e18b3b5ed", "bridge": "br-int", "label": "tempest-ServersTestJSON-766397782-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dfff08982dad4790bf4d555e2b4db5e4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f3a2eb5-353f-45c5-a73b-869626f4bb13", "external-id": "nsx-vlan-transportzone-411", "segmentation_id": 411, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb133d2a7-e5", "ovs_interfaceid": "b133d2a7-e5fb-44fd-b716-20381faecc3d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2242.066351] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: feda7def-7788-4a60-a56a-9353a1475c54] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:61:ac:de', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9f3a2eb5-353f-45c5-a73b-869626f4bb13', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b133d2a7-e5fb-44fd-b716-20381faecc3d', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2242.072036] env[62824]: DEBUG oslo.service.loopingcall [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2242.073230] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: feda7def-7788-4a60-a56a-9353a1475c54] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2242.073658] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-14a21b72-0577-4ceb-a100-d5373e20e389 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2242.093368] env[62824]: DEBUG oslo_vmware.api [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Waiting for the task: (returnval){ [ 2242.093368] env[62824]: value = "task-2146623" [ 2242.093368] env[62824]: _type = "Task" [ 2242.093368] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2242.099389] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2242.099389] env[62824]: value = "task-2146624" [ 2242.099389] env[62824]: _type = "Task" [ 2242.099389] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2242.107348] env[62824]: DEBUG oslo_vmware.api [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Task: {'id': task-2146623, 'name': Rename_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2242.114551] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146624, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2242.140431] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Releasing lock "refresh_cache-0394d368-c846-4f3b-bfcf-2bc8e858052a" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2242.140785] env[62824]: DEBUG nova.compute.manager [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Instance network_info: |[{"id": "2c7a5b60-534e-457c-a27b-29a489877783", "address": "fa:16:3e:df:88:fa", "network": {"id": "b4ed285c-26a7-461b-8a71-d18e07582ab8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-997150974-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "815ecf9340934733a83d4dcb26612fd4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69f65356-c85e-4b7f-ad28-7c7b5e8cf50c", "external-id": "nsx-vlan-transportzone-281", "segmentation_id": 281, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2c7a5b60-53", "ovs_interfaceid": "2c7a5b60-534e-457c-a27b-29a489877783", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2242.141307] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:df:88:fa', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '69f65356-c85e-4b7f-ad28-7c7b5e8cf50c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2c7a5b60-534e-457c-a27b-29a489877783', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2242.150949] env[62824]: DEBUG oslo.service.loopingcall [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2242.151305] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2242.151592] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9972c7a1-a19c-4d6a-9807-d6e9cb7d4fe6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2242.173486] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2242.173486] env[62824]: value = "task-2146625" [ 2242.173486] env[62824]: _type = "Task" [ 2242.173486] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2242.182855] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146625, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2242.262554] env[62824]: DEBUG oslo_concurrency.lockutils [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Releasing lock "refresh_cache-71e9733b-c84b-4501-8faf-a487f089b498" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2242.262554] env[62824]: DEBUG nova.compute.manager [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Instance network_info: |[{"id": "4214bf6e-29f0-4d92-be44-44fccf7ab149", "address": "fa:16:3e:a6:d8:c4", "network": {"id": "fc5d7128-fe9b-44ef-9d91-839190a97226", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-796405330-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35b7289f4746463e9b1cc47fe914a835", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0954fad3-d24d-496c-83e6-a09d3cb556fc", "external-id": "nsx-vlan-transportzone-216", "segmentation_id": 216, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4214bf6e-29", "ovs_interfaceid": "4214bf6e-29f0-4d92-be44-44fccf7ab149", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2242.262554] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a6:d8:c4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0954fad3-d24d-496c-83e6-a09d3cb556fc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4214bf6e-29f0-4d92-be44-44fccf7ab149', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2242.273552] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Creating folder: Project (35b7289f4746463e9b1cc47fe914a835). Parent ref: group-v438503. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 2242.273968] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5b0640c2-e980-42dc-81e2-546856f1d220 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2242.285603] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Created folder: Project (35b7289f4746463e9b1cc47fe914a835) in parent group-v438503. [ 2242.285849] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Creating folder: Instances. Parent ref: group-v438730. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 2242.286154] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-613ad534-1c19-43b4-9623-17e566a8e990 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2242.296489] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Created folder: Instances in parent group-v438730. [ 2242.296740] env[62824]: DEBUG oslo.service.loopingcall [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2242.296954] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2242.297564] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a1abf05a-1386-4f4a-9721-de831c6fc0d0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2242.321666] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2242.321666] env[62824]: value = "task-2146628" [ 2242.321666] env[62824]: _type = "Task" [ 2242.321666] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2242.329918] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146628, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2242.416383] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c948012c-66d5-416f-91d2-75a66106d0c6 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Releasing lock "refresh_cache-fc77c44a-180c-46ff-9690-9072c6213c91" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2242.417163] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c948012c-66d5-416f-91d2-75a66106d0c6 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquiring lock "fc77c44a-180c-46ff-9690-9072c6213c91" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2242.417329] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c948012c-66d5-416f-91d2-75a66106d0c6 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquired lock "fc77c44a-180c-46ff-9690-9072c6213c91" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2242.418205] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c7abd53-98e5-487e-973b-bcc8f1f2f93f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2242.438112] env[62824]: DEBUG nova.virt.hardware [None req-c948012c-66d5-416f-91d2-75a66106d0c6 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2242.438112] env[62824]: DEBUG nova.virt.hardware [None req-c948012c-66d5-416f-91d2-75a66106d0c6 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2242.438112] env[62824]: DEBUG nova.virt.hardware [None req-c948012c-66d5-416f-91d2-75a66106d0c6 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2242.438112] env[62824]: DEBUG nova.virt.hardware [None req-c948012c-66d5-416f-91d2-75a66106d0c6 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2242.438112] env[62824]: DEBUG nova.virt.hardware [None req-c948012c-66d5-416f-91d2-75a66106d0c6 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2242.438112] env[62824]: DEBUG nova.virt.hardware [None req-c948012c-66d5-416f-91d2-75a66106d0c6 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2242.438666] env[62824]: DEBUG nova.virt.hardware [None req-c948012c-66d5-416f-91d2-75a66106d0c6 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2242.441931] env[62824]: DEBUG nova.virt.hardware [None req-c948012c-66d5-416f-91d2-75a66106d0c6 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2242.441931] env[62824]: DEBUG nova.virt.hardware [None req-c948012c-66d5-416f-91d2-75a66106d0c6 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2242.441931] env[62824]: DEBUG nova.virt.hardware [None req-c948012c-66d5-416f-91d2-75a66106d0c6 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2242.441931] env[62824]: DEBUG nova.virt.hardware [None req-c948012c-66d5-416f-91d2-75a66106d0c6 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2242.447360] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c948012c-66d5-416f-91d2-75a66106d0c6 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Reconfiguring VM to attach interface {{(pid=62824) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 2242.447772] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-338aa9f0-35ed-45f5-b36c-27f78e4ca491 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2242.469465] env[62824]: DEBUG oslo_vmware.api [None req-c948012c-66d5-416f-91d2-75a66106d0c6 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Waiting for the task: (returnval){ [ 2242.469465] env[62824]: value = "task-2146629" [ 2242.469465] env[62824]: _type = "Task" [ 2242.469465] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2242.479050] env[62824]: DEBUG oslo_vmware.api [None req-c948012c-66d5-416f-91d2-75a66106d0c6 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146629, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2242.611495] env[62824]: DEBUG oslo_vmware.api [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Task: {'id': task-2146623, 'name': Rename_Task, 'duration_secs': 0.148963} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2242.616655] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2242.616655] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146624, 'name': CreateVM_Task, 'duration_secs': 0.394113} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2242.616811] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c53ace7c-eb8c-41d4-920c-7e6c4e9e42f8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2242.618316] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: feda7def-7788-4a60-a56a-9353a1475c54] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2242.619043] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2242.619217] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2242.619546] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2242.620510] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-52e22dc4-b1de-4c68-a180-28b71eb6d725 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2242.625133] env[62824]: DEBUG oslo_vmware.api [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Waiting for the task: (returnval){ [ 2242.625133] env[62824]: value = "task-2146630" [ 2242.625133] env[62824]: _type = "Task" [ 2242.625133] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2242.626356] env[62824]: DEBUG oslo_vmware.api [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2242.626356] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52c548d9-bc37-7e61-7d44-bec06ad7f473" [ 2242.626356] env[62824]: _type = "Task" [ 2242.626356] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2242.637914] env[62824]: DEBUG oslo_vmware.api [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Task: {'id': task-2146630, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2242.641607] env[62824]: DEBUG oslo_vmware.api [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52c548d9-bc37-7e61-7d44-bec06ad7f473, 'name': SearchDatastore_Task, 'duration_secs': 0.010689} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2242.641702] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2242.641922] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: feda7def-7788-4a60-a56a-9353a1475c54] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2242.642166] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2242.642311] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2242.642487] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2242.642737] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1af6d011-55d4-4547-b106-1f0b31954510 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2242.652935] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2242.653463] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2242.654294] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-92f332ee-d93f-4cdf-820a-347514a34c62 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2242.663079] env[62824]: DEBUG oslo_vmware.api [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2242.663079] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52497154-c65f-dfc5-76c0-8241ae662953" [ 2242.663079] env[62824]: _type = "Task" [ 2242.663079] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2242.673026] env[62824]: DEBUG oslo_vmware.api [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52497154-c65f-dfc5-76c0-8241ae662953, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2242.685203] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146625, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2242.691797] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d410e0b6-9b36-4f68-adc4-2fb7e46c625d tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Acquiring lock "853f3098-881c-402c-8fd7-26540d1ecd88" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2242.692367] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d410e0b6-9b36-4f68-adc4-2fb7e46c625d tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Lock "853f3098-881c-402c-8fd7-26540d1ecd88" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2242.692367] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d410e0b6-9b36-4f68-adc4-2fb7e46c625d tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Acquiring lock "853f3098-881c-402c-8fd7-26540d1ecd88-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2242.693270] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d410e0b6-9b36-4f68-adc4-2fb7e46c625d tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Lock "853f3098-881c-402c-8fd7-26540d1ecd88-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2242.693270] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d410e0b6-9b36-4f68-adc4-2fb7e46c625d tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Lock "853f3098-881c-402c-8fd7-26540d1ecd88-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2242.698635] env[62824]: INFO nova.compute.manager [None req-d410e0b6-9b36-4f68-adc4-2fb7e46c625d tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 853f3098-881c-402c-8fd7-26540d1ecd88] Terminating instance [ 2242.835101] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146628, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2242.979661] env[62824]: DEBUG oslo_vmware.api [None req-c948012c-66d5-416f-91d2-75a66106d0c6 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146629, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2243.091614] env[62824]: DEBUG nova.network.neutron [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 204b866f-e9c5-441d-bf8e-77d324b39ecd] Successfully updated port: e41f4cd6-ab65-4a8d-8514-3be4cde1915b {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2243.142499] env[62824]: DEBUG oslo_vmware.api [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Task: {'id': task-2146630, 'name': PowerOnVM_Task, 'duration_secs': 0.500069} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2243.142570] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2243.146022] env[62824]: INFO nova.compute.manager [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a] Took 16.65 seconds to spawn the instance on the hypervisor. [ 2243.146022] env[62824]: DEBUG nova.compute.manager [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2243.146022] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-baf949a0-7b09-4c84-9d5d-e2181f93f551 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2243.177325] env[62824]: DEBUG oslo_vmware.api [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52497154-c65f-dfc5-76c0-8241ae662953, 'name': SearchDatastore_Task, 'duration_secs': 0.011574} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2243.184712] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7f499c98-ddcd-49d0-9072-fa6d7b9696d5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2243.189711] env[62824]: DEBUG nova.compute.manager [req-0e333493-372e-4daa-bc14-137b5ec5ec26 req-5721aac2-bcb5-4238-a0d4-2bb56e5af13a service nova] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Received event network-vif-plugged-2b6381b4-f3c7-4e22-b698-de190c8199ee {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2243.189956] env[62824]: DEBUG oslo_concurrency.lockutils [req-0e333493-372e-4daa-bc14-137b5ec5ec26 req-5721aac2-bcb5-4238-a0d4-2bb56e5af13a service nova] Acquiring lock "fc77c44a-180c-46ff-9690-9072c6213c91-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2243.190250] env[62824]: DEBUG oslo_concurrency.lockutils [req-0e333493-372e-4daa-bc14-137b5ec5ec26 req-5721aac2-bcb5-4238-a0d4-2bb56e5af13a service nova] Lock "fc77c44a-180c-46ff-9690-9072c6213c91-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2243.191018] env[62824]: DEBUG oslo_concurrency.lockutils [req-0e333493-372e-4daa-bc14-137b5ec5ec26 req-5721aac2-bcb5-4238-a0d4-2bb56e5af13a service nova] Lock "fc77c44a-180c-46ff-9690-9072c6213c91-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2243.191018] env[62824]: DEBUG nova.compute.manager [req-0e333493-372e-4daa-bc14-137b5ec5ec26 req-5721aac2-bcb5-4238-a0d4-2bb56e5af13a service nova] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] No waiting events found dispatching network-vif-plugged-2b6381b4-f3c7-4e22-b698-de190c8199ee {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2243.191018] env[62824]: WARNING nova.compute.manager [req-0e333493-372e-4daa-bc14-137b5ec5ec26 req-5721aac2-bcb5-4238-a0d4-2bb56e5af13a service nova] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Received unexpected event network-vif-plugged-2b6381b4-f3c7-4e22-b698-de190c8199ee for instance with vm_state active and task_state None. [ 2243.192241] env[62824]: DEBUG nova.compute.manager [req-0e333493-372e-4daa-bc14-137b5ec5ec26 req-5721aac2-bcb5-4238-a0d4-2bb56e5af13a service nova] [instance: feda7def-7788-4a60-a56a-9353a1475c54] Received event network-changed-b133d2a7-e5fb-44fd-b716-20381faecc3d {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2243.193039] env[62824]: DEBUG nova.compute.manager [req-0e333493-372e-4daa-bc14-137b5ec5ec26 req-5721aac2-bcb5-4238-a0d4-2bb56e5af13a service nova] [instance: feda7def-7788-4a60-a56a-9353a1475c54] Refreshing instance network info cache due to event network-changed-b133d2a7-e5fb-44fd-b716-20381faecc3d. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2243.193039] env[62824]: DEBUG oslo_concurrency.lockutils [req-0e333493-372e-4daa-bc14-137b5ec5ec26 req-5721aac2-bcb5-4238-a0d4-2bb56e5af13a service nova] Acquiring lock "refresh_cache-feda7def-7788-4a60-a56a-9353a1475c54" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2243.193039] env[62824]: DEBUG oslo_concurrency.lockutils [req-0e333493-372e-4daa-bc14-137b5ec5ec26 req-5721aac2-bcb5-4238-a0d4-2bb56e5af13a service nova] Acquired lock "refresh_cache-feda7def-7788-4a60-a56a-9353a1475c54" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2243.193039] env[62824]: DEBUG nova.network.neutron [req-0e333493-372e-4daa-bc14-137b5ec5ec26 req-5721aac2-bcb5-4238-a0d4-2bb56e5af13a service nova] [instance: feda7def-7788-4a60-a56a-9353a1475c54] Refreshing network info cache for port b133d2a7-e5fb-44fd-b716-20381faecc3d {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2243.201156] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146625, 'name': CreateVM_Task, 'duration_secs': 0.634234} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2243.202577] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2243.202937] env[62824]: DEBUG oslo_vmware.api [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2243.202937] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]526c449a-a612-0000-0587-bd0a04bcfbf7" [ 2243.202937] env[62824]: _type = "Task" [ 2243.202937] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2243.206244] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2243.206244] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2243.206244] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2243.206244] env[62824]: DEBUG nova.compute.manager [None req-d410e0b6-9b36-4f68-adc4-2fb7e46c625d tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 853f3098-881c-402c-8fd7-26540d1ecd88] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2243.206244] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-d410e0b6-9b36-4f68-adc4-2fb7e46c625d tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 853f3098-881c-402c-8fd7-26540d1ecd88] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2243.206244] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e86af8c5-da55-4c6f-be10-440467678b94 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2243.210481] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f51b5f18-daa8-4414-9cb1-790ea88d8338 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2243.225856] env[62824]: DEBUG oslo_vmware.api [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]526c449a-a612-0000-0587-bd0a04bcfbf7, 'name': SearchDatastore_Task, 'duration_secs': 0.011398} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2243.229628] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2243.229944] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] feda7def-7788-4a60-a56a-9353a1475c54/feda7def-7788-4a60-a56a-9353a1475c54.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2243.230322] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d410e0b6-9b36-4f68-adc4-2fb7e46c625d tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 853f3098-881c-402c-8fd7-26540d1ecd88] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2243.230769] env[62824]: DEBUG oslo_vmware.api [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2243.230769] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52b10749-cec5-0d64-fc8c-27da727e09e8" [ 2243.230769] env[62824]: _type = "Task" [ 2243.230769] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2243.231379] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f520dc4f-39d4-4380-87ae-d765ed91893b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2243.233571] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-850f84aa-8d6b-422f-ba75-e4c58b584ef4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2243.246072] env[62824]: DEBUG oslo_vmware.api [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52b10749-cec5-0d64-fc8c-27da727e09e8, 'name': SearchDatastore_Task, 'duration_secs': 0.011829} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2243.249591] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2243.249591] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2243.249591] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2243.249591] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2243.250324] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2243.250324] env[62824]: DEBUG oslo_vmware.api [None req-d410e0b6-9b36-4f68-adc4-2fb7e46c625d tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the task: (returnval){ [ 2243.250324] env[62824]: value = "task-2146633" [ 2243.250324] env[62824]: _type = "Task" [ 2243.250324] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2243.250324] env[62824]: DEBUG oslo_vmware.api [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2243.250324] env[62824]: value = "task-2146632" [ 2243.250324] env[62824]: _type = "Task" [ 2243.250324] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2243.250579] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-beda49b5-69f2-4f7d-9839-092d7cefb137 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2243.269894] env[62824]: DEBUG oslo_vmware.api [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146632, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2243.270801] env[62824]: DEBUG oslo_vmware.api [None req-d410e0b6-9b36-4f68-adc4-2fb7e46c625d tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146633, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2243.272267] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2243.272897] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2243.273356] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-538f1941-007b-46f9-ab08-3e623e5c0cbc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2243.280693] env[62824]: DEBUG oslo_vmware.api [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2243.280693] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]523e7d0d-824b-833a-117a-5a62cd969a48" [ 2243.280693] env[62824]: _type = "Task" [ 2243.280693] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2243.290786] env[62824]: DEBUG oslo_vmware.api [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]523e7d0d-824b-833a-117a-5a62cd969a48, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2243.335961] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146628, 'name': CreateVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2243.409241] env[62824]: DEBUG nova.compute.manager [req-a503968a-8d79-4396-96de-bf6b99b836e3 req-a87d61d0-be9a-4d40-8823-234e55f751f0 service nova] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Received event network-changed-2c7a5b60-534e-457c-a27b-29a489877783 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2243.409627] env[62824]: DEBUG nova.compute.manager [req-a503968a-8d79-4396-96de-bf6b99b836e3 req-a87d61d0-be9a-4d40-8823-234e55f751f0 service nova] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Refreshing instance network info cache due to event network-changed-2c7a5b60-534e-457c-a27b-29a489877783. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2243.409938] env[62824]: DEBUG oslo_concurrency.lockutils [req-a503968a-8d79-4396-96de-bf6b99b836e3 req-a87d61d0-be9a-4d40-8823-234e55f751f0 service nova] Acquiring lock "refresh_cache-0394d368-c846-4f3b-bfcf-2bc8e858052a" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2243.410103] env[62824]: DEBUG oslo_concurrency.lockutils [req-a503968a-8d79-4396-96de-bf6b99b836e3 req-a87d61d0-be9a-4d40-8823-234e55f751f0 service nova] Acquired lock "refresh_cache-0394d368-c846-4f3b-bfcf-2bc8e858052a" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2243.410566] env[62824]: DEBUG nova.network.neutron [req-a503968a-8d79-4396-96de-bf6b99b836e3 req-a87d61d0-be9a-4d40-8823-234e55f751f0 service nova] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Refreshing network info cache for port 2c7a5b60-534e-457c-a27b-29a489877783 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2243.484344] env[62824]: DEBUG oslo_vmware.api [None req-c948012c-66d5-416f-91d2-75a66106d0c6 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146629, 'name': ReconfigVM_Task, 'duration_secs': 1.004185} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2243.484878] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c948012c-66d5-416f-91d2-75a66106d0c6 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Releasing lock "fc77c44a-180c-46ff-9690-9072c6213c91" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2243.484959] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c948012c-66d5-416f-91d2-75a66106d0c6 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Reconfigured VM to attach interface {{(pid=62824) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 2243.595420] env[62824]: DEBUG oslo_concurrency.lockutils [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquiring lock "refresh_cache-204b866f-e9c5-441d-bf8e-77d324b39ecd" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2243.595567] env[62824]: DEBUG oslo_concurrency.lockutils [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquired lock "refresh_cache-204b866f-e9c5-441d-bf8e-77d324b39ecd" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2243.595720] env[62824]: DEBUG nova.network.neutron [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 204b866f-e9c5-441d-bf8e-77d324b39ecd] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2243.669365] env[62824]: INFO nova.compute.manager [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a] Took 28.36 seconds to build instance. [ 2243.772628] env[62824]: DEBUG oslo_vmware.api [None req-d410e0b6-9b36-4f68-adc4-2fb7e46c625d tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146633, 'name': PowerOffVM_Task, 'duration_secs': 0.329704} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2243.777808] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d410e0b6-9b36-4f68-adc4-2fb7e46c625d tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 853f3098-881c-402c-8fd7-26540d1ecd88] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2243.778393] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-d410e0b6-9b36-4f68-adc4-2fb7e46c625d tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 853f3098-881c-402c-8fd7-26540d1ecd88] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2243.778908] env[62824]: DEBUG oslo_vmware.api [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146632, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2243.779389] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4d73cb17-9958-4824-aa5f-f52ce8e98f72 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2243.795474] env[62824]: DEBUG oslo_vmware.api [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]523e7d0d-824b-833a-117a-5a62cd969a48, 'name': SearchDatastore_Task, 'duration_secs': 0.01217} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2243.796481] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-41413112-dcbd-428a-9a30-86be03af8286 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2243.802093] env[62824]: DEBUG oslo_vmware.api [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2243.802093] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52097ca7-a27a-703e-ff64-d375f82cd364" [ 2243.802093] env[62824]: _type = "Task" [ 2243.802093] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2243.812216] env[62824]: DEBUG oslo_vmware.api [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52097ca7-a27a-703e-ff64-d375f82cd364, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2243.834354] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146628, 'name': CreateVM_Task, 'duration_secs': 1.155634} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2243.836619] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2243.837292] env[62824]: DEBUG oslo_concurrency.lockutils [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2243.837495] env[62824]: DEBUG oslo_concurrency.lockutils [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2243.837759] env[62824]: DEBUG oslo_concurrency.lockutils [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2243.838501] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-18dbc5a9-a816-475c-9ade-e5abb87ee48e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2243.843858] env[62824]: DEBUG oslo_vmware.api [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Waiting for the task: (returnval){ [ 2243.843858] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52e3a261-79fc-cc64-baf0-79d40182290f" [ 2243.843858] env[62824]: _type = "Task" [ 2243.843858] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2243.852929] env[62824]: DEBUG oslo_vmware.api [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52e3a261-79fc-cc64-baf0-79d40182290f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2243.965781] env[62824]: DEBUG nova.network.neutron [req-0e333493-372e-4daa-bc14-137b5ec5ec26 req-5721aac2-bcb5-4238-a0d4-2bb56e5af13a service nova] [instance: feda7def-7788-4a60-a56a-9353a1475c54] Updated VIF entry in instance network info cache for port b133d2a7-e5fb-44fd-b716-20381faecc3d. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2243.965781] env[62824]: DEBUG nova.network.neutron [req-0e333493-372e-4daa-bc14-137b5ec5ec26 req-5721aac2-bcb5-4238-a0d4-2bb56e5af13a service nova] [instance: feda7def-7788-4a60-a56a-9353a1475c54] Updating instance_info_cache with network_info: [{"id": "b133d2a7-e5fb-44fd-b716-20381faecc3d", "address": "fa:16:3e:61:ac:de", "network": {"id": "0f512108-670f-4a85-9381-108e18b3b5ed", "bridge": "br-int", "label": "tempest-ServersTestJSON-766397782-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dfff08982dad4790bf4d555e2b4db5e4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f3a2eb5-353f-45c5-a73b-869626f4bb13", "external-id": "nsx-vlan-transportzone-411", "segmentation_id": 411, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb133d2a7-e5", "ovs_interfaceid": "b133d2a7-e5fb-44fd-b716-20381faecc3d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2243.990307] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c948012c-66d5-416f-91d2-75a66106d0c6 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lock "interface-fc77c44a-180c-46ff-9690-9072c6213c91-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 14.807s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2244.023125] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-d410e0b6-9b36-4f68-adc4-2fb7e46c625d tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 853f3098-881c-402c-8fd7-26540d1ecd88] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2244.023390] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-d410e0b6-9b36-4f68-adc4-2fb7e46c625d tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 853f3098-881c-402c-8fd7-26540d1ecd88] Deleting contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2244.023660] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-d410e0b6-9b36-4f68-adc4-2fb7e46c625d tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Deleting the datastore file [datastore1] 853f3098-881c-402c-8fd7-26540d1ecd88 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2244.023785] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cefd02b8-b464-4e9b-9cc6-e2b7b7cd09ad {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2244.032689] env[62824]: DEBUG oslo_vmware.api [None req-d410e0b6-9b36-4f68-adc4-2fb7e46c625d tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for the task: (returnval){ [ 2244.032689] env[62824]: value = "task-2146635" [ 2244.032689] env[62824]: _type = "Task" [ 2244.032689] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2244.045685] env[62824]: DEBUG oslo_vmware.api [None req-d410e0b6-9b36-4f68-adc4-2fb7e46c625d tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146635, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2244.147137] env[62824]: DEBUG nova.network.neutron [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 204b866f-e9c5-441d-bf8e-77d324b39ecd] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2244.171391] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3a687207-392a-4503-bdf0-1fff833d515c tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Lock "b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.884s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2244.267584] env[62824]: DEBUG oslo_vmware.api [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146632, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.55693} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2244.267949] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] feda7def-7788-4a60-a56a-9353a1475c54/feda7def-7788-4a60-a56a-9353a1475c54.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2244.268472] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: feda7def-7788-4a60-a56a-9353a1475c54] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2244.268836] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9bed5495-cd45-47af-aaf3-4b340048a11a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2244.274624] env[62824]: DEBUG nova.network.neutron [req-a503968a-8d79-4396-96de-bf6b99b836e3 req-a87d61d0-be9a-4d40-8823-234e55f751f0 service nova] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Updated VIF entry in instance network info cache for port 2c7a5b60-534e-457c-a27b-29a489877783. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2244.276209] env[62824]: DEBUG nova.network.neutron [req-a503968a-8d79-4396-96de-bf6b99b836e3 req-a87d61d0-be9a-4d40-8823-234e55f751f0 service nova] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Updating instance_info_cache with network_info: [{"id": "2c7a5b60-534e-457c-a27b-29a489877783", "address": "fa:16:3e:df:88:fa", "network": {"id": "b4ed285c-26a7-461b-8a71-d18e07582ab8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-997150974-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "815ecf9340934733a83d4dcb26612fd4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69f65356-c85e-4b7f-ad28-7c7b5e8cf50c", "external-id": "nsx-vlan-transportzone-281", "segmentation_id": 281, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2c7a5b60-53", "ovs_interfaceid": "2c7a5b60-534e-457c-a27b-29a489877783", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2244.279665] env[62824]: DEBUG oslo_vmware.api [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2244.279665] env[62824]: value = "task-2146636" [ 2244.279665] env[62824]: _type = "Task" [ 2244.279665] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2244.296695] env[62824]: DEBUG oslo_vmware.api [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146636, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2244.314269] env[62824]: DEBUG oslo_vmware.api [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52097ca7-a27a-703e-ff64-d375f82cd364, 'name': SearchDatastore_Task, 'duration_secs': 0.02483} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2244.314565] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2244.314840] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 0394d368-c846-4f3b-bfcf-2bc8e858052a/0394d368-c846-4f3b-bfcf-2bc8e858052a.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2244.315143] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-43a8fb42-12c9-480e-8e90-661c1ef33994 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2244.323213] env[62824]: DEBUG oslo_vmware.api [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2244.323213] env[62824]: value = "task-2146637" [ 2244.323213] env[62824]: _type = "Task" [ 2244.323213] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2244.335186] env[62824]: DEBUG oslo_vmware.api [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146637, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2244.353348] env[62824]: DEBUG oslo_vmware.api [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52e3a261-79fc-cc64-baf0-79d40182290f, 'name': SearchDatastore_Task, 'duration_secs': 0.018234} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2244.353690] env[62824]: DEBUG oslo_concurrency.lockutils [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2244.353938] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2244.354204] env[62824]: DEBUG oslo_concurrency.lockutils [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2244.354365] env[62824]: DEBUG oslo_concurrency.lockutils [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2244.354553] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2244.354820] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-02d3ab1b-f92c-4c59-8299-2a92e57a4e73 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2244.358398] env[62824]: DEBUG nova.network.neutron [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 204b866f-e9c5-441d-bf8e-77d324b39ecd] Updating instance_info_cache with network_info: [{"id": "e41f4cd6-ab65-4a8d-8514-3be4cde1915b", "address": "fa:16:3e:29:83:fc", "network": {"id": "b995934c-4934-4469-884c-73cb4de1796f", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1892307918-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c11aef7fb6247cb9b5272a6a063cd12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15165046-2de9-4ada-9e99-0126e20854a9", "external-id": "nsx-vlan-transportzone-974", "segmentation_id": 974, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape41f4cd6-ab", "ovs_interfaceid": "e41f4cd6-ab65-4a8d-8514-3be4cde1915b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2244.377528] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2244.377528] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2244.378803] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-36f4a454-d822-401e-90ce-7e8890eda3f1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2244.385437] env[62824]: DEBUG oslo_vmware.api [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Waiting for the task: (returnval){ [ 2244.385437] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5253919a-30c8-0fcf-41ff-ea1a07211e74" [ 2244.385437] env[62824]: _type = "Task" [ 2244.385437] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2244.393711] env[62824]: DEBUG oslo_vmware.api [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5253919a-30c8-0fcf-41ff-ea1a07211e74, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2244.468790] env[62824]: DEBUG oslo_concurrency.lockutils [req-0e333493-372e-4daa-bc14-137b5ec5ec26 req-5721aac2-bcb5-4238-a0d4-2bb56e5af13a service nova] Releasing lock "refresh_cache-feda7def-7788-4a60-a56a-9353a1475c54" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2244.469146] env[62824]: DEBUG nova.compute.manager [req-0e333493-372e-4daa-bc14-137b5ec5ec26 req-5721aac2-bcb5-4238-a0d4-2bb56e5af13a service nova] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Received event network-changed-2b6381b4-f3c7-4e22-b698-de190c8199ee {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2244.469345] env[62824]: DEBUG nova.compute.manager [req-0e333493-372e-4daa-bc14-137b5ec5ec26 req-5721aac2-bcb5-4238-a0d4-2bb56e5af13a service nova] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Refreshing instance network info cache due to event network-changed-2b6381b4-f3c7-4e22-b698-de190c8199ee. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2244.469565] env[62824]: DEBUG oslo_concurrency.lockutils [req-0e333493-372e-4daa-bc14-137b5ec5ec26 req-5721aac2-bcb5-4238-a0d4-2bb56e5af13a service nova] Acquiring lock "refresh_cache-fc77c44a-180c-46ff-9690-9072c6213c91" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2244.469710] env[62824]: DEBUG oslo_concurrency.lockutils [req-0e333493-372e-4daa-bc14-137b5ec5ec26 req-5721aac2-bcb5-4238-a0d4-2bb56e5af13a service nova] Acquired lock "refresh_cache-fc77c44a-180c-46ff-9690-9072c6213c91" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2244.469873] env[62824]: DEBUG nova.network.neutron [req-0e333493-372e-4daa-bc14-137b5ec5ec26 req-5721aac2-bcb5-4238-a0d4-2bb56e5af13a service nova] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Refreshing network info cache for port 2b6381b4-f3c7-4e22-b698-de190c8199ee {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2244.544894] env[62824]: DEBUG oslo_vmware.api [None req-d410e0b6-9b36-4f68-adc4-2fb7e46c625d tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146635, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2244.610922] env[62824]: INFO nova.compute.manager [None req-697b1a23-d14a-4d94-b716-8c4b25f6df27 tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Rebuilding instance [ 2244.670108] env[62824]: DEBUG nova.compute.manager [None req-697b1a23-d14a-4d94-b716-8c4b25f6df27 tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2244.670108] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60060738-5427-45bc-b3a4-5dbdf88b1e66 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2244.779849] env[62824]: DEBUG oslo_concurrency.lockutils [req-a503968a-8d79-4396-96de-bf6b99b836e3 req-a87d61d0-be9a-4d40-8823-234e55f751f0 service nova] Releasing lock "refresh_cache-0394d368-c846-4f3b-bfcf-2bc8e858052a" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2244.779849] env[62824]: DEBUG nova.compute.manager [req-a503968a-8d79-4396-96de-bf6b99b836e3 req-a87d61d0-be9a-4d40-8823-234e55f751f0 service nova] [instance: 204b866f-e9c5-441d-bf8e-77d324b39ecd] Received event network-vif-plugged-e41f4cd6-ab65-4a8d-8514-3be4cde1915b {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2244.779849] env[62824]: DEBUG oslo_concurrency.lockutils [req-a503968a-8d79-4396-96de-bf6b99b836e3 req-a87d61d0-be9a-4d40-8823-234e55f751f0 service nova] Acquiring lock "204b866f-e9c5-441d-bf8e-77d324b39ecd-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2244.779849] env[62824]: DEBUG oslo_concurrency.lockutils [req-a503968a-8d79-4396-96de-bf6b99b836e3 req-a87d61d0-be9a-4d40-8823-234e55f751f0 service nova] Lock "204b866f-e9c5-441d-bf8e-77d324b39ecd-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2244.779849] env[62824]: DEBUG oslo_concurrency.lockutils [req-a503968a-8d79-4396-96de-bf6b99b836e3 req-a87d61d0-be9a-4d40-8823-234e55f751f0 service nova] Lock "204b866f-e9c5-441d-bf8e-77d324b39ecd-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2244.779849] env[62824]: DEBUG nova.compute.manager [req-a503968a-8d79-4396-96de-bf6b99b836e3 req-a87d61d0-be9a-4d40-8823-234e55f751f0 service nova] [instance: 204b866f-e9c5-441d-bf8e-77d324b39ecd] No waiting events found dispatching network-vif-plugged-e41f4cd6-ab65-4a8d-8514-3be4cde1915b {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2244.779849] env[62824]: WARNING nova.compute.manager [req-a503968a-8d79-4396-96de-bf6b99b836e3 req-a87d61d0-be9a-4d40-8823-234e55f751f0 service nova] [instance: 204b866f-e9c5-441d-bf8e-77d324b39ecd] Received unexpected event network-vif-plugged-e41f4cd6-ab65-4a8d-8514-3be4cde1915b for instance with vm_state building and task_state spawning. [ 2244.779849] env[62824]: DEBUG nova.compute.manager [req-a503968a-8d79-4396-96de-bf6b99b836e3 req-a87d61d0-be9a-4d40-8823-234e55f751f0 service nova] [instance: 204b866f-e9c5-441d-bf8e-77d324b39ecd] Received event network-changed-e41f4cd6-ab65-4a8d-8514-3be4cde1915b {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2244.780224] env[62824]: DEBUG nova.compute.manager [req-a503968a-8d79-4396-96de-bf6b99b836e3 req-a87d61d0-be9a-4d40-8823-234e55f751f0 service nova] [instance: 204b866f-e9c5-441d-bf8e-77d324b39ecd] Refreshing instance network info cache due to event network-changed-e41f4cd6-ab65-4a8d-8514-3be4cde1915b. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2244.780708] env[62824]: DEBUG oslo_concurrency.lockutils [req-a503968a-8d79-4396-96de-bf6b99b836e3 req-a87d61d0-be9a-4d40-8823-234e55f751f0 service nova] Acquiring lock "refresh_cache-204b866f-e9c5-441d-bf8e-77d324b39ecd" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2244.789649] env[62824]: DEBUG oslo_vmware.api [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146636, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065039} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2244.789981] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: feda7def-7788-4a60-a56a-9353a1475c54] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2244.790842] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9769c2f-8ee1-468b-b2de-d646b677b332 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2244.820498] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: feda7def-7788-4a60-a56a-9353a1475c54] Reconfiguring VM instance instance-0000005b to attach disk [datastore2] feda7def-7788-4a60-a56a-9353a1475c54/feda7def-7788-4a60-a56a-9353a1475c54.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2244.821145] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d585b845-2437-460a-8436-640027c4d04b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2244.852048] env[62824]: DEBUG oslo_vmware.api [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146637, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2244.853494] env[62824]: DEBUG oslo_vmware.api [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2244.853494] env[62824]: value = "task-2146638" [ 2244.853494] env[62824]: _type = "Task" [ 2244.853494] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2244.862051] env[62824]: DEBUG oslo_concurrency.lockutils [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Releasing lock "refresh_cache-204b866f-e9c5-441d-bf8e-77d324b39ecd" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2244.862480] env[62824]: DEBUG nova.compute.manager [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 204b866f-e9c5-441d-bf8e-77d324b39ecd] Instance network_info: |[{"id": "e41f4cd6-ab65-4a8d-8514-3be4cde1915b", "address": "fa:16:3e:29:83:fc", "network": {"id": "b995934c-4934-4469-884c-73cb4de1796f", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1892307918-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c11aef7fb6247cb9b5272a6a063cd12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15165046-2de9-4ada-9e99-0126e20854a9", "external-id": "nsx-vlan-transportzone-974", "segmentation_id": 974, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape41f4cd6-ab", "ovs_interfaceid": "e41f4cd6-ab65-4a8d-8514-3be4cde1915b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2244.863199] env[62824]: DEBUG oslo_vmware.api [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146638, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2244.863692] env[62824]: DEBUG oslo_concurrency.lockutils [req-a503968a-8d79-4396-96de-bf6b99b836e3 req-a87d61d0-be9a-4d40-8823-234e55f751f0 service nova] Acquired lock "refresh_cache-204b866f-e9c5-441d-bf8e-77d324b39ecd" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2244.864322] env[62824]: DEBUG nova.network.neutron [req-a503968a-8d79-4396-96de-bf6b99b836e3 req-a87d61d0-be9a-4d40-8823-234e55f751f0 service nova] [instance: 204b866f-e9c5-441d-bf8e-77d324b39ecd] Refreshing network info cache for port e41f4cd6-ab65-4a8d-8514-3be4cde1915b {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2244.865817] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 204b866f-e9c5-441d-bf8e-77d324b39ecd] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:29:83:fc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '15165046-2de9-4ada-9e99-0126e20854a9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e41f4cd6-ab65-4a8d-8514-3be4cde1915b', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2244.875322] env[62824]: DEBUG oslo.service.loopingcall [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2244.876250] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 204b866f-e9c5-441d-bf8e-77d324b39ecd] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2244.876557] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7f55e97a-7125-4d09-8656-7ce892b11eae {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2244.904707] env[62824]: DEBUG oslo_vmware.api [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5253919a-30c8-0fcf-41ff-ea1a07211e74, 'name': SearchDatastore_Task, 'duration_secs': 0.039085} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2244.907368] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2244.907368] env[62824]: value = "task-2146639" [ 2244.907368] env[62824]: _type = "Task" [ 2244.907368] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2244.907831] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4141405a-a2ce-4490-8df7-dd577172ceb2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2244.916593] env[62824]: DEBUG oslo_vmware.api [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Waiting for the task: (returnval){ [ 2244.916593] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5256ddb7-6a18-e3cb-bcaa-70f5ff6276e1" [ 2244.916593] env[62824]: _type = "Task" [ 2244.916593] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2244.920335] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146639, 'name': CreateVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2244.929385] env[62824]: DEBUG oslo_vmware.api [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5256ddb7-6a18-e3cb-bcaa-70f5ff6276e1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2245.048186] env[62824]: DEBUG oslo_vmware.api [None req-d410e0b6-9b36-4f68-adc4-2fb7e46c625d tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Task: {'id': task-2146635, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.593545} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2245.048186] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-d410e0b6-9b36-4f68-adc4-2fb7e46c625d tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2245.048186] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-d410e0b6-9b36-4f68-adc4-2fb7e46c625d tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 853f3098-881c-402c-8fd7-26540d1ecd88] Deleted contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2245.048186] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-d410e0b6-9b36-4f68-adc4-2fb7e46c625d tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 853f3098-881c-402c-8fd7-26540d1ecd88] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2245.048186] env[62824]: INFO nova.compute.manager [None req-d410e0b6-9b36-4f68-adc4-2fb7e46c625d tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] [instance: 853f3098-881c-402c-8fd7-26540d1ecd88] Took 1.84 seconds to destroy the instance on the hypervisor. [ 2245.048186] env[62824]: DEBUG oslo.service.loopingcall [None req-d410e0b6-9b36-4f68-adc4-2fb7e46c625d tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2245.048186] env[62824]: DEBUG nova.compute.manager [-] [instance: 853f3098-881c-402c-8fd7-26540d1ecd88] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2245.048186] env[62824]: DEBUG nova.network.neutron [-] [instance: 853f3098-881c-402c-8fd7-26540d1ecd88] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2245.352174] env[62824]: DEBUG oslo_vmware.api [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146637, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2245.364055] env[62824]: DEBUG oslo_vmware.api [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146638, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2245.422969] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146639, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2245.435937] env[62824]: DEBUG oslo_vmware.api [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5256ddb7-6a18-e3cb-bcaa-70f5ff6276e1, 'name': SearchDatastore_Task, 'duration_secs': 0.022326} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2245.436441] env[62824]: DEBUG oslo_concurrency.lockutils [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2245.436788] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 71e9733b-c84b-4501-8faf-a487f089b498/71e9733b-c84b-4501-8faf-a487f089b498.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2245.438440] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1490eba7-4481-42db-9db5-c78a53fe16e1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2245.446191] env[62824]: DEBUG oslo_vmware.api [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Waiting for the task: (returnval){ [ 2245.446191] env[62824]: value = "task-2146640" [ 2245.446191] env[62824]: _type = "Task" [ 2245.446191] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2245.455881] env[62824]: DEBUG oslo_vmware.api [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146640, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2245.528383] env[62824]: DEBUG nova.compute.manager [req-af1186a1-89d9-4ca7-8e28-857186e1ce36 req-6732ec93-207d-48bc-97dc-3a4e2178e2ea service nova] [instance: 853f3098-881c-402c-8fd7-26540d1ecd88] Received event network-vif-deleted-03eb70c1-fc19-4e7c-874d-2f4fcd77fe6b {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2245.528583] env[62824]: INFO nova.compute.manager [req-af1186a1-89d9-4ca7-8e28-857186e1ce36 req-6732ec93-207d-48bc-97dc-3a4e2178e2ea service nova] [instance: 853f3098-881c-402c-8fd7-26540d1ecd88] Neutron deleted interface 03eb70c1-fc19-4e7c-874d-2f4fcd77fe6b; detaching it from the instance and deleting it from the info cache [ 2245.528758] env[62824]: DEBUG nova.network.neutron [req-af1186a1-89d9-4ca7-8e28-857186e1ce36 req-6732ec93-207d-48bc-97dc-3a4e2178e2ea service nova] [instance: 853f3098-881c-402c-8fd7-26540d1ecd88] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2245.541124] env[62824]: DEBUG nova.network.neutron [req-0e333493-372e-4daa-bc14-137b5ec5ec26 req-5721aac2-bcb5-4238-a0d4-2bb56e5af13a service nova] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Updated VIF entry in instance network info cache for port 2b6381b4-f3c7-4e22-b698-de190c8199ee. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2245.541574] env[62824]: DEBUG nova.network.neutron [req-0e333493-372e-4daa-bc14-137b5ec5ec26 req-5721aac2-bcb5-4238-a0d4-2bb56e5af13a service nova] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Updating instance_info_cache with network_info: [{"id": "e285a842-d54b-4614-84ea-239e34f5b41f", "address": "fa:16:3e:9d:17:cd", "network": {"id": "aa1c09e5-8d72-43ad-a903-a0f5e711da80", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1643616572-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c987c5f844be42c99cc9d57a9ca84c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9bc2632-36f9-4912-8782-8bbb789f909d", "external-id": "nsx-vlan-transportzone-897", "segmentation_id": 897, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape285a842-d5", "ovs_interfaceid": "e285a842-d54b-4614-84ea-239e34f5b41f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "2b6381b4-f3c7-4e22-b698-de190c8199ee", "address": "fa:16:3e:f4:8a:c1", "network": {"id": "aa1c09e5-8d72-43ad-a903-a0f5e711da80", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1643616572-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c987c5f844be42c99cc9d57a9ca84c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9bc2632-36f9-4912-8782-8bbb789f909d", "external-id": "nsx-vlan-transportzone-897", "segmentation_id": 897, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b6381b4-f3", "ovs_interfaceid": "2b6381b4-f3c7-4e22-b698-de190c8199ee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2245.604095] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-4304e8a4-2b17-4cea-b613-125b1018f207 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Volume attach. Driver type: vmdk {{(pid=62824) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 2245.604303] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-4304e8a4-2b17-4cea-b613-125b1018f207 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-438727', 'volume_id': '10a6ba42-78a0-4e92-b936-f88a95560707', 'name': 'volume-10a6ba42-78a0-4e92-b936-f88a95560707', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'bbfcb2e3-9326-4548-b15b-e054cbfd192e', 'attached_at': '', 'detached_at': '', 'volume_id': '10a6ba42-78a0-4e92-b936-f88a95560707', 'serial': '10a6ba42-78a0-4e92-b936-f88a95560707'} {{(pid=62824) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 2245.605237] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2095c5a-85b0-455a-9b50-4807ed19e0c2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2245.630225] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49b7b810-dbad-4251-bfac-48f4a9ef2d69 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2245.657818] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-4304e8a4-2b17-4cea-b613-125b1018f207 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Reconfiguring VM instance instance-00000039 to attach disk [datastore2] volume-10a6ba42-78a0-4e92-b936-f88a95560707/volume-10a6ba42-78a0-4e92-b936-f88a95560707.vmdk or device None with type thin {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2245.657818] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-50620bb2-2045-410c-aecf-cedafa002db0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2245.671743] env[62824]: DEBUG nova.network.neutron [req-a503968a-8d79-4396-96de-bf6b99b836e3 req-a87d61d0-be9a-4d40-8823-234e55f751f0 service nova] [instance: 204b866f-e9c5-441d-bf8e-77d324b39ecd] Updated VIF entry in instance network info cache for port e41f4cd6-ab65-4a8d-8514-3be4cde1915b. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2245.672225] env[62824]: DEBUG nova.network.neutron [req-a503968a-8d79-4396-96de-bf6b99b836e3 req-a87d61d0-be9a-4d40-8823-234e55f751f0 service nova] [instance: 204b866f-e9c5-441d-bf8e-77d324b39ecd] Updating instance_info_cache with network_info: [{"id": "e41f4cd6-ab65-4a8d-8514-3be4cde1915b", "address": "fa:16:3e:29:83:fc", "network": {"id": "b995934c-4934-4469-884c-73cb4de1796f", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1892307918-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c11aef7fb6247cb9b5272a6a063cd12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15165046-2de9-4ada-9e99-0126e20854a9", "external-id": "nsx-vlan-transportzone-974", "segmentation_id": 974, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape41f4cd6-ab", "ovs_interfaceid": "e41f4cd6-ab65-4a8d-8514-3be4cde1915b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2245.680833] env[62824]: DEBUG oslo_vmware.api [None req-4304e8a4-2b17-4cea-b613-125b1018f207 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for the task: (returnval){ [ 2245.680833] env[62824]: value = "task-2146641" [ 2245.680833] env[62824]: _type = "Task" [ 2245.680833] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2245.686900] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-697b1a23-d14a-4d94-b716-8c4b25f6df27 tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2245.687388] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b693f495-1f99-4f5b-9702-41d400661f10 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2245.694617] env[62824]: DEBUG oslo_vmware.api [None req-4304e8a4-2b17-4cea-b613-125b1018f207 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2146641, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2245.696035] env[62824]: DEBUG oslo_vmware.api [None req-697b1a23-d14a-4d94-b716-8c4b25f6df27 tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Waiting for the task: (returnval){ [ 2245.696035] env[62824]: value = "task-2146642" [ 2245.696035] env[62824]: _type = "Task" [ 2245.696035] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2245.706213] env[62824]: DEBUG oslo_vmware.api [None req-697b1a23-d14a-4d94-b716-8c4b25f6df27 tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Task: {'id': task-2146642, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2245.853325] env[62824]: DEBUG oslo_vmware.api [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146637, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.043836} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2245.853854] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 0394d368-c846-4f3b-bfcf-2bc8e858052a/0394d368-c846-4f3b-bfcf-2bc8e858052a.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2245.854185] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2245.854586] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3d14f9a9-4526-45dc-8ddd-4e72511fae46 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2245.868830] env[62824]: DEBUG oslo_vmware.api [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146638, 'name': ReconfigVM_Task, 'duration_secs': 0.518316} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2245.870576] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: feda7def-7788-4a60-a56a-9353a1475c54] Reconfigured VM instance instance-0000005b to attach disk [datastore2] feda7def-7788-4a60-a56a-9353a1475c54/feda7def-7788-4a60-a56a-9353a1475c54.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2245.871510] env[62824]: DEBUG oslo_vmware.api [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2245.871510] env[62824]: value = "task-2146643" [ 2245.871510] env[62824]: _type = "Task" [ 2245.871510] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2245.871745] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e01f78a4-7f1c-40a5-9b3c-91279760a883 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2245.886683] env[62824]: DEBUG oslo_vmware.api [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146643, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2245.889438] env[62824]: DEBUG oslo_vmware.api [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2245.889438] env[62824]: value = "task-2146644" [ 2245.889438] env[62824]: _type = "Task" [ 2245.889438] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2245.900833] env[62824]: DEBUG oslo_vmware.api [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146644, 'name': Rename_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2245.926034] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146639, 'name': CreateVM_Task, 'duration_secs': 0.646799} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2245.926034] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 204b866f-e9c5-441d-bf8e-77d324b39ecd] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2245.926761] env[62824]: DEBUG oslo_concurrency.lockutils [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2245.927217] env[62824]: DEBUG oslo_concurrency.lockutils [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2245.927330] env[62824]: DEBUG oslo_concurrency.lockutils [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2245.927622] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1dd0f41f-d6ab-4556-9ffe-8df2a1997f8a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2245.933716] env[62824]: DEBUG oslo_vmware.api [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 2245.933716] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]529c3195-ae95-762b-c925-6244c6fb1af5" [ 2245.933716] env[62824]: _type = "Task" [ 2245.933716] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2245.950031] env[62824]: DEBUG oslo_vmware.api [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]529c3195-ae95-762b-c925-6244c6fb1af5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2245.960933] env[62824]: DEBUG oslo_vmware.api [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146640, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2246.004812] env[62824]: DEBUG nova.network.neutron [-] [instance: 853f3098-881c-402c-8fd7-26540d1ecd88] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2246.037034] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1321dca0-aae3-44ff-8975-a84a145b1dbf {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2246.047583] env[62824]: DEBUG oslo_concurrency.lockutils [req-0e333493-372e-4daa-bc14-137b5ec5ec26 req-5721aac2-bcb5-4238-a0d4-2bb56e5af13a service nova] Releasing lock "refresh_cache-fc77c44a-180c-46ff-9690-9072c6213c91" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2246.047583] env[62824]: DEBUG nova.compute.manager [req-0e333493-372e-4daa-bc14-137b5ec5ec26 req-5721aac2-bcb5-4238-a0d4-2bb56e5af13a service nova] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Received event network-vif-plugged-4214bf6e-29f0-4d92-be44-44fccf7ab149 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2246.047583] env[62824]: DEBUG oslo_concurrency.lockutils [req-0e333493-372e-4daa-bc14-137b5ec5ec26 req-5721aac2-bcb5-4238-a0d4-2bb56e5af13a service nova] Acquiring lock "71e9733b-c84b-4501-8faf-a487f089b498-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2246.047583] env[62824]: DEBUG oslo_concurrency.lockutils [req-0e333493-372e-4daa-bc14-137b5ec5ec26 req-5721aac2-bcb5-4238-a0d4-2bb56e5af13a service nova] Lock "71e9733b-c84b-4501-8faf-a487f089b498-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2246.047583] env[62824]: DEBUG oslo_concurrency.lockutils [req-0e333493-372e-4daa-bc14-137b5ec5ec26 req-5721aac2-bcb5-4238-a0d4-2bb56e5af13a service nova] Lock "71e9733b-c84b-4501-8faf-a487f089b498-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2246.047583] env[62824]: DEBUG nova.compute.manager [req-0e333493-372e-4daa-bc14-137b5ec5ec26 req-5721aac2-bcb5-4238-a0d4-2bb56e5af13a service nova] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] No waiting events found dispatching network-vif-plugged-4214bf6e-29f0-4d92-be44-44fccf7ab149 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2246.047583] env[62824]: WARNING nova.compute.manager [req-0e333493-372e-4daa-bc14-137b5ec5ec26 req-5721aac2-bcb5-4238-a0d4-2bb56e5af13a service nova] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Received unexpected event network-vif-plugged-4214bf6e-29f0-4d92-be44-44fccf7ab149 for instance with vm_state building and task_state spawning. [ 2246.047956] env[62824]: DEBUG nova.compute.manager [req-0e333493-372e-4daa-bc14-137b5ec5ec26 req-5721aac2-bcb5-4238-a0d4-2bb56e5af13a service nova] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Received event network-changed-4214bf6e-29f0-4d92-be44-44fccf7ab149 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2246.047956] env[62824]: DEBUG nova.compute.manager [req-0e333493-372e-4daa-bc14-137b5ec5ec26 req-5721aac2-bcb5-4238-a0d4-2bb56e5af13a service nova] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Refreshing instance network info cache due to event network-changed-4214bf6e-29f0-4d92-be44-44fccf7ab149. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2246.048156] env[62824]: DEBUG oslo_concurrency.lockutils [req-0e333493-372e-4daa-bc14-137b5ec5ec26 req-5721aac2-bcb5-4238-a0d4-2bb56e5af13a service nova] Acquiring lock "refresh_cache-71e9733b-c84b-4501-8faf-a487f089b498" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2246.048281] env[62824]: DEBUG oslo_concurrency.lockutils [req-0e333493-372e-4daa-bc14-137b5ec5ec26 req-5721aac2-bcb5-4238-a0d4-2bb56e5af13a service nova] Acquired lock "refresh_cache-71e9733b-c84b-4501-8faf-a487f089b498" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2246.048439] env[62824]: DEBUG nova.network.neutron [req-0e333493-372e-4daa-bc14-137b5ec5ec26 req-5721aac2-bcb5-4238-a0d4-2bb56e5af13a service nova] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Refreshing network info cache for port 4214bf6e-29f0-4d92-be44-44fccf7ab149 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2246.052506] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9105126-a504-4aaa-858c-1a997a0ef04f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2246.099997] env[62824]: DEBUG nova.compute.manager [req-af1186a1-89d9-4ca7-8e28-857186e1ce36 req-6732ec93-207d-48bc-97dc-3a4e2178e2ea service nova] [instance: 853f3098-881c-402c-8fd7-26540d1ecd88] Detach interface failed, port_id=03eb70c1-fc19-4e7c-874d-2f4fcd77fe6b, reason: Instance 853f3098-881c-402c-8fd7-26540d1ecd88 could not be found. {{(pid=62824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11482}} [ 2246.181942] env[62824]: DEBUG oslo_concurrency.lockutils [req-a503968a-8d79-4396-96de-bf6b99b836e3 req-a87d61d0-be9a-4d40-8823-234e55f751f0 service nova] Releasing lock "refresh_cache-204b866f-e9c5-441d-bf8e-77d324b39ecd" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2246.198308] env[62824]: DEBUG oslo_vmware.api [None req-4304e8a4-2b17-4cea-b613-125b1018f207 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2146641, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2246.207828] env[62824]: DEBUG oslo_vmware.api [None req-697b1a23-d14a-4d94-b716-8c4b25f6df27 tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Task: {'id': task-2146642, 'name': PowerOffVM_Task, 'duration_secs': 0.273103} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2246.208204] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-697b1a23-d14a-4d94-b716-8c4b25f6df27 tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2246.208959] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-697b1a23-d14a-4d94-b716-8c4b25f6df27 tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2246.209255] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a25a4510-3a36-4d67-ae88-ece492f7a93d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2246.215301] env[62824]: DEBUG oslo_vmware.api [None req-697b1a23-d14a-4d94-b716-8c4b25f6df27 tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Waiting for the task: (returnval){ [ 2246.215301] env[62824]: value = "task-2146645" [ 2246.215301] env[62824]: _type = "Task" [ 2246.215301] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2246.226180] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-697b1a23-d14a-4d94-b716-8c4b25f6df27 tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] VM already powered off {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 2246.226419] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-697b1a23-d14a-4d94-b716-8c4b25f6df27 tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Volume detach. Driver type: vmdk {{(pid=62824) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 2246.226631] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-697b1a23-d14a-4d94-b716-8c4b25f6df27 tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-438591', 'volume_id': '8d462122-e7ae-4257-8ce4-0a79a4acae13', 'name': 'volume-8d462122-e7ae-4257-8ce4-0a79a4acae13', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '2f315f9e-94e4-47ef-9503-ee92b59e5452', 'attached_at': '', 'detached_at': '', 'volume_id': '8d462122-e7ae-4257-8ce4-0a79a4acae13', 'serial': '8d462122-e7ae-4257-8ce4-0a79a4acae13'} {{(pid=62824) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 2246.228318] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dcc3f92-a66e-40df-b79d-90d6368a3c99 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2246.252158] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1db78a96-c526-4ddd-9756-b19c9caac21e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2246.265482] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76e705f9-ff90-4d05-908c-d553ad451fe8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2246.291655] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51520238-adca-41d4-93c8-6836af8d6401 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2246.310064] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-697b1a23-d14a-4d94-b716-8c4b25f6df27 tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] The volume has not been displaced from its original location: [datastore2] volume-8d462122-e7ae-4257-8ce4-0a79a4acae13/volume-8d462122-e7ae-4257-8ce4-0a79a4acae13.vmdk. No consolidation needed. {{(pid=62824) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 2246.316741] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-697b1a23-d14a-4d94-b716-8c4b25f6df27 tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Reconfiguring VM instance instance-0000003a to detach disk 2000 {{(pid=62824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 2246.319887] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-61c15b7d-5b9d-4da8-80e3-ee423e38a275 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2246.344396] env[62824]: DEBUG oslo_vmware.api [None req-697b1a23-d14a-4d94-b716-8c4b25f6df27 tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Waiting for the task: (returnval){ [ 2246.344396] env[62824]: value = "task-2146646" [ 2246.344396] env[62824]: _type = "Task" [ 2246.344396] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2246.355993] env[62824]: DEBUG oslo_vmware.api [None req-697b1a23-d14a-4d94-b716-8c4b25f6df27 tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Task: {'id': task-2146646, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2246.384397] env[62824]: DEBUG oslo_vmware.api [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146643, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.092233} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2246.384658] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2246.385618] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d522faf9-28c6-4f79-b9d4-5bd7841d5d87 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2246.422573] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Reconfiguring VM instance instance-0000005d to attach disk [datastore2] 0394d368-c846-4f3b-bfcf-2bc8e858052a/0394d368-c846-4f3b-bfcf-2bc8e858052a.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2246.423116] env[62824]: DEBUG oslo_vmware.api [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146644, 'name': Rename_Task, 'duration_secs': 0.326673} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2246.423822] env[62824]: DEBUG nova.network.neutron [req-0e333493-372e-4daa-bc14-137b5ec5ec26 req-5721aac2-bcb5-4238-a0d4-2bb56e5af13a service nova] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Updated VIF entry in instance network info cache for port 4214bf6e-29f0-4d92-be44-44fccf7ab149. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2246.424166] env[62824]: DEBUG nova.network.neutron [req-0e333493-372e-4daa-bc14-137b5ec5ec26 req-5721aac2-bcb5-4238-a0d4-2bb56e5af13a service nova] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Updating instance_info_cache with network_info: [{"id": "4214bf6e-29f0-4d92-be44-44fccf7ab149", "address": "fa:16:3e:a6:d8:c4", "network": {"id": "fc5d7128-fe9b-44ef-9d91-839190a97226", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-796405330-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35b7289f4746463e9b1cc47fe914a835", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0954fad3-d24d-496c-83e6-a09d3cb556fc", "external-id": "nsx-vlan-transportzone-216", "segmentation_id": 216, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4214bf6e-29", "ovs_interfaceid": "4214bf6e-29f0-4d92-be44-44fccf7ab149", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2246.425574] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-59a76674-c8d9-47b2-97d0-b3fb8cc6e2a5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2246.441109] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: feda7def-7788-4a60-a56a-9353a1475c54] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2246.442769] env[62824]: DEBUG oslo_concurrency.lockutils [req-0e333493-372e-4daa-bc14-137b5ec5ec26 req-5721aac2-bcb5-4238-a0d4-2bb56e5af13a service nova] Releasing lock "refresh_cache-71e9733b-c84b-4501-8faf-a487f089b498" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2246.442902] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-63e0ad57-8a14-45ce-a5ee-88888f89d68b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2246.458282] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7536d703-08c8-482a-8dac-7e0a3b2d05db tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquiring lock "interface-fc77c44a-180c-46ff-9690-9072c6213c91-9e4a9375-566d-4c66-8165-2d01b62d9a15" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2246.458746] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7536d703-08c8-482a-8dac-7e0a3b2d05db tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lock "interface-fc77c44a-180c-46ff-9690-9072c6213c91-9e4a9375-566d-4c66-8165-2d01b62d9a15" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2246.459015] env[62824]: DEBUG nova.objects.instance [None req-7536d703-08c8-482a-8dac-7e0a3b2d05db tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lazy-loading 'flavor' on Instance uuid fc77c44a-180c-46ff-9690-9072c6213c91 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2246.460252] env[62824]: DEBUG oslo_vmware.api [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2246.460252] env[62824]: value = "task-2146648" [ 2246.460252] env[62824]: _type = "Task" [ 2246.460252] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2246.460487] env[62824]: DEBUG oslo_vmware.api [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2246.460487] env[62824]: value = "task-2146647" [ 2246.460487] env[62824]: _type = "Task" [ 2246.460487] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2246.460691] env[62824]: DEBUG oslo_vmware.api [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]529c3195-ae95-762b-c925-6244c6fb1af5, 'name': SearchDatastore_Task, 'duration_secs': 0.069461} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2246.461659] env[62824]: DEBUG oslo_concurrency.lockutils [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2246.462294] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 204b866f-e9c5-441d-bf8e-77d324b39ecd] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2246.466186] env[62824]: DEBUG oslo_concurrency.lockutils [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2246.466186] env[62824]: DEBUG oslo_concurrency.lockutils [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2246.466186] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2246.476843] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5a23346a-f6b7-4608-86a9-790bfcb93913 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2246.478953] env[62824]: DEBUG oslo_vmware.api [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146640, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.748383} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2246.479252] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 71e9733b-c84b-4501-8faf-a487f089b498/71e9733b-c84b-4501-8faf-a487f089b498.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2246.479454] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2246.480601] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fd2bbb73-830a-4eae-9de1-55f62b708bb5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2246.490929] env[62824]: DEBUG oslo_vmware.api [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146648, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2246.494008] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2246.494220] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2246.494971] env[62824]: DEBUG oslo_vmware.api [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146647, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2246.495231] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-24114656-9ef0-4f77-85f2-cd695e579e41 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2246.500251] env[62824]: DEBUG oslo_vmware.api [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Waiting for the task: (returnval){ [ 2246.500251] env[62824]: value = "task-2146649" [ 2246.500251] env[62824]: _type = "Task" [ 2246.500251] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2246.505766] env[62824]: DEBUG oslo_vmware.api [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 2246.505766] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]524fca1e-aa09-84dc-3d5a-23f2ccd67910" [ 2246.505766] env[62824]: _type = "Task" [ 2246.505766] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2246.510296] env[62824]: INFO nova.compute.manager [-] [instance: 853f3098-881c-402c-8fd7-26540d1ecd88] Took 1.46 seconds to deallocate network for instance. [ 2246.514201] env[62824]: DEBUG oslo_vmware.api [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146649, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2246.522184] env[62824]: DEBUG oslo_vmware.api [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]524fca1e-aa09-84dc-3d5a-23f2ccd67910, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2246.694722] env[62824]: DEBUG oslo_vmware.api [None req-4304e8a4-2b17-4cea-b613-125b1018f207 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2146641, 'name': ReconfigVM_Task, 'duration_secs': 0.793935} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2246.695036] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-4304e8a4-2b17-4cea-b613-125b1018f207 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Reconfigured VM instance instance-00000039 to attach disk [datastore2] volume-10a6ba42-78a0-4e92-b936-f88a95560707/volume-10a6ba42-78a0-4e92-b936-f88a95560707.vmdk or device None with type thin {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2246.699817] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-80b2b4cd-4b8d-447a-ac17-1f85a4b15b90 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2246.715316] env[62824]: DEBUG oslo_vmware.api [None req-4304e8a4-2b17-4cea-b613-125b1018f207 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for the task: (returnval){ [ 2246.715316] env[62824]: value = "task-2146650" [ 2246.715316] env[62824]: _type = "Task" [ 2246.715316] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2246.723506] env[62824]: DEBUG oslo_vmware.api [None req-4304e8a4-2b17-4cea-b613-125b1018f207 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2146650, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2246.807107] env[62824]: DEBUG oslo_concurrency.lockutils [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Acquiring lock "c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2246.807432] env[62824]: DEBUG oslo_concurrency.lockutils [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Lock "c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2246.856061] env[62824]: DEBUG oslo_vmware.api [None req-697b1a23-d14a-4d94-b716-8c4b25f6df27 tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Task: {'id': task-2146646, 'name': ReconfigVM_Task, 'duration_secs': 0.253562} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2246.856555] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-697b1a23-d14a-4d94-b716-8c4b25f6df27 tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Reconfigured VM instance instance-0000003a to detach disk 2000 {{(pid=62824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 2246.862197] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d8154cea-3aa0-4a57-a0ad-3a7c136f94e6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2246.879049] env[62824]: DEBUG oslo_vmware.api [None req-697b1a23-d14a-4d94-b716-8c4b25f6df27 tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Waiting for the task: (returnval){ [ 2246.879049] env[62824]: value = "task-2146651" [ 2246.879049] env[62824]: _type = "Task" [ 2246.879049] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2246.887970] env[62824]: DEBUG oslo_vmware.api [None req-697b1a23-d14a-4d94-b716-8c4b25f6df27 tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Task: {'id': task-2146651, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2246.980378] env[62824]: DEBUG oslo_vmware.api [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146648, 'name': PowerOnVM_Task, 'duration_secs': 0.498564} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2246.983148] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: feda7def-7788-4a60-a56a-9353a1475c54] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2246.983354] env[62824]: INFO nova.compute.manager [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: feda7def-7788-4a60-a56a-9353a1475c54] Took 15.73 seconds to spawn the instance on the hypervisor. [ 2246.983526] env[62824]: DEBUG nova.compute.manager [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: feda7def-7788-4a60-a56a-9353a1475c54] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2246.983792] env[62824]: DEBUG oslo_vmware.api [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146647, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2246.984497] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa31b3bc-e6ac-4be8-bf2e-333c8b9bc313 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2247.009752] env[62824]: DEBUG oslo_vmware.api [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146649, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068109} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2247.015954] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2247.015954] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3f648bf-0204-4ecc-99f5-9e44892b71ce {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2247.020265] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d410e0b6-9b36-4f68-adc4-2fb7e46c625d tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2247.020265] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d410e0b6-9b36-4f68-adc4-2fb7e46c625d tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2247.020265] env[62824]: DEBUG nova.objects.instance [None req-d410e0b6-9b36-4f68-adc4-2fb7e46c625d tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Lazy-loading 'resources' on Instance uuid 853f3098-881c-402c-8fd7-26540d1ecd88 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2247.045173] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Reconfiguring VM instance instance-0000005c to attach disk [datastore2] 71e9733b-c84b-4501-8faf-a487f089b498/71e9733b-c84b-4501-8faf-a487f089b498.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2247.045587] env[62824]: DEBUG oslo_vmware.api [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]524fca1e-aa09-84dc-3d5a-23f2ccd67910, 'name': SearchDatastore_Task, 'duration_secs': 0.018901} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2247.047456] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c9b31d94-1ebb-45f1-b746-29d24c145371 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2247.065279] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d6ced92b-a6b4-4f4d-bff8-4fac5f515156 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2247.072267] env[62824]: DEBUG oslo_vmware.api [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 2247.072267] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]529b73cb-f640-1b57-2726-07612985ca8b" [ 2247.072267] env[62824]: _type = "Task" [ 2247.072267] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2247.076826] env[62824]: DEBUG oslo_vmware.api [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Waiting for the task: (returnval){ [ 2247.076826] env[62824]: value = "task-2146652" [ 2247.076826] env[62824]: _type = "Task" [ 2247.076826] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2247.083228] env[62824]: DEBUG oslo_vmware.api [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]529b73cb-f640-1b57-2726-07612985ca8b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2247.088179] env[62824]: DEBUG oslo_vmware.api [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146652, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2247.106954] env[62824]: DEBUG nova.objects.instance [None req-7536d703-08c8-482a-8dac-7e0a3b2d05db tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lazy-loading 'pci_requests' on Instance uuid fc77c44a-180c-46ff-9690-9072c6213c91 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2247.226929] env[62824]: DEBUG oslo_vmware.api [None req-4304e8a4-2b17-4cea-b613-125b1018f207 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2146650, 'name': ReconfigVM_Task, 'duration_secs': 0.166568} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2247.227928] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-4304e8a4-2b17-4cea-b613-125b1018f207 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-438727', 'volume_id': '10a6ba42-78a0-4e92-b936-f88a95560707', 'name': 'volume-10a6ba42-78a0-4e92-b936-f88a95560707', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'bbfcb2e3-9326-4548-b15b-e054cbfd192e', 'attached_at': '', 'detached_at': '', 'volume_id': '10a6ba42-78a0-4e92-b936-f88a95560707', 'serial': '10a6ba42-78a0-4e92-b936-f88a95560707'} {{(pid=62824) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 2247.309875] env[62824]: DEBUG nova.compute.manager [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] [instance: c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2247.388703] env[62824]: DEBUG oslo_vmware.api [None req-697b1a23-d14a-4d94-b716-8c4b25f6df27 tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Task: {'id': task-2146651, 'name': ReconfigVM_Task, 'duration_secs': 0.22211} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2247.389068] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-697b1a23-d14a-4d94-b716-8c4b25f6df27 tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-438591', 'volume_id': '8d462122-e7ae-4257-8ce4-0a79a4acae13', 'name': 'volume-8d462122-e7ae-4257-8ce4-0a79a4acae13', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '2f315f9e-94e4-47ef-9503-ee92b59e5452', 'attached_at': '', 'detached_at': '', 'volume_id': '8d462122-e7ae-4257-8ce4-0a79a4acae13', 'serial': '8d462122-e7ae-4257-8ce4-0a79a4acae13'} {{(pid=62824) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 2247.389331] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-697b1a23-d14a-4d94-b716-8c4b25f6df27 tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2247.390081] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34a5b153-5f6b-49f6-b3a2-cc22be72974c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2247.396497] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-697b1a23-d14a-4d94-b716-8c4b25f6df27 tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2247.396706] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e2c9b2bc-39d2-4051-b416-11c9ac69e535 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2247.476381] env[62824]: DEBUG oslo_vmware.api [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146647, 'name': ReconfigVM_Task, 'duration_secs': 0.751554} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2247.476675] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Reconfigured VM instance instance-0000005d to attach disk [datastore2] 0394d368-c846-4f3b-bfcf-2bc8e858052a/0394d368-c846-4f3b-bfcf-2bc8e858052a.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2247.477328] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f2b91542-2505-431b-98c4-45275702f65a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2247.484223] env[62824]: DEBUG oslo_vmware.api [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2247.484223] env[62824]: value = "task-2146654" [ 2247.484223] env[62824]: _type = "Task" [ 2247.484223] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2247.501024] env[62824]: DEBUG oslo_vmware.api [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146654, 'name': Rename_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2247.502675] env[62824]: INFO nova.compute.manager [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: feda7def-7788-4a60-a56a-9353a1475c54] Took 29.87 seconds to build instance. [ 2247.586532] env[62824]: DEBUG oslo_vmware.api [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]529b73cb-f640-1b57-2726-07612985ca8b, 'name': SearchDatastore_Task, 'duration_secs': 0.020979} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2247.586698] env[62824]: DEBUG oslo_concurrency.lockutils [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2247.586815] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 204b866f-e9c5-441d-bf8e-77d324b39ecd/204b866f-e9c5-441d-bf8e-77d324b39ecd.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2247.587078] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-aa3e8355-4a68-40c5-afeb-5493fbf090e8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2247.592471] env[62824]: DEBUG oslo_vmware.api [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146652, 'name': ReconfigVM_Task, 'duration_secs': 0.282821} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2247.593891] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Reconfigured VM instance instance-0000005c to attach disk [datastore2] 71e9733b-c84b-4501-8faf-a487f089b498/71e9733b-c84b-4501-8faf-a487f089b498.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2247.595655] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-43e05e55-b188-4ce1-9ad5-9bf6c38e649e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2247.598151] env[62824]: DEBUG oslo_vmware.api [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 2247.598151] env[62824]: value = "task-2146655" [ 2247.598151] env[62824]: _type = "Task" [ 2247.598151] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2247.603376] env[62824]: DEBUG oslo_vmware.api [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Waiting for the task: (returnval){ [ 2247.603376] env[62824]: value = "task-2146656" [ 2247.603376] env[62824]: _type = "Task" [ 2247.603376] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2247.609263] env[62824]: DEBUG oslo_vmware.api [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146655, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2247.611518] env[62824]: DEBUG nova.objects.base [None req-7536d703-08c8-482a-8dac-7e0a3b2d05db tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=62824) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 2247.611700] env[62824]: DEBUG nova.network.neutron [None req-7536d703-08c8-482a-8dac-7e0a3b2d05db tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2247.621645] env[62824]: DEBUG oslo_vmware.api [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146656, 'name': Rename_Task} progress is 10%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2247.702136] env[62824]: DEBUG nova.policy [None req-7536d703-08c8-482a-8dac-7e0a3b2d05db tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b989eb89498b45dab6604116870d2fa1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c987c5f844be42c99cc9d57a9ca84c31', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2247.827651] env[62824]: DEBUG oslo_concurrency.lockutils [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2247.841024] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7a5c0f5-01c3-4aa1-abad-ff0728b46f2c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2247.851640] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4ed8f00-2b12-4f17-99a7-1e1a9eb09d4d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2247.891296] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d03a1678-220d-4e23-9549-b8e24d055922 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2247.901032] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-304514ef-a7cc-4514-b784-3b0bcec5153f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2247.916575] env[62824]: DEBUG nova.compute.provider_tree [None req-d410e0b6-9b36-4f68-adc4-2fb7e46c625d tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2247.994068] env[62824]: DEBUG oslo_vmware.api [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146654, 'name': Rename_Task, 'duration_secs': 0.242288} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2247.994068] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2247.994318] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bcf10df5-9015-4f69-a85b-2e341a074020 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2248.001221] env[62824]: DEBUG oslo_vmware.api [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2248.001221] env[62824]: value = "task-2146657" [ 2248.001221] env[62824]: _type = "Task" [ 2248.001221] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2248.005377] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3e9928aa-4e7d-4f1c-a882-7c42101042d6 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "feda7def-7788-4a60-a56a-9353a1475c54" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.383s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2248.011427] env[62824]: DEBUG oslo_vmware.api [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146657, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2248.033207] env[62824]: DEBUG oslo_concurrency.lockutils [None req-61c785a4-c6ea-4ee0-8967-908719fa995a tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Acquiring lock "b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2248.033556] env[62824]: DEBUG oslo_concurrency.lockutils [None req-61c785a4-c6ea-4ee0-8967-908719fa995a tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Lock "b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2248.033806] env[62824]: DEBUG oslo_concurrency.lockutils [None req-61c785a4-c6ea-4ee0-8967-908719fa995a tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Acquiring lock "b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2248.034034] env[62824]: DEBUG oslo_concurrency.lockutils [None req-61c785a4-c6ea-4ee0-8967-908719fa995a tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Lock "b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2248.034246] env[62824]: DEBUG oslo_concurrency.lockutils [None req-61c785a4-c6ea-4ee0-8967-908719fa995a tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Lock "b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2248.036823] env[62824]: INFO nova.compute.manager [None req-61c785a4-c6ea-4ee0-8967-908719fa995a tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a] Terminating instance [ 2248.111773] env[62824]: DEBUG oslo_vmware.api [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146655, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2248.121774] env[62824]: DEBUG oslo_vmware.api [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146656, 'name': Rename_Task, 'duration_secs': 0.1623} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2248.122075] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2248.122367] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7e141774-591a-4608-b6f9-927bd4b2d3fb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2248.129225] env[62824]: DEBUG oslo_vmware.api [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Waiting for the task: (returnval){ [ 2248.129225] env[62824]: value = "task-2146658" [ 2248.129225] env[62824]: _type = "Task" [ 2248.129225] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2248.137527] env[62824]: DEBUG oslo_vmware.api [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146658, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2248.175586] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-697b1a23-d14a-4d94-b716-8c4b25f6df27 tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2248.175968] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-697b1a23-d14a-4d94-b716-8c4b25f6df27 tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Deleting contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2248.176237] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-697b1a23-d14a-4d94-b716-8c4b25f6df27 tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Deleting the datastore file [datastore2] 2f315f9e-94e4-47ef-9503-ee92b59e5452 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2248.176559] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4c43508d-5667-48d4-b046-4b11da442f21 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2248.188037] env[62824]: DEBUG oslo_vmware.api [None req-697b1a23-d14a-4d94-b716-8c4b25f6df27 tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Waiting for the task: (returnval){ [ 2248.188037] env[62824]: value = "task-2146659" [ 2248.188037] env[62824]: _type = "Task" [ 2248.188037] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2248.196759] env[62824]: DEBUG oslo_vmware.api [None req-697b1a23-d14a-4d94-b716-8c4b25f6df27 tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Task: {'id': task-2146659, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2248.284070] env[62824]: DEBUG nova.objects.instance [None req-4304e8a4-2b17-4cea-b613-125b1018f207 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Lazy-loading 'flavor' on Instance uuid bbfcb2e3-9326-4548-b15b-e054cbfd192e {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2248.420450] env[62824]: DEBUG nova.scheduler.client.report [None req-d410e0b6-9b36-4f68-adc4-2fb7e46c625d tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2248.511786] env[62824]: DEBUG oslo_vmware.api [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146657, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2248.541944] env[62824]: DEBUG nova.compute.manager [None req-61c785a4-c6ea-4ee0-8967-908719fa995a tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2248.542187] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-61c785a4-c6ea-4ee0-8967-908719fa995a tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2248.543102] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5648df9e-6e97-4299-9c85-dc7f52d52301 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2248.550733] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-61c785a4-c6ea-4ee0-8967-908719fa995a tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2248.550973] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-37359233-77af-4650-934d-f4cbaddf4b29 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2248.556590] env[62824]: DEBUG oslo_vmware.api [None req-61c785a4-c6ea-4ee0-8967-908719fa995a tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Waiting for the task: (returnval){ [ 2248.556590] env[62824]: value = "task-2146660" [ 2248.556590] env[62824]: _type = "Task" [ 2248.556590] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2248.564472] env[62824]: DEBUG oslo_vmware.api [None req-61c785a4-c6ea-4ee0-8967-908719fa995a tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Task: {'id': task-2146660, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2248.611548] env[62824]: DEBUG oslo_vmware.api [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146655, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.730693} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2248.611903] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 204b866f-e9c5-441d-bf8e-77d324b39ecd/204b866f-e9c5-441d-bf8e-77d324b39ecd.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2248.612248] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 204b866f-e9c5-441d-bf8e-77d324b39ecd] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2248.612592] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2b60a642-b8fb-4320-801e-d5ee50e08585 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2248.620124] env[62824]: DEBUG oslo_vmware.api [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 2248.620124] env[62824]: value = "task-2146661" [ 2248.620124] env[62824]: _type = "Task" [ 2248.620124] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2248.632965] env[62824]: DEBUG oslo_vmware.api [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146661, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2248.644383] env[62824]: DEBUG oslo_vmware.api [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146658, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2248.699167] env[62824]: DEBUG oslo_vmware.api [None req-697b1a23-d14a-4d94-b716-8c4b25f6df27 tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Task: {'id': task-2146659, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.226334} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2248.699505] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-697b1a23-d14a-4d94-b716-8c4b25f6df27 tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2248.699734] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-697b1a23-d14a-4d94-b716-8c4b25f6df27 tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Deleted contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2248.700141] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-697b1a23-d14a-4d94-b716-8c4b25f6df27 tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2248.762458] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-697b1a23-d14a-4d94-b716-8c4b25f6df27 tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Volume detach. Driver type: vmdk {{(pid=62824) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 2248.762826] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9a2f4d79-0020-46d5-b0b6-d0e718625b8f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2248.773134] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3d612eb-70df-45fc-a7bd-4adea81a6418 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2248.792981] env[62824]: DEBUG oslo_concurrency.lockutils [None req-4304e8a4-2b17-4cea-b613-125b1018f207 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Lock "bbfcb2e3-9326-4548-b15b-e054cbfd192e" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.309s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2248.815018] env[62824]: ERROR nova.compute.manager [None req-697b1a23-d14a-4d94-b716-8c4b25f6df27 tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Failed to detach volume 8d462122-e7ae-4257-8ce4-0a79a4acae13 from /dev/sda: nova.exception.InstanceNotFound: Instance 2f315f9e-94e4-47ef-9503-ee92b59e5452 could not be found. [ 2248.815018] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Traceback (most recent call last): [ 2248.815018] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] File "/opt/stack/nova/nova/compute/manager.py", line 4185, in _do_rebuild_instance [ 2248.815018] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] self.driver.rebuild(**kwargs) [ 2248.815018] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] File "/opt/stack/nova/nova/virt/driver.py", line 497, in rebuild [ 2248.815018] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] raise NotImplementedError() [ 2248.815018] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] NotImplementedError [ 2248.815018] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] [ 2248.815018] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] During handling of the above exception, another exception occurred: [ 2248.815018] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] [ 2248.815018] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Traceback (most recent call last): [ 2248.815018] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] File "/opt/stack/nova/nova/compute/manager.py", line 3608, in _detach_root_volume [ 2248.815018] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] self.driver.detach_volume(context, old_connection_info, [ 2248.815018] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 561, in detach_volume [ 2248.815018] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] return self._volumeops.detach_volume(connection_info, instance) [ 2248.815018] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 2248.815018] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] self._detach_volume_vmdk(connection_info, instance) [ 2248.815018] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 2248.815018] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 2248.815018] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1145, in get_vm_ref [ 2248.815018] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] stable_ref.fetch_moref(session) [ 2248.815018] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1136, in fetch_moref [ 2248.815018] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] raise exception.InstanceNotFound(instance_id=self._uuid) [ 2248.815018] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] nova.exception.InstanceNotFound: Instance 2f315f9e-94e4-47ef-9503-ee92b59e5452 could not be found. [ 2248.815018] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] [ 2248.928169] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d410e0b6-9b36-4f68-adc4-2fb7e46c625d tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.908s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2248.930114] env[62824]: DEBUG oslo_concurrency.lockutils [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.103s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2248.932265] env[62824]: INFO nova.compute.claims [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] [instance: c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2248.950745] env[62824]: INFO nova.scheduler.client.report [None req-d410e0b6-9b36-4f68-adc4-2fb7e46c625d tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Deleted allocations for instance 853f3098-881c-402c-8fd7-26540d1ecd88 [ 2248.955753] env[62824]: DEBUG nova.compute.utils [None req-697b1a23-d14a-4d94-b716-8c4b25f6df27 tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Build of instance 2f315f9e-94e4-47ef-9503-ee92b59e5452 aborted: Failed to rebuild volume backed instance. {{(pid=62824) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 2248.958426] env[62824]: ERROR nova.compute.manager [None req-697b1a23-d14a-4d94-b716-8c4b25f6df27 tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Setting instance vm_state to ERROR: nova.exception.BuildAbortException: Build of instance 2f315f9e-94e4-47ef-9503-ee92b59e5452 aborted: Failed to rebuild volume backed instance. [ 2248.958426] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Traceback (most recent call last): [ 2248.958426] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] File "/opt/stack/nova/nova/compute/manager.py", line 4185, in _do_rebuild_instance [ 2248.958426] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] self.driver.rebuild(**kwargs) [ 2248.958426] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] File "/opt/stack/nova/nova/virt/driver.py", line 497, in rebuild [ 2248.958426] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] raise NotImplementedError() [ 2248.958426] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] NotImplementedError [ 2248.958426] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] [ 2248.958426] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] During handling of the above exception, another exception occurred: [ 2248.958426] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] [ 2248.958426] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Traceback (most recent call last): [ 2248.958426] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] File "/opt/stack/nova/nova/compute/manager.py", line 3643, in _rebuild_volume_backed_instance [ 2248.958426] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] self._detach_root_volume(context, instance, root_bdm) [ 2248.958426] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] File "/opt/stack/nova/nova/compute/manager.py", line 3622, in _detach_root_volume [ 2248.958426] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] with excutils.save_and_reraise_exception(): [ 2248.958426] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 2248.958426] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] self.force_reraise() [ 2248.958426] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 2248.958426] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] raise self.value [ 2248.958426] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] File "/opt/stack/nova/nova/compute/manager.py", line 3608, in _detach_root_volume [ 2248.958426] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] self.driver.detach_volume(context, old_connection_info, [ 2248.958426] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 561, in detach_volume [ 2248.958426] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] return self._volumeops.detach_volume(connection_info, instance) [ 2248.958426] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 2248.958426] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] self._detach_volume_vmdk(connection_info, instance) [ 2248.958426] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 2248.958426] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 2248.958426] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1145, in get_vm_ref [ 2248.958426] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] stable_ref.fetch_moref(session) [ 2248.958426] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1136, in fetch_moref [ 2248.958426] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] raise exception.InstanceNotFound(instance_id=self._uuid) [ 2248.958426] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] nova.exception.InstanceNotFound: Instance 2f315f9e-94e4-47ef-9503-ee92b59e5452 could not be found. [ 2248.958426] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] [ 2248.958426] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] During handling of the above exception, another exception occurred: [ 2248.958426] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] [ 2248.958426] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Traceback (most recent call last): [ 2248.958426] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] File "/opt/stack/nova/nova/compute/manager.py", line 11382, in _error_out_instance_on_exception [ 2248.958426] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] yield [ 2248.958426] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] File "/opt/stack/nova/nova/compute/manager.py", line 3911, in rebuild_instance [ 2248.958426] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] self._do_rebuild_instance_with_claim( [ 2248.959639] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] File "/opt/stack/nova/nova/compute/manager.py", line 3997, in _do_rebuild_instance_with_claim [ 2248.959639] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] self._do_rebuild_instance( [ 2248.959639] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] File "/opt/stack/nova/nova/compute/manager.py", line 4189, in _do_rebuild_instance [ 2248.959639] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] self._rebuild_default_impl(**kwargs) [ 2248.959639] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] File "/opt/stack/nova/nova/compute/manager.py", line 3766, in _rebuild_default_impl [ 2248.959639] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] self._rebuild_volume_backed_instance( [ 2248.959639] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] File "/opt/stack/nova/nova/compute/manager.py", line 3658, in _rebuild_volume_backed_instance [ 2248.959639] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] raise exception.BuildAbortException( [ 2248.959639] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] nova.exception.BuildAbortException: Build of instance 2f315f9e-94e4-47ef-9503-ee92b59e5452 aborted: Failed to rebuild volume backed instance. [ 2248.959639] env[62824]: ERROR nova.compute.manager [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] [ 2249.012188] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquiring lock "3acdf86e-45e9-4353-a282-7272451389e7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2249.012436] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "3acdf86e-45e9-4353-a282-7272451389e7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2249.020297] env[62824]: DEBUG oslo_vmware.api [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146657, 'name': PowerOnVM_Task, 'duration_secs': 0.833088} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2249.020529] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2249.020718] env[62824]: INFO nova.compute.manager [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Took 11.25 seconds to spawn the instance on the hypervisor. [ 2249.021059] env[62824]: DEBUG nova.compute.manager [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2249.021707] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9098e670-31ac-44fc-aa9e-eb1c94780116 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2249.066697] env[62824]: DEBUG oslo_vmware.api [None req-61c785a4-c6ea-4ee0-8967-908719fa995a tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Task: {'id': task-2146660, 'name': PowerOffVM_Task, 'duration_secs': 0.241779} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2249.066907] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-61c785a4-c6ea-4ee0-8967-908719fa995a tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2249.067180] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-61c785a4-c6ea-4ee0-8967-908719fa995a tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2249.068015] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d96dd90b-ecce-404b-8054-68a9723599eb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2249.130521] env[62824]: DEBUG oslo_vmware.api [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146661, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07551} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2249.130739] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 204b866f-e9c5-441d-bf8e-77d324b39ecd] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2249.131507] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58d2bb29-8213-4556-bfb4-dea63741dae5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2249.141521] env[62824]: DEBUG oslo_vmware.api [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146658, 'name': PowerOnVM_Task, 'duration_secs': 0.576522} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2249.150751] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2249.150959] env[62824]: INFO nova.compute.manager [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Took 15.50 seconds to spawn the instance on the hypervisor. [ 2249.151160] env[62824]: DEBUG nova.compute.manager [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2249.160239] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 204b866f-e9c5-441d-bf8e-77d324b39ecd] Reconfiguring VM instance instance-0000005e to attach disk [datastore2] 204b866f-e9c5-441d-bf8e-77d324b39ecd/204b866f-e9c5-441d-bf8e-77d324b39ecd.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2249.160920] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-599908d9-8d36-41b0-93f6-64ffff0632cb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2249.163336] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-029428ee-8552-4aad-900e-a5e53c104cd7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2249.186157] env[62824]: DEBUG oslo_vmware.api [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 2249.186157] env[62824]: value = "task-2146663" [ 2249.186157] env[62824]: _type = "Task" [ 2249.186157] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2249.193674] env[62824]: DEBUG oslo_vmware.api [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146663, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2249.235130] env[62824]: INFO nova.compute.manager [None req-22a9cd95-c63b-4cbe-b831-e647907ecb3f tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Rescuing [ 2249.235405] env[62824]: DEBUG oslo_concurrency.lockutils [None req-22a9cd95-c63b-4cbe-b831-e647907ecb3f tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Acquiring lock "refresh_cache-bbfcb2e3-9326-4548-b15b-e054cbfd192e" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2249.235560] env[62824]: DEBUG oslo_concurrency.lockutils [None req-22a9cd95-c63b-4cbe-b831-e647907ecb3f tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Acquired lock "refresh_cache-bbfcb2e3-9326-4548-b15b-e054cbfd192e" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2249.235726] env[62824]: DEBUG nova.network.neutron [None req-22a9cd95-c63b-4cbe-b831-e647907ecb3f tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2249.251118] env[62824]: DEBUG nova.network.neutron [None req-7536d703-08c8-482a-8dac-7e0a3b2d05db tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Successfully updated port: 9e4a9375-566d-4c66-8165-2d01b62d9a15 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2249.253306] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-61c785a4-c6ea-4ee0-8967-908719fa995a tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2249.253306] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-61c785a4-c6ea-4ee0-8967-908719fa995a tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a] Deleting contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2249.253306] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-61c785a4-c6ea-4ee0-8967-908719fa995a tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Deleting the datastore file [datastore2] b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2249.253306] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-372aec8a-f431-43d4-9148-c4a125d12b59 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2249.259801] env[62824]: DEBUG oslo_vmware.api [None req-61c785a4-c6ea-4ee0-8967-908719fa995a tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Waiting for the task: (returnval){ [ 2249.259801] env[62824]: value = "task-2146664" [ 2249.259801] env[62824]: _type = "Task" [ 2249.259801] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2249.268836] env[62824]: DEBUG oslo_vmware.api [None req-61c785a4-c6ea-4ee0-8967-908719fa995a tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Task: {'id': task-2146664, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2249.380184] env[62824]: DEBUG nova.compute.manager [req-32e63aab-ab8e-4e3c-92e6-1262e1b95060 req-9174f3df-8cb1-4131-8bcf-6bc66b637380 service nova] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Received event network-vif-plugged-9e4a9375-566d-4c66-8165-2d01b62d9a15 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2249.380390] env[62824]: DEBUG oslo_concurrency.lockutils [req-32e63aab-ab8e-4e3c-92e6-1262e1b95060 req-9174f3df-8cb1-4131-8bcf-6bc66b637380 service nova] Acquiring lock "fc77c44a-180c-46ff-9690-9072c6213c91-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2249.380606] env[62824]: DEBUG oslo_concurrency.lockutils [req-32e63aab-ab8e-4e3c-92e6-1262e1b95060 req-9174f3df-8cb1-4131-8bcf-6bc66b637380 service nova] Lock "fc77c44a-180c-46ff-9690-9072c6213c91-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2249.380818] env[62824]: DEBUG oslo_concurrency.lockutils [req-32e63aab-ab8e-4e3c-92e6-1262e1b95060 req-9174f3df-8cb1-4131-8bcf-6bc66b637380 service nova] Lock "fc77c44a-180c-46ff-9690-9072c6213c91-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2249.380969] env[62824]: DEBUG nova.compute.manager [req-32e63aab-ab8e-4e3c-92e6-1262e1b95060 req-9174f3df-8cb1-4131-8bcf-6bc66b637380 service nova] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] No waiting events found dispatching network-vif-plugged-9e4a9375-566d-4c66-8165-2d01b62d9a15 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2249.381197] env[62824]: WARNING nova.compute.manager [req-32e63aab-ab8e-4e3c-92e6-1262e1b95060 req-9174f3df-8cb1-4131-8bcf-6bc66b637380 service nova] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Received unexpected event network-vif-plugged-9e4a9375-566d-4c66-8165-2d01b62d9a15 for instance with vm_state active and task_state None. [ 2249.460458] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d410e0b6-9b36-4f68-adc4-2fb7e46c625d tempest-ImagesTestJSON-1432969189 tempest-ImagesTestJSON-1432969189-project-member] Lock "853f3098-881c-402c-8fd7-26540d1ecd88" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.768s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2249.515164] env[62824]: DEBUG nova.compute.manager [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 3acdf86e-45e9-4353-a282-7272451389e7] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2249.539015] env[62824]: INFO nova.compute.manager [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Took 22.62 seconds to build instance. [ 2249.698221] env[62824]: INFO nova.compute.manager [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Took 31.99 seconds to build instance. [ 2249.703463] env[62824]: DEBUG oslo_vmware.api [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146663, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2249.754272] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7536d703-08c8-482a-8dac-7e0a3b2d05db tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquiring lock "refresh_cache-fc77c44a-180c-46ff-9690-9072c6213c91" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2249.754424] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7536d703-08c8-482a-8dac-7e0a3b2d05db tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquired lock "refresh_cache-fc77c44a-180c-46ff-9690-9072c6213c91" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2249.754639] env[62824]: DEBUG nova.network.neutron [None req-7536d703-08c8-482a-8dac-7e0a3b2d05db tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2249.771097] env[62824]: DEBUG oslo_vmware.api [None req-61c785a4-c6ea-4ee0-8967-908719fa995a tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Task: {'id': task-2146664, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.47032} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2249.771437] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-61c785a4-c6ea-4ee0-8967-908719fa995a tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2249.771632] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-61c785a4-c6ea-4ee0-8967-908719fa995a tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a] Deleted contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2249.771809] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-61c785a4-c6ea-4ee0-8967-908719fa995a tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2249.771986] env[62824]: INFO nova.compute.manager [None req-61c785a4-c6ea-4ee0-8967-908719fa995a tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a] Took 1.23 seconds to destroy the instance on the hypervisor. [ 2249.772244] env[62824]: DEBUG oslo.service.loopingcall [None req-61c785a4-c6ea-4ee0-8967-908719fa995a tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2249.772472] env[62824]: DEBUG nova.compute.manager [-] [instance: b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2249.772567] env[62824]: DEBUG nova.network.neutron [-] [instance: b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2249.995457] env[62824]: DEBUG nova.network.neutron [None req-22a9cd95-c63b-4cbe-b831-e647907ecb3f tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Updating instance_info_cache with network_info: [{"id": "d85a0acc-fed6-4797-9f81-2aafa5bb6967", "address": "fa:16:3e:d8:c8:ab", "network": {"id": "a410f4f3-d072-4034-b797-8210e1a79093", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1482574391-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.211", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e10208040df34eb5adfe8dcbc76043d5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f4399275-8e92-4448-be9e-d4984e93e89c", "external-id": "nsx-vlan-transportzone-192", "segmentation_id": 192, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd85a0acc-fe", "ovs_interfaceid": "d85a0acc-fed6-4797-9f81-2aafa5bb6967", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2250.041041] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2250.043672] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dfa94fa8-ad5e-444a-9010-64e7db107623 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "0394d368-c846-4f3b-bfcf-2bc8e858052a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.141s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2250.199883] env[62824]: DEBUG oslo_concurrency.lockutils [None req-08f4ec4c-d4de-4f5f-89e8-ba52838cdaa1 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Lock "71e9733b-c84b-4501-8faf-a487f089b498" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 33.505s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2250.203448] env[62824]: DEBUG oslo_vmware.api [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146663, 'name': ReconfigVM_Task, 'duration_secs': 0.810863} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2250.204566] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 204b866f-e9c5-441d-bf8e-77d324b39ecd] Reconfigured VM instance instance-0000005e to attach disk [datastore2] 204b866f-e9c5-441d-bf8e-77d324b39ecd/204b866f-e9c5-441d-bf8e-77d324b39ecd.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2250.205832] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d6d08df-2caf-4eb5-a018-c4bd340f4d54 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2250.208269] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5bcd3ded-1bfd-4b62-b836-34a865c9ae17 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2250.215022] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7af9495e-40e7-43dd-9878-a30e69ae0e95 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2250.219136] env[62824]: DEBUG oslo_vmware.api [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 2250.219136] env[62824]: value = "task-2146665" [ 2250.219136] env[62824]: _type = "Task" [ 2250.219136] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2250.249011] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1aafd06-2314-4f27-a7ff-8bc530f8c688 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2250.255106] env[62824]: DEBUG oslo_vmware.api [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146665, 'name': Rename_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2250.261889] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9510b01f-8b79-4307-b99c-4071fb8af77d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2250.276091] env[62824]: DEBUG nova.compute.provider_tree [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2250.305824] env[62824]: WARNING nova.network.neutron [None req-7536d703-08c8-482a-8dac-7e0a3b2d05db tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] aa1c09e5-8d72-43ad-a903-a0f5e711da80 already exists in list: networks containing: ['aa1c09e5-8d72-43ad-a903-a0f5e711da80']. ignoring it [ 2250.305864] env[62824]: WARNING nova.network.neutron [None req-7536d703-08c8-482a-8dac-7e0a3b2d05db tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] aa1c09e5-8d72-43ad-a903-a0f5e711da80 already exists in list: networks containing: ['aa1c09e5-8d72-43ad-a903-a0f5e711da80']. ignoring it [ 2250.341380] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Acquiring lock "815bd7ae-f461-4c74-9401-7c5fe43679ee" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2250.341695] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Lock "815bd7ae-f461-4c74-9401-7c5fe43679ee" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2250.503396] env[62824]: DEBUG oslo_concurrency.lockutils [None req-22a9cd95-c63b-4cbe-b831-e647907ecb3f tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Releasing lock "refresh_cache-bbfcb2e3-9326-4548-b15b-e054cbfd192e" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2250.633097] env[62824]: DEBUG nova.network.neutron [-] [instance: b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2250.729415] env[62824]: DEBUG oslo_vmware.api [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146665, 'name': Rename_Task, 'duration_secs': 0.29443} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2250.729775] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 204b866f-e9c5-441d-bf8e-77d324b39ecd] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2250.729911] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-50e16209-0dc4-4a43-b4b8-7faf15ea80fa {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2250.736209] env[62824]: DEBUG oslo_vmware.api [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 2250.736209] env[62824]: value = "task-2146666" [ 2250.736209] env[62824]: _type = "Task" [ 2250.736209] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2250.745439] env[62824]: DEBUG oslo_vmware.api [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146666, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2250.757124] env[62824]: DEBUG nova.compute.manager [req-10dc221e-b774-46d8-81f3-5932bb4067c2 req-9fb07704-5feb-4f98-a669-2520069dd077 service nova] [instance: b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a] Received event network-vif-deleted-a6e03f6a-b693-48dc-a8a5-b853d424639f {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2250.779300] env[62824]: DEBUG nova.scheduler.client.report [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2250.844515] env[62824]: DEBUG nova.compute.manager [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 815bd7ae-f461-4c74-9401-7c5fe43679ee] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2250.909159] env[62824]: DEBUG nova.network.neutron [None req-7536d703-08c8-482a-8dac-7e0a3b2d05db tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Updating instance_info_cache with network_info: [{"id": "e285a842-d54b-4614-84ea-239e34f5b41f", "address": "fa:16:3e:9d:17:cd", "network": {"id": "aa1c09e5-8d72-43ad-a903-a0f5e711da80", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1643616572-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c987c5f844be42c99cc9d57a9ca84c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9bc2632-36f9-4912-8782-8bbb789f909d", "external-id": "nsx-vlan-transportzone-897", "segmentation_id": 897, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape285a842-d5", "ovs_interfaceid": "e285a842-d54b-4614-84ea-239e34f5b41f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "2b6381b4-f3c7-4e22-b698-de190c8199ee", "address": "fa:16:3e:f4:8a:c1", "network": {"id": "aa1c09e5-8d72-43ad-a903-a0f5e711da80", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1643616572-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c987c5f844be42c99cc9d57a9ca84c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9bc2632-36f9-4912-8782-8bbb789f909d", "external-id": "nsx-vlan-transportzone-897", "segmentation_id": 897, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b6381b4-f3", "ovs_interfaceid": "2b6381b4-f3c7-4e22-b698-de190c8199ee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "9e4a9375-566d-4c66-8165-2d01b62d9a15", "address": "fa:16:3e:31:14:52", "network": {"id": "aa1c09e5-8d72-43ad-a903-a0f5e711da80", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1643616572-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c987c5f844be42c99cc9d57a9ca84c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9bc2632-36f9-4912-8782-8bbb789f909d", "external-id": "nsx-vlan-transportzone-897", "segmentation_id": 897, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9e4a9375-56", "ovs_interfaceid": "9e4a9375-566d-4c66-8165-2d01b62d9a15", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2250.973571] env[62824]: DEBUG oslo_concurrency.lockutils [None req-697b1a23-d14a-4d94-b716-8c4b25f6df27 tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2251.136526] env[62824]: INFO nova.compute.manager [-] [instance: b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a] Took 1.36 seconds to deallocate network for instance. [ 2251.247103] env[62824]: DEBUG oslo_vmware.api [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146666, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2251.284417] env[62824]: DEBUG oslo_concurrency.lockutils [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.354s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2251.285086] env[62824]: DEBUG nova.compute.manager [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] [instance: c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2251.288124] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.247s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2251.289630] env[62824]: INFO nova.compute.claims [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 3acdf86e-45e9-4353-a282-7272451389e7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2251.368597] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2251.412173] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7536d703-08c8-482a-8dac-7e0a3b2d05db tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Releasing lock "refresh_cache-fc77c44a-180c-46ff-9690-9072c6213c91" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2251.412879] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7536d703-08c8-482a-8dac-7e0a3b2d05db tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquiring lock "fc77c44a-180c-46ff-9690-9072c6213c91" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2251.413070] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7536d703-08c8-482a-8dac-7e0a3b2d05db tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquired lock "fc77c44a-180c-46ff-9690-9072c6213c91" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2251.413973] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b06e7b8-d24b-4e41-9190-417d269044f4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2251.418121] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c7c3ebd6-2b83-4c36-8295-ff7e1039c04f tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquiring lock "0394d368-c846-4f3b-bfcf-2bc8e858052a" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2251.418422] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c7c3ebd6-2b83-4c36-8295-ff7e1039c04f tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "0394d368-c846-4f3b-bfcf-2bc8e858052a" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2251.418619] env[62824]: DEBUG nova.compute.manager [None req-c7c3ebd6-2b83-4c36-8295-ff7e1039c04f tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2251.419466] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f774b05-d5d0-46e7-a20e-a38523ac0b20 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2251.439743] env[62824]: DEBUG nova.virt.hardware [None req-7536d703-08c8-482a-8dac-7e0a3b2d05db tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2251.440135] env[62824]: DEBUG nova.virt.hardware [None req-7536d703-08c8-482a-8dac-7e0a3b2d05db tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2251.440373] env[62824]: DEBUG nova.virt.hardware [None req-7536d703-08c8-482a-8dac-7e0a3b2d05db tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2251.440627] env[62824]: DEBUG nova.virt.hardware [None req-7536d703-08c8-482a-8dac-7e0a3b2d05db tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2251.440814] env[62824]: DEBUG nova.virt.hardware [None req-7536d703-08c8-482a-8dac-7e0a3b2d05db tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2251.441241] env[62824]: DEBUG nova.virt.hardware [None req-7536d703-08c8-482a-8dac-7e0a3b2d05db tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2251.441544] env[62824]: DEBUG nova.virt.hardware [None req-7536d703-08c8-482a-8dac-7e0a3b2d05db tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2251.441735] env[62824]: DEBUG nova.virt.hardware [None req-7536d703-08c8-482a-8dac-7e0a3b2d05db tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2251.441963] env[62824]: DEBUG nova.virt.hardware [None req-7536d703-08c8-482a-8dac-7e0a3b2d05db tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2251.442228] env[62824]: DEBUG nova.virt.hardware [None req-7536d703-08c8-482a-8dac-7e0a3b2d05db tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2251.442466] env[62824]: DEBUG nova.virt.hardware [None req-7536d703-08c8-482a-8dac-7e0a3b2d05db tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2251.449736] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-7536d703-08c8-482a-8dac-7e0a3b2d05db tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Reconfiguring VM to attach interface {{(pid=62824) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 2251.451780] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e01ae380-2b00-4457-8dbf-37b9d07013f9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2251.464520] env[62824]: DEBUG nova.compute.manager [None req-c7c3ebd6-2b83-4c36-8295-ff7e1039c04f tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62824) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3404}} [ 2251.464987] env[62824]: DEBUG nova.objects.instance [None req-c7c3ebd6-2b83-4c36-8295-ff7e1039c04f tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lazy-loading 'flavor' on Instance uuid 0394d368-c846-4f3b-bfcf-2bc8e858052a {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2251.474027] env[62824]: DEBUG oslo_vmware.api [None req-7536d703-08c8-482a-8dac-7e0a3b2d05db tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Waiting for the task: (returnval){ [ 2251.474027] env[62824]: value = "task-2146667" [ 2251.474027] env[62824]: _type = "Task" [ 2251.474027] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2251.482634] env[62824]: DEBUG oslo_vmware.api [None req-7536d703-08c8-482a-8dac-7e0a3b2d05db tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146667, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2251.499850] env[62824]: DEBUG nova.compute.manager [req-cefa78f4-ea18-47f2-8c36-c3fdbbdeebce req-16b40a80-a3be-4729-9a51-9ff550332148 service nova] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Received event network-changed-9e4a9375-566d-4c66-8165-2d01b62d9a15 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2251.500269] env[62824]: DEBUG nova.compute.manager [req-cefa78f4-ea18-47f2-8c36-c3fdbbdeebce req-16b40a80-a3be-4729-9a51-9ff550332148 service nova] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Refreshing instance network info cache due to event network-changed-9e4a9375-566d-4c66-8165-2d01b62d9a15. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2251.500463] env[62824]: DEBUG oslo_concurrency.lockutils [req-cefa78f4-ea18-47f2-8c36-c3fdbbdeebce req-16b40a80-a3be-4729-9a51-9ff550332148 service nova] Acquiring lock "refresh_cache-fc77c44a-180c-46ff-9690-9072c6213c91" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2251.500463] env[62824]: DEBUG oslo_concurrency.lockutils [req-cefa78f4-ea18-47f2-8c36-c3fdbbdeebce req-16b40a80-a3be-4729-9a51-9ff550332148 service nova] Acquired lock "refresh_cache-fc77c44a-180c-46ff-9690-9072c6213c91" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2251.500666] env[62824]: DEBUG nova.network.neutron [req-cefa78f4-ea18-47f2-8c36-c3fdbbdeebce req-16b40a80-a3be-4729-9a51-9ff550332148 service nova] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Refreshing network info cache for port 9e4a9375-566d-4c66-8165-2d01b62d9a15 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2251.642528] env[62824]: DEBUG oslo_concurrency.lockutils [None req-61c785a4-c6ea-4ee0-8967-908719fa995a tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2251.749989] env[62824]: DEBUG oslo_vmware.api [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146666, 'name': PowerOnVM_Task, 'duration_secs': 0.735471} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2251.750374] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 204b866f-e9c5-441d-bf8e-77d324b39ecd] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2251.750675] env[62824]: INFO nova.compute.manager [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 204b866f-e9c5-441d-bf8e-77d324b39ecd] Took 9.88 seconds to spawn the instance on the hypervisor. [ 2251.750920] env[62824]: DEBUG nova.compute.manager [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 204b866f-e9c5-441d-bf8e-77d324b39ecd] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2251.752027] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a172d81-5c8b-49df-94ca-63a2a0372a37 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2251.795023] env[62824]: DEBUG nova.compute.utils [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2251.800362] env[62824]: DEBUG nova.compute.manager [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] [instance: c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2251.800362] env[62824]: DEBUG nova.network.neutron [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] [instance: c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2251.983821] env[62824]: DEBUG oslo_vmware.api [None req-7536d703-08c8-482a-8dac-7e0a3b2d05db tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146667, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2252.039369] env[62824]: DEBUG nova.policy [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3c33344bd0fe4a088801138a887d912f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '58f9886b55b64e90af5d3b646429a232', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2252.043715] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-22a9cd95-c63b-4cbe-b831-e647907ecb3f tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2252.043936] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e2d9a13b-501f-4817-8335-7df4082caafd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2252.051176] env[62824]: DEBUG oslo_vmware.api [None req-22a9cd95-c63b-4cbe-b831-e647907ecb3f tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for the task: (returnval){ [ 2252.051176] env[62824]: value = "task-2146668" [ 2252.051176] env[62824]: _type = "Task" [ 2252.051176] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2252.062534] env[62824]: DEBUG oslo_vmware.api [None req-22a9cd95-c63b-4cbe-b831-e647907ecb3f tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2146668, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2252.239516] env[62824]: DEBUG nova.network.neutron [req-cefa78f4-ea18-47f2-8c36-c3fdbbdeebce req-16b40a80-a3be-4729-9a51-9ff550332148 service nova] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Updated VIF entry in instance network info cache for port 9e4a9375-566d-4c66-8165-2d01b62d9a15. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2252.240050] env[62824]: DEBUG nova.network.neutron [req-cefa78f4-ea18-47f2-8c36-c3fdbbdeebce req-16b40a80-a3be-4729-9a51-9ff550332148 service nova] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Updating instance_info_cache with network_info: [{"id": "e285a842-d54b-4614-84ea-239e34f5b41f", "address": "fa:16:3e:9d:17:cd", "network": {"id": "aa1c09e5-8d72-43ad-a903-a0f5e711da80", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1643616572-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c987c5f844be42c99cc9d57a9ca84c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9bc2632-36f9-4912-8782-8bbb789f909d", "external-id": "nsx-vlan-transportzone-897", "segmentation_id": 897, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape285a842-d5", "ovs_interfaceid": "e285a842-d54b-4614-84ea-239e34f5b41f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "2b6381b4-f3c7-4e22-b698-de190c8199ee", "address": "fa:16:3e:f4:8a:c1", "network": {"id": "aa1c09e5-8d72-43ad-a903-a0f5e711da80", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1643616572-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c987c5f844be42c99cc9d57a9ca84c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9bc2632-36f9-4912-8782-8bbb789f909d", "external-id": "nsx-vlan-transportzone-897", "segmentation_id": 897, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b6381b4-f3", "ovs_interfaceid": "2b6381b4-f3c7-4e22-b698-de190c8199ee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "9e4a9375-566d-4c66-8165-2d01b62d9a15", "address": "fa:16:3e:31:14:52", "network": {"id": "aa1c09e5-8d72-43ad-a903-a0f5e711da80", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1643616572-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c987c5f844be42c99cc9d57a9ca84c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9bc2632-36f9-4912-8782-8bbb789f909d", "external-id": "nsx-vlan-transportzone-897", "segmentation_id": 897, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9e4a9375-56", "ovs_interfaceid": "9e4a9375-566d-4c66-8165-2d01b62d9a15", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2252.274913] env[62824]: INFO nova.compute.manager [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 204b866f-e9c5-441d-bf8e-77d324b39ecd] Took 19.79 seconds to build instance. [ 2252.301184] env[62824]: DEBUG nova.compute.manager [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] [instance: c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2252.475713] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7c3ebd6-2b83-4c36-8295-ff7e1039c04f tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2252.476129] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8ba8fcaf-b631-4b8a-a886-3316ca18cb3b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2252.491426] env[62824]: DEBUG oslo_vmware.api [None req-c7c3ebd6-2b83-4c36-8295-ff7e1039c04f tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2252.491426] env[62824]: value = "task-2146669" [ 2252.491426] env[62824]: _type = "Task" [ 2252.491426] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2252.491692] env[62824]: DEBUG oslo_vmware.api [None req-7536d703-08c8-482a-8dac-7e0a3b2d05db tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146667, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2252.504943] env[62824]: DEBUG oslo_vmware.api [None req-c7c3ebd6-2b83-4c36-8295-ff7e1039c04f tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146669, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2252.567438] env[62824]: DEBUG oslo_vmware.api [None req-22a9cd95-c63b-4cbe-b831-e647907ecb3f tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2146668, 'name': PowerOffVM_Task, 'duration_secs': 0.305457} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2252.567714] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-22a9cd95-c63b-4cbe-b831-e647907ecb3f tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2252.574020] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-622d790d-85fa-4fea-94eb-7e777483f5d3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2252.598025] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47a3fc6c-31ed-41a6-835b-6248ab2c6b50 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2252.643127] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5baaf861-a875-4e3c-9fec-11b12e7adc0b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2252.656905] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f85e51f-913f-4913-b183-94e5b3759455 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2252.662380] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-22a9cd95-c63b-4cbe-b831-e647907ecb3f tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2252.663114] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0869963d-f07d-4d6b-8881-232fe98334d4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2252.695172] env[62824]: DEBUG nova.network.neutron [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] [instance: c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9] Successfully created port: 78cd208b-52d0-44fb-af4b-7012963b8b85 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2252.697770] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95e29982-7e23-4c15-8d78-060dcf191591 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2252.700559] env[62824]: DEBUG oslo_vmware.api [None req-22a9cd95-c63b-4cbe-b831-e647907ecb3f tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for the task: (returnval){ [ 2252.700559] env[62824]: value = "task-2146670" [ 2252.700559] env[62824]: _type = "Task" [ 2252.700559] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2252.710132] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c25505ae-6f62-4cba-9554-71d86d61d0db {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2252.719421] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-22a9cd95-c63b-4cbe-b831-e647907ecb3f tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] VM already powered off {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 2252.719788] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-22a9cd95-c63b-4cbe-b831-e647907ecb3f tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2252.720061] env[62824]: DEBUG oslo_concurrency.lockutils [None req-22a9cd95-c63b-4cbe-b831-e647907ecb3f tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2252.720198] env[62824]: DEBUG oslo_concurrency.lockutils [None req-22a9cd95-c63b-4cbe-b831-e647907ecb3f tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2252.720752] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-22a9cd95-c63b-4cbe-b831-e647907ecb3f tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2252.728560] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b624848e-fbfb-41fe-a3cf-f1f36e8ac12f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2252.730883] env[62824]: DEBUG nova.compute.provider_tree [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2252.739312] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-22a9cd95-c63b-4cbe-b831-e647907ecb3f tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2252.739650] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-22a9cd95-c63b-4cbe-b831-e647907ecb3f tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2252.740896] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d3044ed3-f6cf-4d8c-acfe-0610b4f0c44e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2252.743754] env[62824]: DEBUG oslo_concurrency.lockutils [req-cefa78f4-ea18-47f2-8c36-c3fdbbdeebce req-16b40a80-a3be-4729-9a51-9ff550332148 service nova] Releasing lock "refresh_cache-fc77c44a-180c-46ff-9690-9072c6213c91" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2252.748728] env[62824]: DEBUG oslo_vmware.api [None req-22a9cd95-c63b-4cbe-b831-e647907ecb3f tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for the task: (returnval){ [ 2252.748728] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]526da1c9-b82e-3b89-57e8-e15f984be9e1" [ 2252.748728] env[62824]: _type = "Task" [ 2252.748728] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2252.758911] env[62824]: DEBUG oslo_vmware.api [None req-22a9cd95-c63b-4cbe-b831-e647907ecb3f tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]526da1c9-b82e-3b89-57e8-e15f984be9e1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2252.777265] env[62824]: DEBUG oslo_concurrency.lockutils [None req-81999bc8-a159-4324-84d3-6737d8a373f0 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "204b866f-e9c5-441d-bf8e-77d324b39ecd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.304s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2252.986696] env[62824]: DEBUG oslo_vmware.api [None req-7536d703-08c8-482a-8dac-7e0a3b2d05db tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146667, 'name': ReconfigVM_Task, 'duration_secs': 1.169509} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2252.987315] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7536d703-08c8-482a-8dac-7e0a3b2d05db tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Releasing lock "fc77c44a-180c-46ff-9690-9072c6213c91" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2252.987533] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-7536d703-08c8-482a-8dac-7e0a3b2d05db tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Reconfigured VM to attach interface {{(pid=62824) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 2253.003605] env[62824]: DEBUG oslo_vmware.api [None req-c7c3ebd6-2b83-4c36-8295-ff7e1039c04f tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146669, 'name': PowerOffVM_Task, 'duration_secs': 0.221342} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2253.004700] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7c3ebd6-2b83-4c36-8295-ff7e1039c04f tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2253.004854] env[62824]: DEBUG nova.compute.manager [None req-c7c3ebd6-2b83-4c36-8295-ff7e1039c04f tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2253.005657] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8171b2d-e1df-4e19-9a5d-57a762251dc3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2253.134878] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d6968317-d8b9-4e00-a6a9-ee3a22a078b2 tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Acquiring lock "2f315f9e-94e4-47ef-9503-ee92b59e5452" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2253.135175] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d6968317-d8b9-4e00-a6a9-ee3a22a078b2 tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Lock "2f315f9e-94e4-47ef-9503-ee92b59e5452" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2253.135395] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d6968317-d8b9-4e00-a6a9-ee3a22a078b2 tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Acquiring lock "2f315f9e-94e4-47ef-9503-ee92b59e5452-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2253.135578] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d6968317-d8b9-4e00-a6a9-ee3a22a078b2 tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Lock "2f315f9e-94e4-47ef-9503-ee92b59e5452-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2253.135809] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d6968317-d8b9-4e00-a6a9-ee3a22a078b2 tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Lock "2f315f9e-94e4-47ef-9503-ee92b59e5452-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2253.137965] env[62824]: INFO nova.compute.manager [None req-d6968317-d8b9-4e00-a6a9-ee3a22a078b2 tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Terminating instance [ 2253.234570] env[62824]: DEBUG nova.scheduler.client.report [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2253.259994] env[62824]: DEBUG oslo_vmware.api [None req-22a9cd95-c63b-4cbe-b831-e647907ecb3f tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]526da1c9-b82e-3b89-57e8-e15f984be9e1, 'name': SearchDatastore_Task, 'duration_secs': 0.033997} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2253.260815] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a6ca7508-6ede-43c2-a98b-e1e75e8c5714 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2253.266509] env[62824]: DEBUG oslo_vmware.api [None req-22a9cd95-c63b-4cbe-b831-e647907ecb3f tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for the task: (returnval){ [ 2253.266509] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5201b730-baf2-449e-ef8d-8991752b16d7" [ 2253.266509] env[62824]: _type = "Task" [ 2253.266509] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2253.274940] env[62824]: DEBUG oslo_vmware.api [None req-22a9cd95-c63b-4cbe-b831-e647907ecb3f tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5201b730-baf2-449e-ef8d-8991752b16d7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2253.313773] env[62824]: DEBUG nova.compute.manager [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] [instance: c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2253.347896] env[62824]: DEBUG nova.virt.hardware [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2253.348141] env[62824]: DEBUG nova.virt.hardware [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2253.348312] env[62824]: DEBUG nova.virt.hardware [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2253.348498] env[62824]: DEBUG nova.virt.hardware [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2253.348645] env[62824]: DEBUG nova.virt.hardware [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2253.348793] env[62824]: DEBUG nova.virt.hardware [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2253.349014] env[62824]: DEBUG nova.virt.hardware [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2253.352937] env[62824]: DEBUG nova.virt.hardware [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2253.353173] env[62824]: DEBUG nova.virt.hardware [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2253.353366] env[62824]: DEBUG nova.virt.hardware [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2253.353547] env[62824]: DEBUG nova.virt.hardware [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2253.355770] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3595abd7-57b8-4248-a03a-d2b95a3430ee {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2253.363035] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0696a965-c0f8-434d-959c-585475b09e33 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2253.492822] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7536d703-08c8-482a-8dac-7e0a3b2d05db tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lock "interface-fc77c44a-180c-46ff-9690-9072c6213c91-9e4a9375-566d-4c66-8165-2d01b62d9a15" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.034s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2253.518330] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c7c3ebd6-2b83-4c36-8295-ff7e1039c04f tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "0394d368-c846-4f3b-bfcf-2bc8e858052a" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.100s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2253.644484] env[62824]: DEBUG nova.compute.manager [None req-d6968317-d8b9-4e00-a6a9-ee3a22a078b2 tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2253.645514] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-db8bc9f7-fc60-4dc3-b77e-d12f0d3d9ca8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2253.654701] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5370c5c3-9a49-4dce-b33f-3580f5f3ee35 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2253.691911] env[62824]: WARNING nova.virt.vmwareapi.driver [None req-d6968317-d8b9-4e00-a6a9-ee3a22a078b2 tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 2f315f9e-94e4-47ef-9503-ee92b59e5452 could not be found. [ 2253.692859] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-d6968317-d8b9-4e00-a6a9-ee3a22a078b2 tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2253.692859] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b515bf4f-f1d3-4d96-a4f9-8d1073f69093 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2253.703573] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-257b43de-060c-4f0f-a6b3-28dcfa3ea52f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2253.738983] env[62824]: WARNING nova.virt.vmwareapi.vmops [None req-d6968317-d8b9-4e00-a6a9-ee3a22a078b2 tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2f315f9e-94e4-47ef-9503-ee92b59e5452 could not be found. [ 2253.739223] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-d6968317-d8b9-4e00-a6a9-ee3a22a078b2 tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2253.739477] env[62824]: INFO nova.compute.manager [None req-d6968317-d8b9-4e00-a6a9-ee3a22a078b2 tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Took 0.09 seconds to destroy the instance on the hypervisor. [ 2253.739799] env[62824]: DEBUG oslo.service.loopingcall [None req-d6968317-d8b9-4e00-a6a9-ee3a22a078b2 tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2253.740716] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.453s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2253.741200] env[62824]: DEBUG nova.compute.manager [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 3acdf86e-45e9-4353-a282-7272451389e7] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2253.743714] env[62824]: DEBUG nova.compute.manager [-] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2253.743819] env[62824]: DEBUG nova.network.neutron [-] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2253.745826] env[62824]: DEBUG oslo_concurrency.lockutils [None req-697b1a23-d14a-4d94-b716-8c4b25f6df27 tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.772s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2253.781122] env[62824]: DEBUG oslo_vmware.api [None req-22a9cd95-c63b-4cbe-b831-e647907ecb3f tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5201b730-baf2-449e-ef8d-8991752b16d7, 'name': SearchDatastore_Task, 'duration_secs': 0.009942} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2253.782176] env[62824]: DEBUG oslo_concurrency.lockutils [None req-22a9cd95-c63b-4cbe-b831-e647907ecb3f tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2253.782451] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-22a9cd95-c63b-4cbe-b831-e647907ecb3f tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] bbfcb2e3-9326-4548-b15b-e054cbfd192e/9e2a7d30-212d-4ab8-9606-c5c6d52629e8-rescue.vmdk. {{(pid=62824) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 2253.782794] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-84f54015-e2a4-4b35-8098-e1c4ffe43b6b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2253.793524] env[62824]: DEBUG oslo_vmware.api [None req-22a9cd95-c63b-4cbe-b831-e647907ecb3f tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for the task: (returnval){ [ 2253.793524] env[62824]: value = "task-2146671" [ 2253.793524] env[62824]: _type = "Task" [ 2253.793524] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2253.805997] env[62824]: DEBUG oslo_vmware.api [None req-22a9cd95-c63b-4cbe-b831-e647907ecb3f tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2146671, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2254.002914] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bb7e3c42-05f1-407f-af38-2c6c282611aa tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquiring lock "204b866f-e9c5-441d-bf8e-77d324b39ecd" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2254.003181] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bb7e3c42-05f1-407f-af38-2c6c282611aa tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "204b866f-e9c5-441d-bf8e-77d324b39ecd" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2254.003529] env[62824]: DEBUG nova.compute.manager [None req-bb7e3c42-05f1-407f-af38-2c6c282611aa tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 204b866f-e9c5-441d-bf8e-77d324b39ecd] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2254.005342] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6de2224f-2303-4c41-afd0-035daa65f525 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2254.015714] env[62824]: DEBUG nova.compute.manager [None req-bb7e3c42-05f1-407f-af38-2c6c282611aa tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 204b866f-e9c5-441d-bf8e-77d324b39ecd] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62824) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3404}} [ 2254.016659] env[62824]: DEBUG nova.objects.instance [None req-bb7e3c42-05f1-407f-af38-2c6c282611aa tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lazy-loading 'flavor' on Instance uuid 204b866f-e9c5-441d-bf8e-77d324b39ecd {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2254.038887] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf150ff0-2f75-48f4-9cf2-573d5cf14e42 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2254.048876] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91dcdac7-f950-4170-ad4f-266ff746a080 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2254.088906] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-683dcd1b-93c8-4892-9c61-a61e08218596 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2254.099189] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96122fc3-85a8-4d7d-9f96-801f8698e447 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2254.114371] env[62824]: DEBUG nova.compute.provider_tree [None req-697b1a23-d14a-4d94-b716-8c4b25f6df27 tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2254.248288] env[62824]: DEBUG nova.compute.utils [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2254.250046] env[62824]: DEBUG nova.compute.manager [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 3acdf86e-45e9-4353-a282-7272451389e7] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2254.250279] env[62824]: DEBUG nova.network.neutron [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 3acdf86e-45e9-4353-a282-7272451389e7] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2254.303857] env[62824]: DEBUG oslo_vmware.api [None req-22a9cd95-c63b-4cbe-b831-e647907ecb3f tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2146671, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.495082} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2254.304159] env[62824]: INFO nova.virt.vmwareapi.ds_util [None req-22a9cd95-c63b-4cbe-b831-e647907ecb3f tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] bbfcb2e3-9326-4548-b15b-e054cbfd192e/9e2a7d30-212d-4ab8-9606-c5c6d52629e8-rescue.vmdk. [ 2254.305071] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9f0b817-62e7-417f-a52a-d141176fd95a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2254.333483] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-22a9cd95-c63b-4cbe-b831-e647907ecb3f tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Reconfiguring VM instance instance-00000039 to attach disk [datastore2] bbfcb2e3-9326-4548-b15b-e054cbfd192e/9e2a7d30-212d-4ab8-9606-c5c6d52629e8-rescue.vmdk or device None with type thin {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2254.333931] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a9ea7ee8-b7f4-40d3-b427-f3520dc349e9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2254.350832] env[62824]: DEBUG nova.policy [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '959d127a7d144b33a0cae94db5c11846', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dfff08982dad4790bf4d555e2b4db5e4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2254.358538] env[62824]: DEBUG oslo_vmware.api [None req-22a9cd95-c63b-4cbe-b831-e647907ecb3f tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for the task: (returnval){ [ 2254.358538] env[62824]: value = "task-2146672" [ 2254.358538] env[62824]: _type = "Task" [ 2254.358538] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2254.368357] env[62824]: DEBUG oslo_vmware.api [None req-22a9cd95-c63b-4cbe-b831-e647907ecb3f tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2146672, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2254.620685] env[62824]: DEBUG nova.scheduler.client.report [None req-697b1a23-d14a-4d94-b716-8c4b25f6df27 tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2254.628284] env[62824]: DEBUG nova.compute.manager [req-26e4f5b4-f378-450c-a9af-e2ccef034a22 req-eca94dce-1d36-43fc-8c5e-6b897d3aab02 service nova] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Received event network-vif-deleted-d358ab3c-f4fb-41f3-954b-19de8b22ef15 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2254.628384] env[62824]: INFO nova.compute.manager [req-26e4f5b4-f378-450c-a9af-e2ccef034a22 req-eca94dce-1d36-43fc-8c5e-6b897d3aab02 service nova] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Neutron deleted interface d358ab3c-f4fb-41f3-954b-19de8b22ef15; detaching it from the instance and deleting it from the info cache [ 2254.628558] env[62824]: DEBUG nova.network.neutron [req-26e4f5b4-f378-450c-a9af-e2ccef034a22 req-eca94dce-1d36-43fc-8c5e-6b897d3aab02 service nova] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2254.643484] env[62824]: DEBUG nova.network.neutron [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 3acdf86e-45e9-4353-a282-7272451389e7] Successfully created port: f35780f9-2827-45ec-abd6-0737b92ff5a3 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2254.741158] env[62824]: DEBUG nova.compute.manager [req-5bb82ae9-e753-456d-a6a2-481dfba31ede req-0cd5a13d-e5ee-4dcd-b22b-788c661e41e1 service nova] [instance: c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9] Received event network-vif-plugged-78cd208b-52d0-44fb-af4b-7012963b8b85 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2254.741158] env[62824]: DEBUG oslo_concurrency.lockutils [req-5bb82ae9-e753-456d-a6a2-481dfba31ede req-0cd5a13d-e5ee-4dcd-b22b-788c661e41e1 service nova] Acquiring lock "c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2254.741158] env[62824]: DEBUG oslo_concurrency.lockutils [req-5bb82ae9-e753-456d-a6a2-481dfba31ede req-0cd5a13d-e5ee-4dcd-b22b-788c661e41e1 service nova] Lock "c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2254.741158] env[62824]: DEBUG oslo_concurrency.lockutils [req-5bb82ae9-e753-456d-a6a2-481dfba31ede req-0cd5a13d-e5ee-4dcd-b22b-788c661e41e1 service nova] Lock "c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2254.741158] env[62824]: DEBUG nova.compute.manager [req-5bb82ae9-e753-456d-a6a2-481dfba31ede req-0cd5a13d-e5ee-4dcd-b22b-788c661e41e1 service nova] [instance: c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9] No waiting events found dispatching network-vif-plugged-78cd208b-52d0-44fb-af4b-7012963b8b85 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2254.742516] env[62824]: WARNING nova.compute.manager [req-5bb82ae9-e753-456d-a6a2-481dfba31ede req-0cd5a13d-e5ee-4dcd-b22b-788c661e41e1 service nova] [instance: c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9] Received unexpected event network-vif-plugged-78cd208b-52d0-44fb-af4b-7012963b8b85 for instance with vm_state building and task_state spawning. [ 2254.744065] env[62824]: DEBUG nova.network.neutron [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] [instance: c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9] Successfully updated port: 78cd208b-52d0-44fb-af4b-7012963b8b85 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2254.755115] env[62824]: DEBUG nova.compute.manager [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 3acdf86e-45e9-4353-a282-7272451389e7] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2254.854586] env[62824]: DEBUG nova.network.neutron [-] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2254.873074] env[62824]: DEBUG oslo_vmware.api [None req-22a9cd95-c63b-4cbe-b831-e647907ecb3f tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2146672, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2254.948745] env[62824]: INFO nova.compute.manager [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Rebuilding instance [ 2254.999479] env[62824]: DEBUG nova.compute.manager [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2255.000520] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbbb2590-64aa-41d9-8c5f-033de9756a4c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2255.029264] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb7e3c42-05f1-407f-af38-2c6c282611aa tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 204b866f-e9c5-441d-bf8e-77d324b39ecd] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2255.029808] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f0db8b1d-84f0-4849-93f4-02d8a0b52e2a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2255.040650] env[62824]: DEBUG oslo_vmware.api [None req-bb7e3c42-05f1-407f-af38-2c6c282611aa tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 2255.040650] env[62824]: value = "task-2146673" [ 2255.040650] env[62824]: _type = "Task" [ 2255.040650] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2255.050683] env[62824]: DEBUG oslo_vmware.api [None req-bb7e3c42-05f1-407f-af38-2c6c282611aa tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146673, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2255.126111] env[62824]: DEBUG oslo_concurrency.lockutils [None req-697b1a23-d14a-4d94-b716-8c4b25f6df27 tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.380s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2255.126811] env[62824]: INFO nova.compute.manager [None req-697b1a23-d14a-4d94-b716-8c4b25f6df27 tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Successfully reverted task state from rebuilding on failure for instance. [ 2255.132832] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.764s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2255.134338] env[62824]: INFO nova.compute.claims [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 815bd7ae-f461-4c74-9401-7c5fe43679ee] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2255.140204] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3d2082d5-ec7d-4403-9163-53a86c7f4bdd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2255.146856] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5b94787-ba64-406c-87cc-9891d881a8c4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2255.185851] env[62824]: DEBUG nova.compute.manager [req-26e4f5b4-f378-450c-a9af-e2ccef034a22 req-eca94dce-1d36-43fc-8c5e-6b897d3aab02 service nova] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Detach interface failed, port_id=d358ab3c-f4fb-41f3-954b-19de8b22ef15, reason: Instance 2f315f9e-94e4-47ef-9503-ee92b59e5452 could not be found. {{(pid=62824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11482}} [ 2255.246798] env[62824]: DEBUG oslo_concurrency.lockutils [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Acquiring lock "refresh_cache-c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2255.247083] env[62824]: DEBUG oslo_concurrency.lockutils [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Acquired lock "refresh_cache-c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2255.247277] env[62824]: DEBUG nova.network.neutron [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] [instance: c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2255.358375] env[62824]: INFO nova.compute.manager [-] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Took 1.61 seconds to deallocate network for instance. [ 2255.372749] env[62824]: DEBUG oslo_vmware.api [None req-22a9cd95-c63b-4cbe-b831-e647907ecb3f tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2146672, 'name': ReconfigVM_Task, 'duration_secs': 0.971885} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2255.372749] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-22a9cd95-c63b-4cbe-b831-e647907ecb3f tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Reconfigured VM instance instance-00000039 to attach disk [datastore2] bbfcb2e3-9326-4548-b15b-e054cbfd192e/9e2a7d30-212d-4ab8-9606-c5c6d52629e8-rescue.vmdk or device None with type thin {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2255.372749] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f20f0900-7b80-49dc-87d9-26efb6159e85 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2255.406523] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-075c68c4-0e07-44c8-bf64-124feb9fdae3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2255.425923] env[62824]: DEBUG oslo_vmware.api [None req-22a9cd95-c63b-4cbe-b831-e647907ecb3f tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for the task: (returnval){ [ 2255.425923] env[62824]: value = "task-2146674" [ 2255.425923] env[62824]: _type = "Task" [ 2255.425923] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2255.430779] env[62824]: DEBUG oslo_vmware.api [None req-22a9cd95-c63b-4cbe-b831-e647907ecb3f tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2146674, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2255.550371] env[62824]: DEBUG oslo_vmware.api [None req-bb7e3c42-05f1-407f-af38-2c6c282611aa tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146673, 'name': PowerOffVM_Task, 'duration_secs': 0.286845} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2255.550793] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb7e3c42-05f1-407f-af38-2c6c282611aa tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 204b866f-e9c5-441d-bf8e-77d324b39ecd] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2255.551129] env[62824]: DEBUG nova.compute.manager [None req-bb7e3c42-05f1-407f-af38-2c6c282611aa tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 204b866f-e9c5-441d-bf8e-77d324b39ecd] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2255.552025] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14f2c525-c0ce-460e-a401-4dac1cab725e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2255.559269] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Acquiring lock "71731b21-c302-4d66-a579-889165673712" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2255.559677] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Lock "71731b21-c302-4d66-a579-889165673712" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2255.763031] env[62824]: DEBUG nova.compute.manager [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 3acdf86e-45e9-4353-a282-7272451389e7] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2255.778257] env[62824]: DEBUG nova.network.neutron [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] [instance: c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2255.792515] env[62824]: DEBUG nova.virt.hardware [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2255.792853] env[62824]: DEBUG nova.virt.hardware [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2255.792995] env[62824]: DEBUG nova.virt.hardware [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2255.793392] env[62824]: DEBUG nova.virt.hardware [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2255.794537] env[62824]: DEBUG nova.virt.hardware [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2255.794537] env[62824]: DEBUG nova.virt.hardware [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2255.794537] env[62824]: DEBUG nova.virt.hardware [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2255.794683] env[62824]: DEBUG nova.virt.hardware [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2255.794964] env[62824]: DEBUG nova.virt.hardware [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2255.795297] env[62824]: DEBUG nova.virt.hardware [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2255.795531] env[62824]: DEBUG nova.virt.hardware [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2255.796653] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51754b4f-19dd-4e44-9ec1-0547d50e14ff {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2255.806618] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22209642-459d-425f-9180-6e2531068a07 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2255.820384] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e0a2c92a-cfc3-43db-a3f8-fd54c22d888b tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquiring lock "interface-fc77c44a-180c-46ff-9690-9072c6213c91-2b6381b4-f3c7-4e22-b698-de190c8199ee" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2255.820695] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e0a2c92a-cfc3-43db-a3f8-fd54c22d888b tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lock "interface-fc77c44a-180c-46ff-9690-9072c6213c91-2b6381b4-f3c7-4e22-b698-de190c8199ee" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2255.934896] env[62824]: INFO nova.compute.manager [None req-d6968317-d8b9-4e00-a6a9-ee3a22a078b2 tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Took 0.58 seconds to detach 1 volumes for instance. [ 2255.936731] env[62824]: DEBUG oslo_vmware.api [None req-22a9cd95-c63b-4cbe-b831-e647907ecb3f tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2146674, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2255.937285] env[62824]: DEBUG nova.compute.manager [None req-d6968317-d8b9-4e00-a6a9-ee3a22a078b2 tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Deleting volume: 8d462122-e7ae-4257-8ce4-0a79a4acae13 {{(pid=62824) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3282}} [ 2255.972437] env[62824]: DEBUG nova.network.neutron [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] [instance: c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9] Updating instance_info_cache with network_info: [{"id": "78cd208b-52d0-44fb-af4b-7012963b8b85", "address": "fa:16:3e:53:36:78", "network": {"id": "e2052ab9-cbbb-4f56-acff-4e0af548a2fc", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-892290291-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58f9886b55b64e90af5d3b646429a232", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3753f451-fa23-4988-9361-074fb0bd3fd4", "external-id": "nsx-vlan-transportzone-440", "segmentation_id": 440, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap78cd208b-52", "ovs_interfaceid": "78cd208b-52d0-44fb-af4b-7012963b8b85", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2256.034201] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2256.034201] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-09404ab4-11aa-472e-a7f8-4562b4345ac1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2256.034201] env[62824]: DEBUG oslo_vmware.api [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2256.034201] env[62824]: value = "task-2146675" [ 2256.034201] env[62824]: _type = "Task" [ 2256.034201] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2256.036979] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] VM already powered off {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 2256.037388] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2256.038131] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-040a72c0-f9fb-45fc-8b93-f44e2337aba8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2256.048136] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2256.048136] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8541ad6b-542c-4e4d-974a-15f924edfbcd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2256.062077] env[62824]: DEBUG nova.compute.manager [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] [instance: 71731b21-c302-4d66-a579-889165673712] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2256.067757] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bb7e3c42-05f1-407f-af38-2c6c282611aa tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "204b866f-e9c5-441d-bf8e-77d324b39ecd" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.064s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2256.168888] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2256.169208] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Deleting contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2256.169395] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Deleting the datastore file [datastore2] 0394d368-c846-4f3b-bfcf-2bc8e858052a {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2256.173793] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-070eaac4-985c-4133-a479-d3d7e5b56d48 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2256.182896] env[62824]: DEBUG oslo_vmware.api [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2256.182896] env[62824]: value = "task-2146678" [ 2256.182896] env[62824]: _type = "Task" [ 2256.182896] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2256.194817] env[62824]: DEBUG oslo_vmware.api [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146678, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2256.324762] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e0a2c92a-cfc3-43db-a3f8-fd54c22d888b tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquiring lock "fc77c44a-180c-46ff-9690-9072c6213c91" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2256.324938] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e0a2c92a-cfc3-43db-a3f8-fd54c22d888b tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquired lock "fc77c44a-180c-46ff-9690-9072c6213c91" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2256.325973] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85035f23-8ac3-4ee9-abfe-c8695151e866 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2256.348974] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-402f60ec-5fe9-45cb-aee6-1ff3de3d6ffd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2256.379328] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e0a2c92a-cfc3-43db-a3f8-fd54c22d888b tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Reconfiguring VM to detach interface {{(pid=62824) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 2256.382597] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8ca941ec-a576-4be6-8b8a-e6af7943bf72 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2256.402961] env[62824]: DEBUG oslo_vmware.api [None req-e0a2c92a-cfc3-43db-a3f8-fd54c22d888b tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Waiting for the task: (returnval){ [ 2256.402961] env[62824]: value = "task-2146679" [ 2256.402961] env[62824]: _type = "Task" [ 2256.402961] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2256.408473] env[62824]: DEBUG nova.network.neutron [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 3acdf86e-45e9-4353-a282-7272451389e7] Successfully updated port: f35780f9-2827-45ec-abd6-0737b92ff5a3 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2256.421696] env[62824]: DEBUG oslo_vmware.api [None req-e0a2c92a-cfc3-43db-a3f8-fd54c22d888b tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146679, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2256.432204] env[62824]: DEBUG oslo_vmware.api [None req-22a9cd95-c63b-4cbe-b831-e647907ecb3f tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2146674, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2256.482160] env[62824]: DEBUG oslo_concurrency.lockutils [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Releasing lock "refresh_cache-c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2256.482495] env[62824]: DEBUG nova.compute.manager [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] [instance: c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9] Instance network_info: |[{"id": "78cd208b-52d0-44fb-af4b-7012963b8b85", "address": "fa:16:3e:53:36:78", "network": {"id": "e2052ab9-cbbb-4f56-acff-4e0af548a2fc", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-892290291-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58f9886b55b64e90af5d3b646429a232", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3753f451-fa23-4988-9361-074fb0bd3fd4", "external-id": "nsx-vlan-transportzone-440", "segmentation_id": 440, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap78cd208b-52", "ovs_interfaceid": "78cd208b-52d0-44fb-af4b-7012963b8b85", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2256.482912] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] [instance: c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:53:36:78', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3753f451-fa23-4988-9361-074fb0bd3fd4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '78cd208b-52d0-44fb-af4b-7012963b8b85', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2256.491174] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Creating folder: Project (58f9886b55b64e90af5d3b646429a232). Parent ref: group-v438503. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 2256.491174] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0ed74586-8513-4c3c-be06-60f7667b4aa8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2256.493861] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d6968317-d8b9-4e00-a6a9-ee3a22a078b2 tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2256.504412] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Created folder: Project (58f9886b55b64e90af5d3b646429a232) in parent group-v438503. [ 2256.504412] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Creating folder: Instances. Parent ref: group-v438734. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 2256.504412] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a1bab952-a099-4013-9c7f-0f1158c2349f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2256.514732] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Created folder: Instances in parent group-v438734. [ 2256.515677] env[62824]: DEBUG oslo.service.loopingcall [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2256.515677] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2256.515677] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1c9eedc8-a638-48a8-84ac-72d508b1600a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2256.532749] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11d8c2b3-1a01-4a83-9bdb-60a18e8b23fc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2256.542158] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-756a8914-4519-4d86-9dc2-da4bf4c9a6d8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2256.545451] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2256.545451] env[62824]: value = "task-2146682" [ 2256.545451] env[62824]: _type = "Task" [ 2256.545451] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2256.588548] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54c10149-97b9-4852-9cdb-7de212b9c36d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2256.596046] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146682, 'name': CreateVM_Task} progress is 15%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2256.601305] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afde8e7b-1854-4d93-8be3-8972065a4af9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2256.606183] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2256.617918] env[62824]: DEBUG nova.compute.provider_tree [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2256.685021] env[62824]: DEBUG nova.compute.manager [req-26ee662f-e59b-4843-ad33-6426f53d635a req-6e263cec-5007-4e05-b0ef-8b4e4a23182b service nova] [instance: 3acdf86e-45e9-4353-a282-7272451389e7] Received event network-vif-plugged-f35780f9-2827-45ec-abd6-0737b92ff5a3 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2256.685021] env[62824]: DEBUG oslo_concurrency.lockutils [req-26ee662f-e59b-4843-ad33-6426f53d635a req-6e263cec-5007-4e05-b0ef-8b4e4a23182b service nova] Acquiring lock "3acdf86e-45e9-4353-a282-7272451389e7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2256.685021] env[62824]: DEBUG oslo_concurrency.lockutils [req-26ee662f-e59b-4843-ad33-6426f53d635a req-6e263cec-5007-4e05-b0ef-8b4e4a23182b service nova] Lock "3acdf86e-45e9-4353-a282-7272451389e7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2256.685021] env[62824]: DEBUG oslo_concurrency.lockutils [req-26ee662f-e59b-4843-ad33-6426f53d635a req-6e263cec-5007-4e05-b0ef-8b4e4a23182b service nova] Lock "3acdf86e-45e9-4353-a282-7272451389e7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2256.685021] env[62824]: DEBUG nova.compute.manager [req-26ee662f-e59b-4843-ad33-6426f53d635a req-6e263cec-5007-4e05-b0ef-8b4e4a23182b service nova] [instance: 3acdf86e-45e9-4353-a282-7272451389e7] No waiting events found dispatching network-vif-plugged-f35780f9-2827-45ec-abd6-0737b92ff5a3 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2256.685021] env[62824]: WARNING nova.compute.manager [req-26ee662f-e59b-4843-ad33-6426f53d635a req-6e263cec-5007-4e05-b0ef-8b4e4a23182b service nova] [instance: 3acdf86e-45e9-4353-a282-7272451389e7] Received unexpected event network-vif-plugged-f35780f9-2827-45ec-abd6-0737b92ff5a3 for instance with vm_state building and task_state spawning. [ 2256.685021] env[62824]: DEBUG nova.compute.manager [req-26ee662f-e59b-4843-ad33-6426f53d635a req-6e263cec-5007-4e05-b0ef-8b4e4a23182b service nova] [instance: 3acdf86e-45e9-4353-a282-7272451389e7] Received event network-changed-f35780f9-2827-45ec-abd6-0737b92ff5a3 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2256.685021] env[62824]: DEBUG nova.compute.manager [req-26ee662f-e59b-4843-ad33-6426f53d635a req-6e263cec-5007-4e05-b0ef-8b4e4a23182b service nova] [instance: 3acdf86e-45e9-4353-a282-7272451389e7] Refreshing instance network info cache due to event network-changed-f35780f9-2827-45ec-abd6-0737b92ff5a3. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2256.685021] env[62824]: DEBUG oslo_concurrency.lockutils [req-26ee662f-e59b-4843-ad33-6426f53d635a req-6e263cec-5007-4e05-b0ef-8b4e4a23182b service nova] Acquiring lock "refresh_cache-3acdf86e-45e9-4353-a282-7272451389e7" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2256.685021] env[62824]: DEBUG oslo_concurrency.lockutils [req-26ee662f-e59b-4843-ad33-6426f53d635a req-6e263cec-5007-4e05-b0ef-8b4e4a23182b service nova] Acquired lock "refresh_cache-3acdf86e-45e9-4353-a282-7272451389e7" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2256.685021] env[62824]: DEBUG nova.network.neutron [req-26ee662f-e59b-4843-ad33-6426f53d635a req-6e263cec-5007-4e05-b0ef-8b4e4a23182b service nova] [instance: 3acdf86e-45e9-4353-a282-7272451389e7] Refreshing network info cache for port f35780f9-2827-45ec-abd6-0737b92ff5a3 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2256.696087] env[62824]: DEBUG oslo_vmware.api [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146678, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.133005} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2256.696846] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2256.697054] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Deleted contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2256.697263] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2256.806837] env[62824]: DEBUG nova.compute.manager [req-79f218b6-6c6a-44b0-a592-52c3720eff01 req-c59aee4e-629a-47ba-b994-f2d78bc61e73 service nova] [instance: c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9] Received event network-changed-78cd208b-52d0-44fb-af4b-7012963b8b85 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2256.807157] env[62824]: DEBUG nova.compute.manager [req-79f218b6-6c6a-44b0-a592-52c3720eff01 req-c59aee4e-629a-47ba-b994-f2d78bc61e73 service nova] [instance: c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9] Refreshing instance network info cache due to event network-changed-78cd208b-52d0-44fb-af4b-7012963b8b85. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2256.807322] env[62824]: DEBUG oslo_concurrency.lockutils [req-79f218b6-6c6a-44b0-a592-52c3720eff01 req-c59aee4e-629a-47ba-b994-f2d78bc61e73 service nova] Acquiring lock "refresh_cache-c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2256.807471] env[62824]: DEBUG oslo_concurrency.lockutils [req-79f218b6-6c6a-44b0-a592-52c3720eff01 req-c59aee4e-629a-47ba-b994-f2d78bc61e73 service nova] Acquired lock "refresh_cache-c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2256.807635] env[62824]: DEBUG nova.network.neutron [req-79f218b6-6c6a-44b0-a592-52c3720eff01 req-c59aee4e-629a-47ba-b994-f2d78bc61e73 service nova] [instance: c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9] Refreshing network info cache for port 78cd208b-52d0-44fb-af4b-7012963b8b85 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2256.912946] env[62824]: DEBUG oslo_vmware.api [None req-e0a2c92a-cfc3-43db-a3f8-fd54c22d888b tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146679, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2256.921630] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquiring lock "refresh_cache-3acdf86e-45e9-4353-a282-7272451389e7" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2256.932932] env[62824]: DEBUG oslo_vmware.api [None req-22a9cd95-c63b-4cbe-b831-e647907ecb3f tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2146674, 'name': ReconfigVM_Task, 'duration_secs': 1.056353} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2256.933223] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-22a9cd95-c63b-4cbe-b831-e647907ecb3f tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2256.933479] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-10800c2a-9a82-4f07-842d-9814d1562f33 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2256.940628] env[62824]: DEBUG oslo_vmware.api [None req-22a9cd95-c63b-4cbe-b831-e647907ecb3f tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for the task: (returnval){ [ 2256.940628] env[62824]: value = "task-2146683" [ 2256.940628] env[62824]: _type = "Task" [ 2256.940628] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2256.948756] env[62824]: DEBUG oslo_vmware.api [None req-22a9cd95-c63b-4cbe-b831-e647907ecb3f tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2146683, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2257.056044] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146682, 'name': CreateVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2257.086384] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e89a221d-5c30-48d4-bfc0-a91f11a74aac tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquiring lock "204b866f-e9c5-441d-bf8e-77d324b39ecd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2257.086384] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e89a221d-5c30-48d4-bfc0-a91f11a74aac tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "204b866f-e9c5-441d-bf8e-77d324b39ecd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2257.086384] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e89a221d-5c30-48d4-bfc0-a91f11a74aac tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquiring lock "204b866f-e9c5-441d-bf8e-77d324b39ecd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2257.086384] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e89a221d-5c30-48d4-bfc0-a91f11a74aac tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "204b866f-e9c5-441d-bf8e-77d324b39ecd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2257.086384] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e89a221d-5c30-48d4-bfc0-a91f11a74aac tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "204b866f-e9c5-441d-bf8e-77d324b39ecd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2257.087378] env[62824]: INFO nova.compute.manager [None req-e89a221d-5c30-48d4-bfc0-a91f11a74aac tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 204b866f-e9c5-441d-bf8e-77d324b39ecd] Terminating instance [ 2257.121244] env[62824]: DEBUG nova.scheduler.client.report [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2257.225101] env[62824]: DEBUG nova.network.neutron [req-26ee662f-e59b-4843-ad33-6426f53d635a req-6e263cec-5007-4e05-b0ef-8b4e4a23182b service nova] [instance: 3acdf86e-45e9-4353-a282-7272451389e7] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2257.332891] env[62824]: DEBUG nova.network.neutron [req-26ee662f-e59b-4843-ad33-6426f53d635a req-6e263cec-5007-4e05-b0ef-8b4e4a23182b service nova] [instance: 3acdf86e-45e9-4353-a282-7272451389e7] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2257.413762] env[62824]: DEBUG oslo_vmware.api [None req-e0a2c92a-cfc3-43db-a3f8-fd54c22d888b tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146679, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2257.450143] env[62824]: DEBUG oslo_vmware.api [None req-22a9cd95-c63b-4cbe-b831-e647907ecb3f tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2146683, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2257.559018] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146682, 'name': CreateVM_Task, 'duration_secs': 0.651785} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2257.559600] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2257.560114] env[62824]: DEBUG oslo_concurrency.lockutils [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2257.560218] env[62824]: DEBUG oslo_concurrency.lockutils [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2257.560474] env[62824]: DEBUG oslo_concurrency.lockutils [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2257.560732] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a4f5aed1-17d0-409a-a7c7-ed2da0fe4c4c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2257.565545] env[62824]: DEBUG oslo_vmware.api [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Waiting for the task: (returnval){ [ 2257.565545] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52e4578a-cd36-e294-f1ad-cf291b4db7b9" [ 2257.565545] env[62824]: _type = "Task" [ 2257.565545] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2257.573682] env[62824]: DEBUG oslo_vmware.api [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52e4578a-cd36-e294-f1ad-cf291b4db7b9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2257.574580] env[62824]: DEBUG nova.network.neutron [req-79f218b6-6c6a-44b0-a592-52c3720eff01 req-c59aee4e-629a-47ba-b994-f2d78bc61e73 service nova] [instance: c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9] Updated VIF entry in instance network info cache for port 78cd208b-52d0-44fb-af4b-7012963b8b85. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2257.574903] env[62824]: DEBUG nova.network.neutron [req-79f218b6-6c6a-44b0-a592-52c3720eff01 req-c59aee4e-629a-47ba-b994-f2d78bc61e73 service nova] [instance: c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9] Updating instance_info_cache with network_info: [{"id": "78cd208b-52d0-44fb-af4b-7012963b8b85", "address": "fa:16:3e:53:36:78", "network": {"id": "e2052ab9-cbbb-4f56-acff-4e0af548a2fc", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-892290291-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "58f9886b55b64e90af5d3b646429a232", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3753f451-fa23-4988-9361-074fb0bd3fd4", "external-id": "nsx-vlan-transportzone-440", "segmentation_id": 440, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap78cd208b-52", "ovs_interfaceid": "78cd208b-52d0-44fb-af4b-7012963b8b85", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2257.592453] env[62824]: DEBUG nova.compute.manager [None req-e89a221d-5c30-48d4-bfc0-a91f11a74aac tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 204b866f-e9c5-441d-bf8e-77d324b39ecd] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2257.592669] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e89a221d-5c30-48d4-bfc0-a91f11a74aac tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 204b866f-e9c5-441d-bf8e-77d324b39ecd] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2257.593715] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6129f363-09b1-462f-b5b5-a5297e03c709 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2257.603661] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e89a221d-5c30-48d4-bfc0-a91f11a74aac tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 204b866f-e9c5-441d-bf8e-77d324b39ecd] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2257.603661] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-19269b82-a171-47a5-a4db-0d982dccb6e7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2257.626700] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.493s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2257.626700] env[62824]: DEBUG nova.compute.manager [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 815bd7ae-f461-4c74-9401-7c5fe43679ee] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2257.629308] env[62824]: DEBUG oslo_concurrency.lockutils [None req-61c785a4-c6ea-4ee0-8967-908719fa995a tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.987s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2257.629540] env[62824]: DEBUG nova.objects.instance [None req-61c785a4-c6ea-4ee0-8967-908719fa995a tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Lazy-loading 'resources' on Instance uuid b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2257.737921] env[62824]: DEBUG nova.virt.hardware [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2257.738190] env[62824]: DEBUG nova.virt.hardware [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2257.738350] env[62824]: DEBUG nova.virt.hardware [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2257.738533] env[62824]: DEBUG nova.virt.hardware [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2257.738681] env[62824]: DEBUG nova.virt.hardware [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2257.738830] env[62824]: DEBUG nova.virt.hardware [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2257.740833] env[62824]: DEBUG nova.virt.hardware [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2257.741124] env[62824]: DEBUG nova.virt.hardware [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2257.741345] env[62824]: DEBUG nova.virt.hardware [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2257.741550] env[62824]: DEBUG nova.virt.hardware [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2257.741741] env[62824]: DEBUG nova.virt.hardware [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2257.745250] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efbef75b-417c-4eef-9561-7248e13c0ca6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2257.765125] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76f7d90c-e2b5-44e7-83f2-988939c7ed83 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2257.780026] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:df:88:fa', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '69f65356-c85e-4b7f-ad28-7c7b5e8cf50c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2c7a5b60-534e-457c-a27b-29a489877783', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2257.785890] env[62824]: DEBUG oslo.service.loopingcall [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2257.786449] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2257.787182] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c14da812-ba86-496d-9201-2c581d26ab3d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2257.806611] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2257.806611] env[62824]: value = "task-2146685" [ 2257.806611] env[62824]: _type = "Task" [ 2257.806611] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2257.816137] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146685, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2257.835875] env[62824]: DEBUG oslo_concurrency.lockutils [req-26ee662f-e59b-4843-ad33-6426f53d635a req-6e263cec-5007-4e05-b0ef-8b4e4a23182b service nova] Releasing lock "refresh_cache-3acdf86e-45e9-4353-a282-7272451389e7" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2257.836446] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquired lock "refresh_cache-3acdf86e-45e9-4353-a282-7272451389e7" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2257.836683] env[62824]: DEBUG nova.network.neutron [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 3acdf86e-45e9-4353-a282-7272451389e7] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2257.914286] env[62824]: DEBUG oslo_vmware.api [None req-e0a2c92a-cfc3-43db-a3f8-fd54c22d888b tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146679, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2257.950474] env[62824]: DEBUG oslo_vmware.api [None req-22a9cd95-c63b-4cbe-b831-e647907ecb3f tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2146683, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2258.076148] env[62824]: DEBUG oslo_vmware.api [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52e4578a-cd36-e294-f1ad-cf291b4db7b9, 'name': SearchDatastore_Task, 'duration_secs': 0.011446} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2258.076480] env[62824]: DEBUG oslo_concurrency.lockutils [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2258.076759] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] [instance: c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2258.077082] env[62824]: DEBUG oslo_concurrency.lockutils [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2258.077290] env[62824]: DEBUG oslo_concurrency.lockutils [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2258.077519] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2258.078084] env[62824]: DEBUG oslo_concurrency.lockutils [req-79f218b6-6c6a-44b0-a592-52c3720eff01 req-c59aee4e-629a-47ba-b994-f2d78bc61e73 service nova] Releasing lock "refresh_cache-c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2258.078449] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8c5d6dfe-7e1c-4763-997c-f70dafa19add {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2258.086648] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2258.086870] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2258.087662] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9c22502b-b9d4-4967-a77a-39a6b8f365ff {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2258.092964] env[62824]: DEBUG oslo_vmware.api [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Waiting for the task: (returnval){ [ 2258.092964] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]527d0de4-17f3-9785-d798-debf5f8f574f" [ 2258.092964] env[62824]: _type = "Task" [ 2258.092964] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2258.100542] env[62824]: DEBUG oslo_vmware.api [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]527d0de4-17f3-9785-d798-debf5f8f574f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2258.132496] env[62824]: DEBUG nova.compute.utils [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2258.137151] env[62824]: DEBUG nova.compute.manager [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 815bd7ae-f461-4c74-9401-7c5fe43679ee] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2258.137151] env[62824]: DEBUG nova.network.neutron [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 815bd7ae-f461-4c74-9401-7c5fe43679ee] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2258.188411] env[62824]: DEBUG nova.policy [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3492c645a51548408a639e3af73c213c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '35b7289f4746463e9b1cc47fe914a835', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2258.322482] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146685, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2258.387719] env[62824]: DEBUG nova.network.neutron [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 3acdf86e-45e9-4353-a282-7272451389e7] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2258.421946] env[62824]: DEBUG oslo_vmware.api [None req-e0a2c92a-cfc3-43db-a3f8-fd54c22d888b tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146679, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2258.453866] env[62824]: DEBUG oslo_vmware.api [None req-22a9cd95-c63b-4cbe-b831-e647907ecb3f tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2146683, 'name': PowerOnVM_Task, 'duration_secs': 1.491069} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2258.454914] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-22a9cd95-c63b-4cbe-b831-e647907ecb3f tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2258.458875] env[62824]: DEBUG nova.compute.manager [None req-22a9cd95-c63b-4cbe-b831-e647907ecb3f tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2258.459659] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f58f3410-6f39-4997-862f-1894678a1bb2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2258.467903] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92f77bcc-b9e6-44f7-a0f9-c30c1a980060 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2258.480019] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a8d769d-8cfb-4b9c-8ae7-e051349496a9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2258.513521] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7534c3fc-9b37-45ee-a59e-bdd605052485 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2258.521645] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2721a3d-8c84-4dc6-9e8e-7e31601c8c4b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2258.535995] env[62824]: DEBUG nova.compute.provider_tree [None req-61c785a4-c6ea-4ee0-8967-908719fa995a tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2258.550362] env[62824]: DEBUG nova.network.neutron [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 3acdf86e-45e9-4353-a282-7272451389e7] Updating instance_info_cache with network_info: [{"id": "f35780f9-2827-45ec-abd6-0737b92ff5a3", "address": "fa:16:3e:29:24:49", "network": {"id": "0f512108-670f-4a85-9381-108e18b3b5ed", "bridge": "br-int", "label": "tempest-ServersTestJSON-766397782-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dfff08982dad4790bf4d555e2b4db5e4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f3a2eb5-353f-45c5-a73b-869626f4bb13", "external-id": "nsx-vlan-transportzone-411", "segmentation_id": 411, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf35780f9-28", "ovs_interfaceid": "f35780f9-2827-45ec-abd6-0737b92ff5a3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2258.603709] env[62824]: DEBUG oslo_vmware.api [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]527d0de4-17f3-9785-d798-debf5f8f574f, 'name': SearchDatastore_Task, 'duration_secs': 0.009594} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2258.604573] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f540310b-c2ab-4e67-8caa-96cbb71612d0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2258.609837] env[62824]: DEBUG oslo_vmware.api [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Waiting for the task: (returnval){ [ 2258.609837] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5223209b-a6b9-36fe-809b-d911d4434662" [ 2258.609837] env[62824]: _type = "Task" [ 2258.609837] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2258.617250] env[62824]: DEBUG oslo_vmware.api [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5223209b-a6b9-36fe-809b-d911d4434662, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2258.637014] env[62824]: DEBUG nova.compute.manager [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 815bd7ae-f461-4c74-9401-7c5fe43679ee] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2258.661605] env[62824]: DEBUG nova.network.neutron [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 815bd7ae-f461-4c74-9401-7c5fe43679ee] Successfully created port: 3cb13172-9b03-4150-8070-b5b5f504d942 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2258.818473] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146685, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2258.916715] env[62824]: DEBUG oslo_vmware.api [None req-e0a2c92a-cfc3-43db-a3f8-fd54c22d888b tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146679, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2259.039460] env[62824]: DEBUG nova.scheduler.client.report [None req-61c785a4-c6ea-4ee0-8967-908719fa995a tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2259.052854] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Releasing lock "refresh_cache-3acdf86e-45e9-4353-a282-7272451389e7" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2259.053197] env[62824]: DEBUG nova.compute.manager [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 3acdf86e-45e9-4353-a282-7272451389e7] Instance network_info: |[{"id": "f35780f9-2827-45ec-abd6-0737b92ff5a3", "address": "fa:16:3e:29:24:49", "network": {"id": "0f512108-670f-4a85-9381-108e18b3b5ed", "bridge": "br-int", "label": "tempest-ServersTestJSON-766397782-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dfff08982dad4790bf4d555e2b4db5e4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f3a2eb5-353f-45c5-a73b-869626f4bb13", "external-id": "nsx-vlan-transportzone-411", "segmentation_id": 411, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf35780f9-28", "ovs_interfaceid": "f35780f9-2827-45ec-abd6-0737b92ff5a3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2259.053752] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 3acdf86e-45e9-4353-a282-7272451389e7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:29:24:49', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9f3a2eb5-353f-45c5-a73b-869626f4bb13', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f35780f9-2827-45ec-abd6-0737b92ff5a3', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2259.061489] env[62824]: DEBUG oslo.service.loopingcall [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2259.062272] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3acdf86e-45e9-4353-a282-7272451389e7] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2259.062500] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c7d35681-87ad-48c9-bdfd-3b7f5d318d31 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2259.083074] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2259.083074] env[62824]: value = "task-2146686" [ 2259.083074] env[62824]: _type = "Task" [ 2259.083074] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2259.091072] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146686, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2259.119623] env[62824]: DEBUG oslo_vmware.api [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5223209b-a6b9-36fe-809b-d911d4434662, 'name': SearchDatastore_Task, 'duration_secs': 0.011014} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2259.119885] env[62824]: DEBUG oslo_concurrency.lockutils [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2259.120166] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9/c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2259.120410] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-769c2cda-93fc-43fe-8bcf-a6b33500786e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2259.126353] env[62824]: DEBUG oslo_vmware.api [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Waiting for the task: (returnval){ [ 2259.126353] env[62824]: value = "task-2146687" [ 2259.126353] env[62824]: _type = "Task" [ 2259.126353] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2259.135644] env[62824]: DEBUG oslo_vmware.api [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Task: {'id': task-2146687, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2259.320129] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146685, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2259.418464] env[62824]: DEBUG oslo_vmware.api [None req-e0a2c92a-cfc3-43db-a3f8-fd54c22d888b tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146679, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2259.545203] env[62824]: DEBUG oslo_concurrency.lockutils [None req-61c785a4-c6ea-4ee0-8967-908719fa995a tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.916s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2259.547782] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d6968317-d8b9-4e00-a6a9-ee3a22a078b2 tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.054s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2259.548092] env[62824]: DEBUG nova.objects.instance [None req-d6968317-d8b9-4e00-a6a9-ee3a22a078b2 tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Lazy-loading 'resources' on Instance uuid 2f315f9e-94e4-47ef-9503-ee92b59e5452 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2259.568672] env[62824]: INFO nova.scheduler.client.report [None req-61c785a4-c6ea-4ee0-8967-908719fa995a tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Deleted allocations for instance b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a [ 2259.593556] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146686, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2259.635740] env[62824]: DEBUG oslo_vmware.api [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Task: {'id': task-2146687, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.441115} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2259.636014] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9/c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2259.636244] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] [instance: c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2259.636493] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1e9f2308-5e0c-4147-9bcd-99a5863d8946 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2259.642921] env[62824]: DEBUG oslo_vmware.api [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Waiting for the task: (returnval){ [ 2259.642921] env[62824]: value = "task-2146688" [ 2259.642921] env[62824]: _type = "Task" [ 2259.642921] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2259.647215] env[62824]: DEBUG nova.compute.manager [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 815bd7ae-f461-4c74-9401-7c5fe43679ee] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2259.653812] env[62824]: DEBUG oslo_vmware.api [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Task: {'id': task-2146688, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2259.671702] env[62824]: DEBUG nova.virt.hardware [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2259.671982] env[62824]: DEBUG nova.virt.hardware [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2259.672159] env[62824]: DEBUG nova.virt.hardware [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2259.672350] env[62824]: DEBUG nova.virt.hardware [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2259.672521] env[62824]: DEBUG nova.virt.hardware [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2259.672691] env[62824]: DEBUG nova.virt.hardware [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2259.672915] env[62824]: DEBUG nova.virt.hardware [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2259.673092] env[62824]: DEBUG nova.virt.hardware [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2259.673287] env[62824]: DEBUG nova.virt.hardware [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2259.673473] env[62824]: DEBUG nova.virt.hardware [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2259.673663] env[62824]: DEBUG nova.virt.hardware [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2259.674557] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd810d8f-5b4e-4f42-8548-3fd81efec097 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2259.683446] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02c046a3-7f7a-4099-90d7-6255c89480ea {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2259.819690] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146685, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2259.918681] env[62824]: DEBUG oslo_vmware.api [None req-e0a2c92a-cfc3-43db-a3f8-fd54c22d888b tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146679, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2260.076020] env[62824]: DEBUG oslo_concurrency.lockutils [None req-61c785a4-c6ea-4ee0-8967-908719fa995a tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Lock "b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.042s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2260.098422] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146686, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2260.153382] env[62824]: DEBUG oslo_vmware.api [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Task: {'id': task-2146688, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070548} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2260.153708] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] [instance: c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2260.154617] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9638f807-2331-4822-8f63-16e299ee3573 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2260.177937] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] [instance: c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9] Reconfiguring VM instance instance-0000005f to attach disk [datastore1] c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9/c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2260.180619] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c576353b-15ad-4a64-a75f-49936f98623a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2260.201769] env[62824]: DEBUG oslo_vmware.api [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Waiting for the task: (returnval){ [ 2260.201769] env[62824]: value = "task-2146689" [ 2260.201769] env[62824]: _type = "Task" [ 2260.201769] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2260.211960] env[62824]: DEBUG oslo_vmware.api [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Task: {'id': task-2146689, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2260.280305] env[62824]: INFO nova.compute.manager [None req-5b20a72b-b561-4d0f-a3b8-8dea853cf384 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Unrescuing [ 2260.280558] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5b20a72b-b561-4d0f-a3b8-8dea853cf384 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Acquiring lock "refresh_cache-bbfcb2e3-9326-4548-b15b-e054cbfd192e" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2260.280708] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5b20a72b-b561-4d0f-a3b8-8dea853cf384 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Acquired lock "refresh_cache-bbfcb2e3-9326-4548-b15b-e054cbfd192e" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2260.280874] env[62824]: DEBUG nova.network.neutron [None req-5b20a72b-b561-4d0f-a3b8-8dea853cf384 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2260.295622] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2777a11-ea3c-4e7c-9e3f-5dcfefbf6cab {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2260.303515] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc93941a-2689-4c4a-b861-677b2ae224ad {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2260.335969] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d3f7dc7-7084-478c-bae3-6a46d761468b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2260.346293] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed3f35ad-249e-480f-9303-1c81af33ed08 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2260.349822] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146685, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2260.360949] env[62824]: DEBUG nova.compute.provider_tree [None req-d6968317-d8b9-4e00-a6a9-ee3a22a078b2 tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2260.418642] env[62824]: DEBUG oslo_vmware.api [None req-e0a2c92a-cfc3-43db-a3f8-fd54c22d888b tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146679, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2260.596883] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146686, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2260.711666] env[62824]: DEBUG oslo_vmware.api [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Task: {'id': task-2146689, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2260.841800] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146685, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2260.864507] env[62824]: DEBUG nova.scheduler.client.report [None req-d6968317-d8b9-4e00-a6a9-ee3a22a078b2 tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2260.919442] env[62824]: DEBUG oslo_vmware.api [None req-e0a2c92a-cfc3-43db-a3f8-fd54c22d888b tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146679, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2261.001052] env[62824]: DEBUG nova.network.neutron [None req-5b20a72b-b561-4d0f-a3b8-8dea853cf384 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Updating instance_info_cache with network_info: [{"id": "d85a0acc-fed6-4797-9f81-2aafa5bb6967", "address": "fa:16:3e:d8:c8:ab", "network": {"id": "a410f4f3-d072-4034-b797-8210e1a79093", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1482574391-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.211", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e10208040df34eb5adfe8dcbc76043d5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f4399275-8e92-4448-be9e-d4984e93e89c", "external-id": "nsx-vlan-transportzone-192", "segmentation_id": 192, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd85a0acc-fe", "ovs_interfaceid": "d85a0acc-fed6-4797-9f81-2aafa5bb6967", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2261.094785] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146686, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2261.211724] env[62824]: DEBUG oslo_vmware.api [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Task: {'id': task-2146689, 'name': ReconfigVM_Task, 'duration_secs': 0.525969} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2261.212023] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] [instance: c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9] Reconfigured VM instance instance-0000005f to attach disk [datastore1] c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9/c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2261.212653] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-713a3682-4486-4db1-b310-0b11360bd7fb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2261.219028] env[62824]: DEBUG oslo_vmware.api [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Waiting for the task: (returnval){ [ 2261.219028] env[62824]: value = "task-2146691" [ 2261.219028] env[62824]: _type = "Task" [ 2261.219028] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2261.226605] env[62824]: DEBUG oslo_vmware.api [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Task: {'id': task-2146691, 'name': Rename_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2261.342997] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146685, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2261.369454] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d6968317-d8b9-4e00-a6a9-ee3a22a078b2 tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.822s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2261.371841] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.766s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2261.373344] env[62824]: INFO nova.compute.claims [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] [instance: 71731b21-c302-4d66-a579-889165673712] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2261.421204] env[62824]: DEBUG oslo_vmware.api [None req-e0a2c92a-cfc3-43db-a3f8-fd54c22d888b tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146679, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2261.504276] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5b20a72b-b561-4d0f-a3b8-8dea853cf384 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Releasing lock "refresh_cache-bbfcb2e3-9326-4548-b15b-e054cbfd192e" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2261.505199] env[62824]: DEBUG nova.objects.instance [None req-5b20a72b-b561-4d0f-a3b8-8dea853cf384 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Lazy-loading 'flavor' on Instance uuid bbfcb2e3-9326-4548-b15b-e054cbfd192e {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2261.598089] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146686, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2261.730107] env[62824]: DEBUG oslo_vmware.api [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Task: {'id': task-2146691, 'name': Rename_Task, 'duration_secs': 0.320924} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2261.730107] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] [instance: c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2261.730107] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b567f9d9-9db5-493d-b59d-ee54af3a0243 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2261.735898] env[62824]: DEBUG oslo_vmware.api [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Waiting for the task: (returnval){ [ 2261.735898] env[62824]: value = "task-2146692" [ 2261.735898] env[62824]: _type = "Task" [ 2261.735898] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2261.742952] env[62824]: DEBUG oslo_vmware.api [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Task: {'id': task-2146692, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2261.842975] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146685, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2261.890326] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d6968317-d8b9-4e00-a6a9-ee3a22a078b2 tempest-ServerActionsV293TestJSON-1121524247 tempest-ServerActionsV293TestJSON-1121524247-project-member] Lock "2f315f9e-94e4-47ef-9503-ee92b59e5452" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.755s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2261.921245] env[62824]: DEBUG oslo_vmware.api [None req-e0a2c92a-cfc3-43db-a3f8-fd54c22d888b tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146679, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2262.012041] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-282bb6b3-5fa0-4104-8eea-cb5681819e69 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2262.036871] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b20a72b-b561-4d0f-a3b8-8dea853cf384 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2262.038203] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-810d141d-cf1e-418d-ba2b-5ee43740105e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2262.044252] env[62824]: DEBUG oslo_vmware.api [None req-5b20a72b-b561-4d0f-a3b8-8dea853cf384 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for the task: (returnval){ [ 2262.044252] env[62824]: value = "task-2146693" [ 2262.044252] env[62824]: _type = "Task" [ 2262.044252] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2262.053526] env[62824]: DEBUG oslo_vmware.api [None req-5b20a72b-b561-4d0f-a3b8-8dea853cf384 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2146693, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2262.096065] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146686, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2262.246589] env[62824]: DEBUG oslo_vmware.api [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Task: {'id': task-2146692, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2262.343633] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146685, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2262.421638] env[62824]: DEBUG oslo_vmware.api [None req-e0a2c92a-cfc3-43db-a3f8-fd54c22d888b tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146679, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2262.553961] env[62824]: DEBUG oslo_vmware.api [None req-5b20a72b-b561-4d0f-a3b8-8dea853cf384 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2146693, 'name': PowerOffVM_Task, 'duration_secs': 0.233015} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2262.557729] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b20a72b-b561-4d0f-a3b8-8dea853cf384 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2262.563252] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-5b20a72b-b561-4d0f-a3b8-8dea853cf384 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Reconfiguring VM instance instance-00000039 to detach disk 2002 {{(pid=62824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 2262.563803] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-134757c7-9730-4b1f-b1f7-00c7f901e97a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2262.585112] env[62824]: DEBUG oslo_vmware.api [None req-5b20a72b-b561-4d0f-a3b8-8dea853cf384 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for the task: (returnval){ [ 2262.585112] env[62824]: value = "task-2146694" [ 2262.585112] env[62824]: _type = "Task" [ 2262.585112] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2262.600354] env[62824]: DEBUG oslo_vmware.api [None req-5b20a72b-b561-4d0f-a3b8-8dea853cf384 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2146694, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2262.603467] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146686, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2262.640891] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e08afc5a-ac78-42c5-a2f8-b097e3df8480 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2262.649069] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a38a9cb-6f81-412b-bca5-d00ebfb29bce {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2262.689665] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac20f18a-fe51-47b7-8da3-75c57bdb399a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2262.697972] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ab15e57-fa4c-4562-9113-50af31e5b6d7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2262.712176] env[62824]: DEBUG nova.compute.provider_tree [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2262.746941] env[62824]: DEBUG oslo_vmware.api [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Task: {'id': task-2146692, 'name': PowerOnVM_Task, 'duration_secs': 0.619314} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2262.747285] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] [instance: c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2262.747507] env[62824]: INFO nova.compute.manager [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] [instance: c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9] Took 9.43 seconds to spawn the instance on the hypervisor. [ 2262.747688] env[62824]: DEBUG nova.compute.manager [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] [instance: c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2262.748518] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c30a161c-671b-4288-a943-936fde6aa718 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2262.850035] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146685, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2262.922600] env[62824]: DEBUG oslo_vmware.api [None req-e0a2c92a-cfc3-43db-a3f8-fd54c22d888b tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146679, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2263.094581] env[62824]: DEBUG oslo_vmware.api [None req-5b20a72b-b561-4d0f-a3b8-8dea853cf384 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2146694, 'name': ReconfigVM_Task, 'duration_secs': 0.253809} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2263.097713] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-5b20a72b-b561-4d0f-a3b8-8dea853cf384 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Reconfigured VM instance instance-00000039 to detach disk 2002 {{(pid=62824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 2263.097913] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b20a72b-b561-4d0f-a3b8-8dea853cf384 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2263.098172] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-52f08800-7032-4dab-bffc-b121de6970d6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2263.104313] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146686, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2263.105400] env[62824]: DEBUG oslo_vmware.api [None req-5b20a72b-b561-4d0f-a3b8-8dea853cf384 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for the task: (returnval){ [ 2263.105400] env[62824]: value = "task-2146695" [ 2263.105400] env[62824]: _type = "Task" [ 2263.105400] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2263.112353] env[62824]: DEBUG oslo_concurrency.lockutils [None req-83ef04f5-6202-449e-9917-20354b7a27cd tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Acquiring lock "b9f30f87-3594-4468-9d29-70890d8761e3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2263.112578] env[62824]: DEBUG oslo_concurrency.lockutils [None req-83ef04f5-6202-449e-9917-20354b7a27cd tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Lock "b9f30f87-3594-4468-9d29-70890d8761e3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2263.112775] env[62824]: DEBUG oslo_concurrency.lockutils [None req-83ef04f5-6202-449e-9917-20354b7a27cd tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Acquiring lock "b9f30f87-3594-4468-9d29-70890d8761e3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2263.112956] env[62824]: DEBUG oslo_concurrency.lockutils [None req-83ef04f5-6202-449e-9917-20354b7a27cd tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Lock "b9f30f87-3594-4468-9d29-70890d8761e3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2263.113139] env[62824]: DEBUG oslo_concurrency.lockutils [None req-83ef04f5-6202-449e-9917-20354b7a27cd tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Lock "b9f30f87-3594-4468-9d29-70890d8761e3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2263.114608] env[62824]: DEBUG oslo_vmware.api [None req-5b20a72b-b561-4d0f-a3b8-8dea853cf384 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2146695, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2263.115047] env[62824]: INFO nova.compute.manager [None req-83ef04f5-6202-449e-9917-20354b7a27cd tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] Terminating instance [ 2263.215590] env[62824]: DEBUG nova.scheduler.client.report [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2263.268745] env[62824]: INFO nova.compute.manager [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] [instance: c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9] Took 15.45 seconds to build instance. [ 2263.346339] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146685, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2263.422869] env[62824]: DEBUG oslo_vmware.api [None req-e0a2c92a-cfc3-43db-a3f8-fd54c22d888b tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146679, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2263.604296] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146686, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2263.615591] env[62824]: DEBUG oslo_vmware.api [None req-5b20a72b-b561-4d0f-a3b8-8dea853cf384 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2146695, 'name': PowerOnVM_Task, 'duration_secs': 0.507594} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2263.615591] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b20a72b-b561-4d0f-a3b8-8dea853cf384 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2263.615591] env[62824]: DEBUG nova.compute.manager [None req-5b20a72b-b561-4d0f-a3b8-8dea853cf384 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2263.616397] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad40e676-f96b-432c-87ef-8a3432fe0723 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2263.619420] env[62824]: DEBUG nova.compute.manager [None req-83ef04f5-6202-449e-9917-20354b7a27cd tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2263.619527] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-83ef04f5-6202-449e-9917-20354b7a27cd tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2263.620311] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eeb2d4f1-dd1d-4c73-ad91-1aa5cf9f807e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2263.627388] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-83ef04f5-6202-449e-9917-20354b7a27cd tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2263.629027] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a82b0f8e-2ac2-442b-adfd-aff42b17749e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2263.638940] env[62824]: DEBUG oslo_vmware.api [None req-83ef04f5-6202-449e-9917-20354b7a27cd tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Waiting for the task: (returnval){ [ 2263.638940] env[62824]: value = "task-2146696" [ 2263.638940] env[62824]: _type = "Task" [ 2263.638940] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2263.650074] env[62824]: DEBUG oslo_vmware.api [None req-83ef04f5-6202-449e-9917-20354b7a27cd tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Task: {'id': task-2146696, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2263.725702] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.353s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2263.725702] env[62824]: DEBUG nova.compute.manager [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] [instance: 71731b21-c302-4d66-a579-889165673712] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2263.770767] env[62824]: DEBUG oslo_concurrency.lockutils [None req-250de2a3-9fc5-45dc-933a-eb3244a71f90 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Lock "c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.963s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2263.849709] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146685, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2263.923933] env[62824]: DEBUG oslo_vmware.api [None req-e0a2c92a-cfc3-43db-a3f8-fd54c22d888b tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146679, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2264.006588] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9b8d58d8-8420-4bcd-b5ae-0f85a3c51179 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Acquiring lock "c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2264.006871] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9b8d58d8-8420-4bcd-b5ae-0f85a3c51179 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Lock "c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2264.007131] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9b8d58d8-8420-4bcd-b5ae-0f85a3c51179 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Acquiring lock "c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2264.007511] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9b8d58d8-8420-4bcd-b5ae-0f85a3c51179 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Lock "c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2264.007712] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9b8d58d8-8420-4bcd-b5ae-0f85a3c51179 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Lock "c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2264.009923] env[62824]: INFO nova.compute.manager [None req-9b8d58d8-8420-4bcd-b5ae-0f85a3c51179 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] [instance: c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9] Terminating instance [ 2264.104801] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146686, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2264.148797] env[62824]: DEBUG oslo_vmware.api [None req-83ef04f5-6202-449e-9917-20354b7a27cd tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Task: {'id': task-2146696, 'name': PowerOffVM_Task, 'duration_secs': 0.294048} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2264.149082] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-83ef04f5-6202-449e-9917-20354b7a27cd tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2264.149291] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-83ef04f5-6202-449e-9917-20354b7a27cd tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2264.149539] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2925e8b2-5201-4a0a-9259-d3efd7bc715a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2264.230738] env[62824]: DEBUG nova.compute.utils [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2264.232264] env[62824]: DEBUG nova.compute.manager [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] [instance: 71731b21-c302-4d66-a579-889165673712] Not allocating networking since 'none' was specified. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1984}} [ 2264.348645] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146685, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2264.425244] env[62824]: DEBUG oslo_vmware.api [None req-e0a2c92a-cfc3-43db-a3f8-fd54c22d888b tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146679, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2264.514042] env[62824]: DEBUG nova.compute.manager [None req-9b8d58d8-8420-4bcd-b5ae-0f85a3c51179 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] [instance: c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2264.514296] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-9b8d58d8-8420-4bcd-b5ae-0f85a3c51179 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] [instance: c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2264.515215] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09c29f06-9f44-4994-ab5d-e57fe66360c0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2264.523054] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b8d58d8-8420-4bcd-b5ae-0f85a3c51179 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] [instance: c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2264.523293] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c753cf2b-9d53-469d-938c-98ba1f8ef34b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2264.531232] env[62824]: DEBUG oslo_vmware.api [None req-9b8d58d8-8420-4bcd-b5ae-0f85a3c51179 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Waiting for the task: (returnval){ [ 2264.531232] env[62824]: value = "task-2146698" [ 2264.531232] env[62824]: _type = "Task" [ 2264.531232] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2264.539259] env[62824]: DEBUG oslo_vmware.api [None req-9b8d58d8-8420-4bcd-b5ae-0f85a3c51179 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Task: {'id': task-2146698, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2264.605823] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146686, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2264.734213] env[62824]: DEBUG nova.compute.manager [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] [instance: 71731b21-c302-4d66-a579-889165673712] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2264.850102] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146685, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2264.897017] env[62824]: DEBUG nova.compute.manager [req-9f6c483b-5547-4f2a-b6e4-17bf8b97e1ce req-49dbf982-ca96-4bb6-ae37-8f4b09cbef6f service nova] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Received event network-changed-d85a0acc-fed6-4797-9f81-2aafa5bb6967 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2264.897189] env[62824]: DEBUG nova.compute.manager [req-9f6c483b-5547-4f2a-b6e4-17bf8b97e1ce req-49dbf982-ca96-4bb6-ae37-8f4b09cbef6f service nova] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Refreshing instance network info cache due to event network-changed-d85a0acc-fed6-4797-9f81-2aafa5bb6967. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2264.897423] env[62824]: DEBUG oslo_concurrency.lockutils [req-9f6c483b-5547-4f2a-b6e4-17bf8b97e1ce req-49dbf982-ca96-4bb6-ae37-8f4b09cbef6f service nova] Acquiring lock "refresh_cache-bbfcb2e3-9326-4548-b15b-e054cbfd192e" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2264.897574] env[62824]: DEBUG oslo_concurrency.lockutils [req-9f6c483b-5547-4f2a-b6e4-17bf8b97e1ce req-49dbf982-ca96-4bb6-ae37-8f4b09cbef6f service nova] Acquired lock "refresh_cache-bbfcb2e3-9326-4548-b15b-e054cbfd192e" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2264.897734] env[62824]: DEBUG nova.network.neutron [req-9f6c483b-5547-4f2a-b6e4-17bf8b97e1ce req-49dbf982-ca96-4bb6-ae37-8f4b09cbef6f service nova] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Refreshing network info cache for port d85a0acc-fed6-4797-9f81-2aafa5bb6967 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2264.915749] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2264.916217] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2264.927759] env[62824]: DEBUG oslo_vmware.api [None req-e0a2c92a-cfc3-43db-a3f8-fd54c22d888b tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146679, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2265.043291] env[62824]: DEBUG oslo_vmware.api [None req-9b8d58d8-8420-4bcd-b5ae-0f85a3c51179 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Task: {'id': task-2146698, 'name': PowerOffVM_Task, 'duration_secs': 0.242962} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2265.043511] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b8d58d8-8420-4bcd-b5ae-0f85a3c51179 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] [instance: c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2265.043679] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-9b8d58d8-8420-4bcd-b5ae-0f85a3c51179 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] [instance: c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2265.043928] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1e9aae96-c2a3-42ef-801b-4de97d0e101f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2265.106272] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146686, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2265.349807] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146685, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2265.429791] env[62824]: DEBUG oslo_vmware.api [None req-e0a2c92a-cfc3-43db-a3f8-fd54c22d888b tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146679, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2265.430739] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2265.430739] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Starting heal instance info cache {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10445}} [ 2265.611069] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146686, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2265.640032] env[62824]: DEBUG nova.network.neutron [req-9f6c483b-5547-4f2a-b6e4-17bf8b97e1ce req-49dbf982-ca96-4bb6-ae37-8f4b09cbef6f service nova] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Updated VIF entry in instance network info cache for port d85a0acc-fed6-4797-9f81-2aafa5bb6967. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2265.640640] env[62824]: DEBUG nova.network.neutron [req-9f6c483b-5547-4f2a-b6e4-17bf8b97e1ce req-49dbf982-ca96-4bb6-ae37-8f4b09cbef6f service nova] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Updating instance_info_cache with network_info: [{"id": "d85a0acc-fed6-4797-9f81-2aafa5bb6967", "address": "fa:16:3e:d8:c8:ab", "network": {"id": "a410f4f3-d072-4034-b797-8210e1a79093", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1482574391-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.211", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e10208040df34eb5adfe8dcbc76043d5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f4399275-8e92-4448-be9e-d4984e93e89c", "external-id": "nsx-vlan-transportzone-192", "segmentation_id": 192, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd85a0acc-fe", "ovs_interfaceid": "d85a0acc-fed6-4797-9f81-2aafa5bb6967", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2265.651348] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-83ef04f5-6202-449e-9917-20354b7a27cd tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2265.651820] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-83ef04f5-6202-449e-9917-20354b7a27cd tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] Deleting contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2265.652151] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-83ef04f5-6202-449e-9917-20354b7a27cd tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Deleting the datastore file [datastore1] b9f30f87-3594-4468-9d29-70890d8761e3 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2265.653092] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-46e0471b-1720-4c2b-b4a0-a39bd048029d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2265.660605] env[62824]: DEBUG oslo_vmware.api [None req-83ef04f5-6202-449e-9917-20354b7a27cd tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Waiting for the task: (returnval){ [ 2265.660605] env[62824]: value = "task-2146700" [ 2265.660605] env[62824]: _type = "Task" [ 2265.660605] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2265.669966] env[62824]: DEBUG oslo_vmware.api [None req-83ef04f5-6202-449e-9917-20354b7a27cd tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Task: {'id': task-2146700, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2265.715958] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-9b8d58d8-8420-4bcd-b5ae-0f85a3c51179 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] [instance: c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2265.715958] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-9b8d58d8-8420-4bcd-b5ae-0f85a3c51179 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] [instance: c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9] Deleting contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2265.716352] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b8d58d8-8420-4bcd-b5ae-0f85a3c51179 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Deleting the datastore file [datastore1] c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2265.716624] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-629d27b2-8400-4644-bf85-d8b64663eda6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2265.724220] env[62824]: DEBUG oslo_vmware.api [None req-9b8d58d8-8420-4bcd-b5ae-0f85a3c51179 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Waiting for the task: (returnval){ [ 2265.724220] env[62824]: value = "task-2146701" [ 2265.724220] env[62824]: _type = "Task" [ 2265.724220] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2265.733354] env[62824]: DEBUG oslo_vmware.api [None req-9b8d58d8-8420-4bcd-b5ae-0f85a3c51179 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Task: {'id': task-2146701, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2265.743835] env[62824]: DEBUG nova.compute.manager [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] [instance: 71731b21-c302-4d66-a579-889165673712] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2265.779268] env[62824]: DEBUG nova.virt.hardware [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2265.779741] env[62824]: DEBUG nova.virt.hardware [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2265.780010] env[62824]: DEBUG nova.virt.hardware [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2265.780467] env[62824]: DEBUG nova.virt.hardware [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2265.780657] env[62824]: DEBUG nova.virt.hardware [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2265.780860] env[62824]: DEBUG nova.virt.hardware [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2265.781208] env[62824]: DEBUG nova.virt.hardware [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2265.781472] env[62824]: DEBUG nova.virt.hardware [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2265.781687] env[62824]: DEBUG nova.virt.hardware [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2265.782071] env[62824]: DEBUG nova.virt.hardware [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2265.782290] env[62824]: DEBUG nova.virt.hardware [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2265.783357] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fd3ec77-bcf8-46c9-88e9-240edbb42931 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2265.793733] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-506ede65-d1c4-40f8-a0fe-e577290e3f06 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2265.812077] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] [instance: 71731b21-c302-4d66-a579-889165673712] Instance VIF info [] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2265.817859] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Creating folder: Project (3e7bf9b4e8a44c889522ff9c8297cb68). Parent ref: group-v438503. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 2265.818327] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5a58be32-7518-472a-88a5-32162d7f7e5d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2265.832138] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Created folder: Project (3e7bf9b4e8a44c889522ff9c8297cb68) in parent group-v438503. [ 2265.832401] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Creating folder: Instances. Parent ref: group-v438739. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 2265.832694] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c0dd8b7f-f801-4d89-bcc2-b55ff7bd268b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2265.850830] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146685, 'name': CreateVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2265.852155] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Created folder: Instances in parent group-v438739. [ 2265.852389] env[62824]: DEBUG oslo.service.loopingcall [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2265.852585] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 71731b21-c302-4d66-a579-889165673712] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2265.852810] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d46a7067-faf0-4f09-b4f9-833b76cd6de4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2265.870409] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2265.870409] env[62824]: value = "task-2146704" [ 2265.870409] env[62824]: _type = "Task" [ 2265.870409] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2265.878789] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146704, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2265.927419] env[62824]: DEBUG oslo_vmware.api [None req-e0a2c92a-cfc3-43db-a3f8-fd54c22d888b tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146679, 'name': ReconfigVM_Task, 'duration_secs': 9.448432} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2265.927809] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e0a2c92a-cfc3-43db-a3f8-fd54c22d888b tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Releasing lock "fc77c44a-180c-46ff-9690-9072c6213c91" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2265.928104] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e0a2c92a-cfc3-43db-a3f8-fd54c22d888b tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Reconfigured VM to detach interface {{(pid=62824) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 2266.108329] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146686, 'name': CreateVM_Task, 'duration_secs': 6.844933} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2266.108511] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3acdf86e-45e9-4353-a282-7272451389e7] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2266.109501] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2266.109702] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2266.110075] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2266.110390] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-abad3b61-4dde-464a-ad83-ef7745b09c76 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2266.116135] env[62824]: DEBUG oslo_vmware.api [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2266.116135] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]522b3805-be7e-0a58-c9ac-3da0db2901fe" [ 2266.116135] env[62824]: _type = "Task" [ 2266.116135] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2266.124976] env[62824]: DEBUG oslo_vmware.api [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]522b3805-be7e-0a58-c9ac-3da0db2901fe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2266.143892] env[62824]: DEBUG oslo_concurrency.lockutils [req-9f6c483b-5547-4f2a-b6e4-17bf8b97e1ce req-49dbf982-ca96-4bb6-ae37-8f4b09cbef6f service nova] Releasing lock "refresh_cache-bbfcb2e3-9326-4548-b15b-e054cbfd192e" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2266.170179] env[62824]: DEBUG oslo_vmware.api [None req-83ef04f5-6202-449e-9917-20354b7a27cd tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Task: {'id': task-2146700, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.168813} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2266.170474] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-83ef04f5-6202-449e-9917-20354b7a27cd tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2266.170681] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-83ef04f5-6202-449e-9917-20354b7a27cd tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] Deleted contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2266.171058] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-83ef04f5-6202-449e-9917-20354b7a27cd tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2266.171058] env[62824]: INFO nova.compute.manager [None req-83ef04f5-6202-449e-9917-20354b7a27cd tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] Took 2.55 seconds to destroy the instance on the hypervisor. [ 2266.171341] env[62824]: DEBUG oslo.service.loopingcall [None req-83ef04f5-6202-449e-9917-20354b7a27cd tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2266.171579] env[62824]: DEBUG nova.compute.manager [-] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2266.171720] env[62824]: DEBUG nova.network.neutron [-] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2266.233977] env[62824]: DEBUG oslo_vmware.api [None req-9b8d58d8-8420-4bcd-b5ae-0f85a3c51179 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Task: {'id': task-2146701, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.188648} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2266.236197] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b8d58d8-8420-4bcd-b5ae-0f85a3c51179 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2266.236393] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-9b8d58d8-8420-4bcd-b5ae-0f85a3c51179 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] [instance: c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9] Deleted contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2266.236570] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-9b8d58d8-8420-4bcd-b5ae-0f85a3c51179 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] [instance: c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2266.236740] env[62824]: INFO nova.compute.manager [None req-9b8d58d8-8420-4bcd-b5ae-0f85a3c51179 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] [instance: c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9] Took 1.72 seconds to destroy the instance on the hypervisor. [ 2266.236974] env[62824]: DEBUG oslo.service.loopingcall [None req-9b8d58d8-8420-4bcd-b5ae-0f85a3c51179 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2266.237184] env[62824]: DEBUG nova.compute.manager [-] [instance: c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2266.237295] env[62824]: DEBUG nova.network.neutron [-] [instance: c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2266.350051] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146685, 'name': CreateVM_Task, 'duration_secs': 8.138691} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2266.350227] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2266.350862] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2266.381495] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146704, 'name': CreateVM_Task, 'duration_secs': 0.312796} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2266.381495] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 71731b21-c302-4d66-a579-889165673712] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2266.381495] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2266.627894] env[62824]: DEBUG oslo_vmware.api [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]522b3805-be7e-0a58-c9ac-3da0db2901fe, 'name': SearchDatastore_Task, 'duration_secs': 0.009144} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2266.628280] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2266.628564] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 3acdf86e-45e9-4353-a282-7272451389e7] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2266.628911] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2266.629127] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2266.629396] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2266.629726] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2266.630428] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2266.630428] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b244a0c5-c88c-4a34-a622-dc8f6dc42104 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2266.633645] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3499aff7-c812-4375-b360-92dce106fc07 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2266.641119] env[62824]: DEBUG oslo_vmware.api [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2266.641119] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]526192a8-5350-57d1-b92c-7c875886e929" [ 2266.641119] env[62824]: _type = "Task" [ 2266.641119] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2266.644821] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2266.645053] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2266.649114] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d62a4cec-d135-4df8-ace6-75886f4d0dc7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2266.651865] env[62824]: DEBUG oslo_vmware.api [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]526192a8-5350-57d1-b92c-7c875886e929, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2266.655202] env[62824]: DEBUG oslo_vmware.api [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2266.655202] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52b2b3bd-4ff1-fc11-bbb8-2cdfb9516f55" [ 2266.655202] env[62824]: _type = "Task" [ 2266.655202] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2266.663521] env[62824]: DEBUG oslo_vmware.api [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52b2b3bd-4ff1-fc11-bbb8-2cdfb9516f55, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2266.888284] env[62824]: DEBUG nova.compute.manager [req-32ee7b84-1112-404a-8337-b83f371c4eef req-6e242573-9123-4732-9c52-7f2b349c2ed1 service nova] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Received event network-vif-deleted-9e4a9375-566d-4c66-8165-2d01b62d9a15 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2266.888419] env[62824]: INFO nova.compute.manager [req-32ee7b84-1112-404a-8337-b83f371c4eef req-6e242573-9123-4732-9c52-7f2b349c2ed1 service nova] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Neutron deleted interface 9e4a9375-566d-4c66-8165-2d01b62d9a15; detaching it from the instance and deleting it from the info cache [ 2266.888675] env[62824]: DEBUG nova.network.neutron [req-32ee7b84-1112-404a-8337-b83f371c4eef req-6e242573-9123-4732-9c52-7f2b349c2ed1 service nova] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Updating instance_info_cache with network_info: [{"id": "e285a842-d54b-4614-84ea-239e34f5b41f", "address": "fa:16:3e:9d:17:cd", "network": {"id": "aa1c09e5-8d72-43ad-a903-a0f5e711da80", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1643616572-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c987c5f844be42c99cc9d57a9ca84c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9bc2632-36f9-4912-8782-8bbb789f909d", "external-id": "nsx-vlan-transportzone-897", "segmentation_id": 897, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape285a842-d5", "ovs_interfaceid": "e285a842-d54b-4614-84ea-239e34f5b41f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "2b6381b4-f3c7-4e22-b698-de190c8199ee", "address": "fa:16:3e:f4:8a:c1", "network": {"id": "aa1c09e5-8d72-43ad-a903-a0f5e711da80", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1643616572-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c987c5f844be42c99cc9d57a9ca84c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9bc2632-36f9-4912-8782-8bbb789f909d", "external-id": "nsx-vlan-transportzone-897", "segmentation_id": 897, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b6381b4-f3", "ovs_interfaceid": "2b6381b4-f3c7-4e22-b698-de190c8199ee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2266.947265] env[62824]: DEBUG nova.compute.manager [req-173a28c0-4b70-47fe-a51f-32b96501007c req-855fc076-b8d3-449d-ad77-3a00b5b589f2 service nova] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Received event network-changed-d85a0acc-fed6-4797-9f81-2aafa5bb6967 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2266.947586] env[62824]: DEBUG nova.compute.manager [req-173a28c0-4b70-47fe-a51f-32b96501007c req-855fc076-b8d3-449d-ad77-3a00b5b589f2 service nova] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Refreshing instance network info cache due to event network-changed-d85a0acc-fed6-4797-9f81-2aafa5bb6967. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2266.947883] env[62824]: DEBUG oslo_concurrency.lockutils [req-173a28c0-4b70-47fe-a51f-32b96501007c req-855fc076-b8d3-449d-ad77-3a00b5b589f2 service nova] Acquiring lock "refresh_cache-bbfcb2e3-9326-4548-b15b-e054cbfd192e" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2266.948208] env[62824]: DEBUG oslo_concurrency.lockutils [req-173a28c0-4b70-47fe-a51f-32b96501007c req-855fc076-b8d3-449d-ad77-3a00b5b589f2 service nova] Acquired lock "refresh_cache-bbfcb2e3-9326-4548-b15b-e054cbfd192e" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2266.948284] env[62824]: DEBUG nova.network.neutron [req-173a28c0-4b70-47fe-a51f-32b96501007c req-855fc076-b8d3-449d-ad77-3a00b5b589f2 service nova] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Refreshing network info cache for port d85a0acc-fed6-4797-9f81-2aafa5bb6967 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2266.957772] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e89a221d-5c30-48d4-bfc0-a91f11a74aac tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 204b866f-e9c5-441d-bf8e-77d324b39ecd] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2266.958350] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e89a221d-5c30-48d4-bfc0-a91f11a74aac tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 204b866f-e9c5-441d-bf8e-77d324b39ecd] Deleting contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2266.959257] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-e89a221d-5c30-48d4-bfc0-a91f11a74aac tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Deleting the datastore file [datastore2] 204b866f-e9c5-441d-bf8e-77d324b39ecd {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2266.959875] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f54eea84-5854-4dae-a7f7-28e0bfbcb36e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2266.968279] env[62824]: DEBUG oslo_vmware.api [None req-e89a221d-5c30-48d4-bfc0-a91f11a74aac tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 2266.968279] env[62824]: value = "task-2146705" [ 2266.968279] env[62824]: _type = "Task" [ 2266.968279] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2266.976697] env[62824]: DEBUG oslo_vmware.api [None req-e89a221d-5c30-48d4-bfc0-a91f11a74aac tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146705, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2267.136728] env[62824]: DEBUG nova.network.neutron [-] [instance: c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2267.150180] env[62824]: DEBUG oslo_vmware.api [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]526192a8-5350-57d1-b92c-7c875886e929, 'name': SearchDatastore_Task, 'duration_secs': 0.011694} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2267.151094] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2267.151343] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2267.151555] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2267.151796] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2267.152108] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2267.152615] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-468d204e-c5ec-40b1-a914-9d48727ab857 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2267.160412] env[62824]: DEBUG oslo_vmware.api [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Waiting for the task: (returnval){ [ 2267.160412] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52d36dcb-55f7-3a93-ce1d-ed4357f116dc" [ 2267.160412] env[62824]: _type = "Task" [ 2267.160412] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2267.167929] env[62824]: DEBUG oslo_vmware.api [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52b2b3bd-4ff1-fc11-bbb8-2cdfb9516f55, 'name': SearchDatastore_Task, 'duration_secs': 0.009217} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2267.169120] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-406efefd-d2c4-4f24-bb0e-21cb73408df0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2267.174652] env[62824]: DEBUG oslo_vmware.api [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52d36dcb-55f7-3a93-ce1d-ed4357f116dc, 'name': SearchDatastore_Task, 'duration_secs': 0.008899} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2267.175426] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2267.175679] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] [instance: 71731b21-c302-4d66-a579-889165673712] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2267.175889] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2267.178097] env[62824]: DEBUG oslo_vmware.api [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2267.178097] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52b6ae31-4cf6-cc81-84ed-ce0a5e072b7f" [ 2267.178097] env[62824]: _type = "Task" [ 2267.178097] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2267.181359] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e0a2c92a-cfc3-43db-a3f8-fd54c22d888b tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquiring lock "refresh_cache-fc77c44a-180c-46ff-9690-9072c6213c91" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2267.181584] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e0a2c92a-cfc3-43db-a3f8-fd54c22d888b tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquired lock "refresh_cache-fc77c44a-180c-46ff-9690-9072c6213c91" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2267.181797] env[62824]: DEBUG nova.network.neutron [None req-e0a2c92a-cfc3-43db-a3f8-fd54c22d888b tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2267.188746] env[62824]: DEBUG oslo_vmware.api [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52b6ae31-4cf6-cc81-84ed-ce0a5e072b7f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2267.261339] env[62824]: DEBUG nova.network.neutron [-] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2267.356506] env[62824]: DEBUG nova.compute.manager [req-beae8d74-755e-47bc-af15-b0b6485753f2 req-4da79f1d-2d49-4147-b3b0-ca56239d8ec2 service nova] [instance: 815bd7ae-f461-4c74-9401-7c5fe43679ee] Received event network-vif-plugged-3cb13172-9b03-4150-8070-b5b5f504d942 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2267.356793] env[62824]: DEBUG oslo_concurrency.lockutils [req-beae8d74-755e-47bc-af15-b0b6485753f2 req-4da79f1d-2d49-4147-b3b0-ca56239d8ec2 service nova] Acquiring lock "815bd7ae-f461-4c74-9401-7c5fe43679ee-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2267.356924] env[62824]: DEBUG oslo_concurrency.lockutils [req-beae8d74-755e-47bc-af15-b0b6485753f2 req-4da79f1d-2d49-4147-b3b0-ca56239d8ec2 service nova] Lock "815bd7ae-f461-4c74-9401-7c5fe43679ee-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2267.357114] env[62824]: DEBUG oslo_concurrency.lockutils [req-beae8d74-755e-47bc-af15-b0b6485753f2 req-4da79f1d-2d49-4147-b3b0-ca56239d8ec2 service nova] Lock "815bd7ae-f461-4c74-9401-7c5fe43679ee-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2267.357358] env[62824]: DEBUG nova.compute.manager [req-beae8d74-755e-47bc-af15-b0b6485753f2 req-4da79f1d-2d49-4147-b3b0-ca56239d8ec2 service nova] [instance: 815bd7ae-f461-4c74-9401-7c5fe43679ee] No waiting events found dispatching network-vif-plugged-3cb13172-9b03-4150-8070-b5b5f504d942 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2267.357550] env[62824]: WARNING nova.compute.manager [req-beae8d74-755e-47bc-af15-b0b6485753f2 req-4da79f1d-2d49-4147-b3b0-ca56239d8ec2 service nova] [instance: 815bd7ae-f461-4c74-9401-7c5fe43679ee] Received unexpected event network-vif-plugged-3cb13172-9b03-4150-8070-b5b5f504d942 for instance with vm_state building and task_state spawning. [ 2267.395111] env[62824]: DEBUG oslo_concurrency.lockutils [req-32ee7b84-1112-404a-8337-b83f371c4eef req-6e242573-9123-4732-9c52-7f2b349c2ed1 service nova] Acquiring lock "fc77c44a-180c-46ff-9690-9072c6213c91" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2267.395297] env[62824]: DEBUG oslo_concurrency.lockutils [req-32ee7b84-1112-404a-8337-b83f371c4eef req-6e242573-9123-4732-9c52-7f2b349c2ed1 service nova] Acquired lock "fc77c44a-180c-46ff-9690-9072c6213c91" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2267.396582] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb7831d0-c875-4ec9-bf9a-7d7969bb0536 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2267.416645] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-daff8fea-554e-436d-b2ee-6d12013a7ce2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2267.446392] env[62824]: DEBUG nova.virt.vmwareapi.vmops [req-32ee7b84-1112-404a-8337-b83f371c4eef req-6e242573-9123-4732-9c52-7f2b349c2ed1 service nova] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Reconfiguring VM to detach interface {{(pid=62824) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 2267.446516] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-489eebdd-a15d-4d84-85af-44e3035d6202 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2267.459642] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8dadfd7c-4da2-4b7c-97a3-d1039115d4f4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquiring lock "fc77c44a-180c-46ff-9690-9072c6213c91" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2267.469906] env[62824]: DEBUG oslo_vmware.api [req-32ee7b84-1112-404a-8337-b83f371c4eef req-6e242573-9123-4732-9c52-7f2b349c2ed1 service nova] Waiting for the task: (returnval){ [ 2267.469906] env[62824]: value = "task-2146706" [ 2267.469906] env[62824]: _type = "Task" [ 2267.469906] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2267.484647] env[62824]: DEBUG oslo_vmware.api [None req-e89a221d-5c30-48d4-bfc0-a91f11a74aac tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146705, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.13401} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2267.484869] env[62824]: DEBUG oslo_vmware.api [req-32ee7b84-1112-404a-8337-b83f371c4eef req-6e242573-9123-4732-9c52-7f2b349c2ed1 service nova] Task: {'id': task-2146706, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2267.485213] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-e89a221d-5c30-48d4-bfc0-a91f11a74aac tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2267.485285] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e89a221d-5c30-48d4-bfc0-a91f11a74aac tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 204b866f-e9c5-441d-bf8e-77d324b39ecd] Deleted contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2267.485469] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e89a221d-5c30-48d4-bfc0-a91f11a74aac tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 204b866f-e9c5-441d-bf8e-77d324b39ecd] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2267.485615] env[62824]: INFO nova.compute.manager [None req-e89a221d-5c30-48d4-bfc0-a91f11a74aac tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 204b866f-e9c5-441d-bf8e-77d324b39ecd] Took 9.89 seconds to destroy the instance on the hypervisor. [ 2267.485854] env[62824]: DEBUG oslo.service.loopingcall [None req-e89a221d-5c30-48d4-bfc0-a91f11a74aac tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2267.486104] env[62824]: DEBUG nova.compute.manager [-] [instance: 204b866f-e9c5-441d-bf8e-77d324b39ecd] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2267.486176] env[62824]: DEBUG nova.network.neutron [-] [instance: 204b866f-e9c5-441d-bf8e-77d324b39ecd] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2267.639142] env[62824]: INFO nova.compute.manager [-] [instance: c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9] Took 1.40 seconds to deallocate network for instance. [ 2267.690245] env[62824]: DEBUG oslo_vmware.api [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52b6ae31-4cf6-cc81-84ed-ce0a5e072b7f, 'name': SearchDatastore_Task, 'duration_secs': 0.010567} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2267.693493] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2267.693493] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 3acdf86e-45e9-4353-a282-7272451389e7/3acdf86e-45e9-4353-a282-7272451389e7.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2267.693493] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2267.693699] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2267.693793] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6d9b7fe1-4ffe-4593-9e13-a7ee37348a90 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2267.696043] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8b8c184f-24af-409c-adfd-1d7e47b0f083 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2267.704017] env[62824]: DEBUG oslo_vmware.api [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2267.704017] env[62824]: value = "task-2146707" [ 2267.704017] env[62824]: _type = "Task" [ 2267.704017] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2267.706292] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2267.706447] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2267.707378] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f03d40d2-5a83-45ea-9dc2-b22c94643676 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2267.711910] env[62824]: DEBUG oslo_vmware.api [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146707, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2267.714896] env[62824]: DEBUG oslo_vmware.api [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2267.714896] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52c8e56e-c872-541a-fa6e-746e5677e565" [ 2267.714896] env[62824]: _type = "Task" [ 2267.714896] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2267.722035] env[62824]: DEBUG oslo_vmware.api [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52c8e56e-c872-541a-fa6e-746e5677e565, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2267.764583] env[62824]: INFO nova.compute.manager [-] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] Took 1.59 seconds to deallocate network for instance. [ 2267.798542] env[62824]: DEBUG nova.network.neutron [req-173a28c0-4b70-47fe-a51f-32b96501007c req-855fc076-b8d3-449d-ad77-3a00b5b589f2 service nova] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Updated VIF entry in instance network info cache for port d85a0acc-fed6-4797-9f81-2aafa5bb6967. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2267.798975] env[62824]: DEBUG nova.network.neutron [req-173a28c0-4b70-47fe-a51f-32b96501007c req-855fc076-b8d3-449d-ad77-3a00b5b589f2 service nova] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Updating instance_info_cache with network_info: [{"id": "d85a0acc-fed6-4797-9f81-2aafa5bb6967", "address": "fa:16:3e:d8:c8:ab", "network": {"id": "a410f4f3-d072-4034-b797-8210e1a79093", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1482574391-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.211", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e10208040df34eb5adfe8dcbc76043d5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f4399275-8e92-4448-be9e-d4984e93e89c", "external-id": "nsx-vlan-transportzone-192", "segmentation_id": 192, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd85a0acc-fe", "ovs_interfaceid": "d85a0acc-fed6-4797-9f81-2aafa5bb6967", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2267.866930] env[62824]: DEBUG nova.network.neutron [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 815bd7ae-f461-4c74-9401-7c5fe43679ee] Successfully updated port: 3cb13172-9b03-4150-8070-b5b5f504d942 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2267.956985] env[62824]: INFO nova.network.neutron [None req-e0a2c92a-cfc3-43db-a3f8-fd54c22d888b tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Port 2b6381b4-f3c7-4e22-b698-de190c8199ee from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 2267.957410] env[62824]: DEBUG nova.network.neutron [None req-e0a2c92a-cfc3-43db-a3f8-fd54c22d888b tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Updating instance_info_cache with network_info: [{"id": "e285a842-d54b-4614-84ea-239e34f5b41f", "address": "fa:16:3e:9d:17:cd", "network": {"id": "aa1c09e5-8d72-43ad-a903-a0f5e711da80", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1643616572-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c987c5f844be42c99cc9d57a9ca84c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9bc2632-36f9-4912-8782-8bbb789f909d", "external-id": "nsx-vlan-transportzone-897", "segmentation_id": 897, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape285a842-d5", "ovs_interfaceid": "e285a842-d54b-4614-84ea-239e34f5b41f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2267.965262] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: b8cc8cd3-ea03-40bf-b867-7ad193365552] Skipping network cache update for instance because it has been migrated to another host. {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10489}} [ 2267.985806] env[62824]: DEBUG oslo_vmware.api [req-32ee7b84-1112-404a-8337-b83f371c4eef req-6e242573-9123-4732-9c52-7f2b349c2ed1 service nova] Task: {'id': task-2146706, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2268.148464] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9b8d58d8-8420-4bcd-b5ae-0f85a3c51179 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2268.148464] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9b8d58d8-8420-4bcd-b5ae-0f85a3c51179 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2268.148464] env[62824]: DEBUG nova.objects.instance [None req-9b8d58d8-8420-4bcd-b5ae-0f85a3c51179 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Lazy-loading 'resources' on Instance uuid c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2268.212270] env[62824]: DEBUG oslo_vmware.api [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146707, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.498819} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2268.212534] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 3acdf86e-45e9-4353-a282-7272451389e7/3acdf86e-45e9-4353-a282-7272451389e7.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2268.212749] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 3acdf86e-45e9-4353-a282-7272451389e7] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2268.212999] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8933d626-4db5-4c36-b9dc-524379703aad {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2268.219791] env[62824]: DEBUG oslo_vmware.api [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2268.219791] env[62824]: value = "task-2146708" [ 2268.219791] env[62824]: _type = "Task" [ 2268.219791] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2268.226300] env[62824]: DEBUG oslo_vmware.api [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52c8e56e-c872-541a-fa6e-746e5677e565, 'name': SearchDatastore_Task, 'duration_secs': 0.00966} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2268.227440] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-87c03d08-b0f3-4eb0-9605-04d32ca978a2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2268.232459] env[62824]: DEBUG oslo_vmware.api [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146708, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2268.235682] env[62824]: DEBUG oslo_vmware.api [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2268.235682] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]521db484-a331-4e39-7f95-3ddbf3e9498f" [ 2268.235682] env[62824]: _type = "Task" [ 2268.235682] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2268.243945] env[62824]: DEBUG oslo_vmware.api [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]521db484-a331-4e39-7f95-3ddbf3e9498f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2268.271149] env[62824]: DEBUG oslo_concurrency.lockutils [None req-83ef04f5-6202-449e-9917-20354b7a27cd tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2268.301398] env[62824]: DEBUG oslo_concurrency.lockutils [req-173a28c0-4b70-47fe-a51f-32b96501007c req-855fc076-b8d3-449d-ad77-3a00b5b589f2 service nova] Releasing lock "refresh_cache-bbfcb2e3-9326-4548-b15b-e054cbfd192e" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2268.301670] env[62824]: DEBUG nova.compute.manager [req-173a28c0-4b70-47fe-a51f-32b96501007c req-855fc076-b8d3-449d-ad77-3a00b5b589f2 service nova] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Received event network-vif-deleted-2b6381b4-f3c7-4e22-b698-de190c8199ee {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2268.301847] env[62824]: INFO nova.compute.manager [req-173a28c0-4b70-47fe-a51f-32b96501007c req-855fc076-b8d3-449d-ad77-3a00b5b589f2 service nova] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Neutron deleted interface 2b6381b4-f3c7-4e22-b698-de190c8199ee; detaching it from the instance and deleting it from the info cache [ 2268.302120] env[62824]: DEBUG nova.network.neutron [req-173a28c0-4b70-47fe-a51f-32b96501007c req-855fc076-b8d3-449d-ad77-3a00b5b589f2 service nova] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Updating instance_info_cache with network_info: [{"id": "e285a842-d54b-4614-84ea-239e34f5b41f", "address": "fa:16:3e:9d:17:cd", "network": {"id": "aa1c09e5-8d72-43ad-a903-a0f5e711da80", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1643616572-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c987c5f844be42c99cc9d57a9ca84c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9bc2632-36f9-4912-8782-8bbb789f909d", "external-id": "nsx-vlan-transportzone-897", "segmentation_id": 897, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape285a842-d5", "ovs_interfaceid": "e285a842-d54b-4614-84ea-239e34f5b41f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2268.370755] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Acquiring lock "refresh_cache-815bd7ae-f461-4c74-9401-7c5fe43679ee" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2268.370885] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Acquired lock "refresh_cache-815bd7ae-f461-4c74-9401-7c5fe43679ee" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2268.371053] env[62824]: DEBUG nova.network.neutron [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 815bd7ae-f461-4c74-9401-7c5fe43679ee] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2268.372220] env[62824]: DEBUG nova.network.neutron [-] [instance: 204b866f-e9c5-441d-bf8e-77d324b39ecd] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2268.460427] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e0a2c92a-cfc3-43db-a3f8-fd54c22d888b tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Releasing lock "refresh_cache-fc77c44a-180c-46ff-9690-9072c6213c91" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2268.468268] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Didn't find any instances for network info cache update. {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10531}} [ 2268.468416] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2268.468574] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2268.468724] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2268.468871] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2268.469021] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2268.469164] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2268.469439] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62824) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11064}} [ 2268.469670] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager.update_available_resource {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2268.485591] env[62824]: DEBUG oslo_vmware.api [req-32ee7b84-1112-404a-8337-b83f371c4eef req-6e242573-9123-4732-9c52-7f2b349c2ed1 service nova] Task: {'id': task-2146706, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2268.663014] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c5fd25f6-e75a-454a-99ae-93fa0aabd927 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquiring lock "5b3df31f-15fe-473f-992c-ddb272661c53" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2268.663226] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c5fd25f6-e75a-454a-99ae-93fa0aabd927 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lock "5b3df31f-15fe-473f-992c-ddb272661c53" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2268.663452] env[62824]: INFO nova.compute.manager [None req-c5fd25f6-e75a-454a-99ae-93fa0aabd927 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Rebooting instance [ 2268.731307] env[62824]: DEBUG oslo_vmware.api [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146708, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.311943} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2268.734509] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 3acdf86e-45e9-4353-a282-7272451389e7] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2268.735249] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-301c46a4-6bec-41ac-bd1c-12ecc21b04a9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2268.757280] env[62824]: DEBUG oslo_vmware.api [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]521db484-a331-4e39-7f95-3ddbf3e9498f, 'name': SearchDatastore_Task, 'duration_secs': 0.011913} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2268.766095] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 3acdf86e-45e9-4353-a282-7272451389e7] Reconfiguring VM instance instance-00000060 to attach disk [datastore1] 3acdf86e-45e9-4353-a282-7272451389e7/3acdf86e-45e9-4353-a282-7272451389e7.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2268.768833] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2268.769112] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 0394d368-c846-4f3b-bfcf-2bc8e858052a/0394d368-c846-4f3b-bfcf-2bc8e858052a.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2268.769389] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b22f2700-22df-49ba-a3c1-1c3b118a93c7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2268.783798] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2268.784010] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2268.784242] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5592c9f7-7b05-4dde-bd8c-6016dd057f9d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2268.786324] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f0c64711-8397-40a2-8943-93c595e9453d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2268.793403] env[62824]: DEBUG oslo_vmware.api [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2268.793403] env[62824]: value = "task-2146710" [ 2268.793403] env[62824]: _type = "Task" [ 2268.793403] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2268.796931] env[62824]: DEBUG oslo_vmware.api [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2268.796931] env[62824]: value = "task-2146709" [ 2268.796931] env[62824]: _type = "Task" [ 2268.796931] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2268.802110] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2268.802110] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2268.804794] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5b13ed98-a12a-4e6e-9c99-1f33b8fb779b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2268.808430] env[62824]: DEBUG oslo_concurrency.lockutils [req-173a28c0-4b70-47fe-a51f-32b96501007c req-855fc076-b8d3-449d-ad77-3a00b5b589f2 service nova] Acquiring lock "fc77c44a-180c-46ff-9690-9072c6213c91" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2268.818059] env[62824]: DEBUG oslo_vmware.api [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146709, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2268.818407] env[62824]: DEBUG oslo_vmware.api [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Waiting for the task: (returnval){ [ 2268.818407] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52aafbf9-1b78-b3cb-9d0d-26f64698835e" [ 2268.818407] env[62824]: _type = "Task" [ 2268.818407] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2268.818622] env[62824]: DEBUG oslo_vmware.api [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146710, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2268.827342] env[62824]: DEBUG oslo_vmware.api [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52aafbf9-1b78-b3cb-9d0d-26f64698835e, 'name': SearchDatastore_Task, 'duration_secs': 0.011888} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2268.831586] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0293e075-5dce-4529-b4f1-677d4e913cc8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2268.835853] env[62824]: DEBUG oslo_vmware.api [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Waiting for the task: (returnval){ [ 2268.835853] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52a5f61b-eaa9-31f7-73d6-7d56b7098212" [ 2268.835853] env[62824]: _type = "Task" [ 2268.835853] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2268.845941] env[62824]: DEBUG oslo_vmware.api [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52a5f61b-eaa9-31f7-73d6-7d56b7098212, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2268.878220] env[62824]: INFO nova.compute.manager [-] [instance: 204b866f-e9c5-441d-bf8e-77d324b39ecd] Took 1.39 seconds to deallocate network for instance. [ 2268.933954] env[62824]: DEBUG nova.compute.manager [req-6aff378b-18e6-4d29-bd16-d368b66b6970 req-26f1a76e-a608-483d-876d-d0a746054819 service nova] [instance: 815bd7ae-f461-4c74-9401-7c5fe43679ee] Received event network-changed-3cb13172-9b03-4150-8070-b5b5f504d942 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2268.934073] env[62824]: DEBUG nova.compute.manager [req-6aff378b-18e6-4d29-bd16-d368b66b6970 req-26f1a76e-a608-483d-876d-d0a746054819 service nova] [instance: 815bd7ae-f461-4c74-9401-7c5fe43679ee] Refreshing instance network info cache due to event network-changed-3cb13172-9b03-4150-8070-b5b5f504d942. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2268.934376] env[62824]: DEBUG oslo_concurrency.lockutils [req-6aff378b-18e6-4d29-bd16-d368b66b6970 req-26f1a76e-a608-483d-876d-d0a746054819 service nova] Acquiring lock "refresh_cache-815bd7ae-f461-4c74-9401-7c5fe43679ee" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2268.935159] env[62824]: DEBUG nova.network.neutron [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 815bd7ae-f461-4c74-9401-7c5fe43679ee] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2268.958233] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfd3e11a-258e-44d2-b34f-62d0c682f50b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2268.965826] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e0a2c92a-cfc3-43db-a3f8-fd54c22d888b tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lock "interface-fc77c44a-180c-46ff-9690-9072c6213c91-2b6381b4-f3c7-4e22-b698-de190c8199ee" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 13.145s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2268.967865] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b43f8641-c782-4e06-ab55-8d7d3eec32ea {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2268.973198] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2269.012570] env[62824]: DEBUG nova.compute.manager [req-0eef4275-ecfd-4ae6-8133-a61dd286208f req-4e97e9d8-c626-4cc0-b1e9-7205cdc23426 service nova] [instance: 204b866f-e9c5-441d-bf8e-77d324b39ecd] Received event network-vif-deleted-e41f4cd6-ab65-4a8d-8514-3be4cde1915b {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2269.013923] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2d78910-d4c5-4059-a8cb-78c0de4e7775 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2269.021725] env[62824]: DEBUG oslo_vmware.api [req-32ee7b84-1112-404a-8337-b83f371c4eef req-6e242573-9123-4732-9c52-7f2b349c2ed1 service nova] Task: {'id': task-2146706, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2269.027786] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-373fa248-d10b-4215-9ddc-e9e56d0aba99 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2269.043849] env[62824]: DEBUG nova.compute.provider_tree [None req-9b8d58d8-8420-4bcd-b5ae-0f85a3c51179 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2269.149178] env[62824]: DEBUG nova.network.neutron [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 815bd7ae-f461-4c74-9401-7c5fe43679ee] Updating instance_info_cache with network_info: [{"id": "3cb13172-9b03-4150-8070-b5b5f504d942", "address": "fa:16:3e:39:e5:73", "network": {"id": "fc5d7128-fe9b-44ef-9d91-839190a97226", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-796405330-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35b7289f4746463e9b1cc47fe914a835", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0954fad3-d24d-496c-83e6-a09d3cb556fc", "external-id": "nsx-vlan-transportzone-216", "segmentation_id": 216, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3cb13172-9b", "ovs_interfaceid": "3cb13172-9b03-4150-8070-b5b5f504d942", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2269.180902] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c5fd25f6-e75a-454a-99ae-93fa0aabd927 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquiring lock "refresh_cache-5b3df31f-15fe-473f-992c-ddb272661c53" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2269.181112] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c5fd25f6-e75a-454a-99ae-93fa0aabd927 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquired lock "refresh_cache-5b3df31f-15fe-473f-992c-ddb272661c53" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2269.181294] env[62824]: DEBUG nova.network.neutron [None req-c5fd25f6-e75a-454a-99ae-93fa0aabd927 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2269.309712] env[62824]: DEBUG oslo_vmware.api [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146710, 'name': ReconfigVM_Task, 'duration_secs': 0.32885} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2269.312991] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 3acdf86e-45e9-4353-a282-7272451389e7] Reconfigured VM instance instance-00000060 to attach disk [datastore1] 3acdf86e-45e9-4353-a282-7272451389e7/3acdf86e-45e9-4353-a282-7272451389e7.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2269.313663] env[62824]: DEBUG oslo_vmware.api [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146709, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.481325} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2269.313871] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6a5721fa-af3d-4e13-a677-5a110f21030d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2269.315381] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 0394d368-c846-4f3b-bfcf-2bc8e858052a/0394d368-c846-4f3b-bfcf-2bc8e858052a.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2269.315598] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2269.315824] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e843e043-51c9-4bd1-898e-58b858289ea5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2269.324801] env[62824]: DEBUG oslo_vmware.api [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2269.324801] env[62824]: value = "task-2146712" [ 2269.324801] env[62824]: _type = "Task" [ 2269.324801] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2269.326057] env[62824]: DEBUG oslo_vmware.api [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2269.326057] env[62824]: value = "task-2146711" [ 2269.326057] env[62824]: _type = "Task" [ 2269.326057] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2269.337041] env[62824]: DEBUG oslo_vmware.api [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146712, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2269.343187] env[62824]: DEBUG oslo_vmware.api [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146711, 'name': Rename_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2269.348201] env[62824]: DEBUG oslo_vmware.api [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52a5f61b-eaa9-31f7-73d6-7d56b7098212, 'name': SearchDatastore_Task, 'duration_secs': 0.010533} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2269.348453] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2269.348716] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 71731b21-c302-4d66-a579-889165673712/71731b21-c302-4d66-a579-889165673712.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2269.349562] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c0800b1e-6412-4539-bb04-1ce4bda7059f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2269.355477] env[62824]: DEBUG oslo_vmware.api [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Waiting for the task: (returnval){ [ 2269.355477] env[62824]: value = "task-2146713" [ 2269.355477] env[62824]: _type = "Task" [ 2269.355477] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2269.363958] env[62824]: DEBUG oslo_vmware.api [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Task: {'id': task-2146713, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2269.385140] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e89a221d-5c30-48d4-bfc0-a91f11a74aac tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2269.484369] env[62824]: DEBUG oslo_vmware.api [req-32ee7b84-1112-404a-8337-b83f371c4eef req-6e242573-9123-4732-9c52-7f2b349c2ed1 service nova] Task: {'id': task-2146706, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2269.547853] env[62824]: DEBUG nova.scheduler.client.report [None req-9b8d58d8-8420-4bcd-b5ae-0f85a3c51179 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2269.652719] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Releasing lock "refresh_cache-815bd7ae-f461-4c74-9401-7c5fe43679ee" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2269.653090] env[62824]: DEBUG nova.compute.manager [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 815bd7ae-f461-4c74-9401-7c5fe43679ee] Instance network_info: |[{"id": "3cb13172-9b03-4150-8070-b5b5f504d942", "address": "fa:16:3e:39:e5:73", "network": {"id": "fc5d7128-fe9b-44ef-9d91-839190a97226", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-796405330-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35b7289f4746463e9b1cc47fe914a835", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0954fad3-d24d-496c-83e6-a09d3cb556fc", "external-id": "nsx-vlan-transportzone-216", "segmentation_id": 216, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3cb13172-9b", "ovs_interfaceid": "3cb13172-9b03-4150-8070-b5b5f504d942", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2269.653429] env[62824]: DEBUG oslo_concurrency.lockutils [req-6aff378b-18e6-4d29-bd16-d368b66b6970 req-26f1a76e-a608-483d-876d-d0a746054819 service nova] Acquired lock "refresh_cache-815bd7ae-f461-4c74-9401-7c5fe43679ee" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2269.653642] env[62824]: DEBUG nova.network.neutron [req-6aff378b-18e6-4d29-bd16-d368b66b6970 req-26f1a76e-a608-483d-876d-d0a746054819 service nova] [instance: 815bd7ae-f461-4c74-9401-7c5fe43679ee] Refreshing network info cache for port 3cb13172-9b03-4150-8070-b5b5f504d942 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2269.654930] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 815bd7ae-f461-4c74-9401-7c5fe43679ee] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:39:e5:73', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0954fad3-d24d-496c-83e6-a09d3cb556fc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3cb13172-9b03-4150-8070-b5b5f504d942', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2269.663179] env[62824]: DEBUG oslo.service.loopingcall [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2269.666476] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 815bd7ae-f461-4c74-9401-7c5fe43679ee] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2269.667061] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ef6c46bc-dc2f-486d-bd2a-6c6b6915bf04 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2269.691682] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2269.691682] env[62824]: value = "task-2146714" [ 2269.691682] env[62824]: _type = "Task" [ 2269.691682] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2269.702833] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146714, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2269.843020] env[62824]: DEBUG oslo_vmware.api [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146711, 'name': Rename_Task, 'duration_secs': 0.151691} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2269.843270] env[62824]: DEBUG oslo_vmware.api [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146712, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.115251} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2269.843513] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 3acdf86e-45e9-4353-a282-7272451389e7] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2269.843815] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2269.844080] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f76e2743-1f3a-4bb7-aa47-6310723307af {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2269.845965] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd1df7f8-ee3d-4584-b217-573cbd0cdac7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2269.868822] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Reconfiguring VM instance instance-0000005d to attach disk [datastore1] 0394d368-c846-4f3b-bfcf-2bc8e858052a/0394d368-c846-4f3b-bfcf-2bc8e858052a.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2269.875160] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-774d7df5-76b0-4788-86d6-94a83f56b351 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2269.889468] env[62824]: DEBUG oslo_vmware.api [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2269.889468] env[62824]: value = "task-2146715" [ 2269.889468] env[62824]: _type = "Task" [ 2269.889468] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2269.896061] env[62824]: DEBUG oslo_vmware.api [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Task: {'id': task-2146713, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2269.897727] env[62824]: DEBUG oslo_vmware.api [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2269.897727] env[62824]: value = "task-2146716" [ 2269.897727] env[62824]: _type = "Task" [ 2269.897727] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2269.903479] env[62824]: DEBUG oslo_vmware.api [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146715, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2269.908267] env[62824]: DEBUG oslo_vmware.api [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146716, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2269.972524] env[62824]: DEBUG nova.network.neutron [None req-c5fd25f6-e75a-454a-99ae-93fa0aabd927 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Updating instance_info_cache with network_info: [{"id": "7be95c5e-6e98-4fec-8550-1c22a787b95b", "address": "fa:16:3e:e8:69:3a", "network": {"id": "bba9b98c-c7eb-4890-8de5-eafacfdebd7d", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1928394728-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.160", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7fcdbfc29df64a1ba8d982bdcc667b64", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b49c5024-2ced-42ca-90cc-6066766d43e6", "external-id": "nsx-vlan-transportzone-239", "segmentation_id": 239, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7be95c5e-6e", "ovs_interfaceid": "7be95c5e-6e98-4fec-8550-1c22a787b95b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2269.986092] env[62824]: DEBUG oslo_vmware.api [req-32ee7b84-1112-404a-8337-b83f371c4eef req-6e242573-9123-4732-9c52-7f2b349c2ed1 service nova] Task: {'id': task-2146706, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2270.052770] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9b8d58d8-8420-4bcd-b5ae-0f85a3c51179 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.905s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2270.054676] env[62824]: DEBUG oslo_concurrency.lockutils [None req-83ef04f5-6202-449e-9917-20354b7a27cd tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.784s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2270.054907] env[62824]: DEBUG nova.objects.instance [None req-83ef04f5-6202-449e-9917-20354b7a27cd tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Lazy-loading 'resources' on Instance uuid b9f30f87-3594-4468-9d29-70890d8761e3 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2270.079906] env[62824]: INFO nova.scheduler.client.report [None req-9b8d58d8-8420-4bcd-b5ae-0f85a3c51179 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Deleted allocations for instance c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9 [ 2270.161942] env[62824]: DEBUG nova.network.neutron [req-6aff378b-18e6-4d29-bd16-d368b66b6970 req-26f1a76e-a608-483d-876d-d0a746054819 service nova] [instance: 815bd7ae-f461-4c74-9401-7c5fe43679ee] Updated VIF entry in instance network info cache for port 3cb13172-9b03-4150-8070-b5b5f504d942. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2270.162367] env[62824]: DEBUG nova.network.neutron [req-6aff378b-18e6-4d29-bd16-d368b66b6970 req-26f1a76e-a608-483d-876d-d0a746054819 service nova] [instance: 815bd7ae-f461-4c74-9401-7c5fe43679ee] Updating instance_info_cache with network_info: [{"id": "3cb13172-9b03-4150-8070-b5b5f504d942", "address": "fa:16:3e:39:e5:73", "network": {"id": "fc5d7128-fe9b-44ef-9d91-839190a97226", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-796405330-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35b7289f4746463e9b1cc47fe914a835", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0954fad3-d24d-496c-83e6-a09d3cb556fc", "external-id": "nsx-vlan-transportzone-216", "segmentation_id": 216, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3cb13172-9b", "ovs_interfaceid": "3cb13172-9b03-4150-8070-b5b5f504d942", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2270.201994] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146714, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2270.369904] env[62824]: DEBUG oslo_vmware.api [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Task: {'id': task-2146713, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2270.398134] env[62824]: DEBUG oslo_vmware.api [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146715, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2270.407314] env[62824]: DEBUG oslo_vmware.api [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146716, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2270.474822] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c5fd25f6-e75a-454a-99ae-93fa0aabd927 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Releasing lock "refresh_cache-5b3df31f-15fe-473f-992c-ddb272661c53" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2270.486761] env[62824]: DEBUG oslo_vmware.api [req-32ee7b84-1112-404a-8337-b83f371c4eef req-6e242573-9123-4732-9c52-7f2b349c2ed1 service nova] Task: {'id': task-2146706, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2270.586466] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9b8d58d8-8420-4bcd-b5ae-0f85a3c51179 tempest-ServersNegativeTestMultiTenantJSON-701481437 tempest-ServersNegativeTestMultiTenantJSON-701481437-project-member] Lock "c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.579s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2270.665619] env[62824]: DEBUG oslo_concurrency.lockutils [req-6aff378b-18e6-4d29-bd16-d368b66b6970 req-26f1a76e-a608-483d-876d-d0a746054819 service nova] Releasing lock "refresh_cache-815bd7ae-f461-4c74-9401-7c5fe43679ee" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2270.707381] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146714, 'name': CreateVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2270.814926] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b1bef88-bd70-4beb-8c11-9caeb9800e37 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2270.823144] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a458fc7b-8772-414e-a820-97de9483fb03 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2270.855974] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a98c1b7-8fc8-48e0-b2f3-7b2aca9a2721 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2270.868019] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e3dd5c1-6352-4de9-b39b-1a511d2bfe2f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2270.877923] env[62824]: DEBUG oslo_vmware.api [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Task: {'id': task-2146713, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.470963} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2270.885674] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 71731b21-c302-4d66-a579-889165673712/71731b21-c302-4d66-a579-889165673712.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2270.885977] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] [instance: 71731b21-c302-4d66-a579-889165673712] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2270.886505] env[62824]: DEBUG nova.compute.provider_tree [None req-83ef04f5-6202-449e-9917-20354b7a27cd tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2270.887855] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d54ba4a7-7fa8-41f2-aec5-cc7aeb1c77c7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2270.894248] env[62824]: DEBUG oslo_vmware.api [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Waiting for the task: (returnval){ [ 2270.894248] env[62824]: value = "task-2146717" [ 2270.894248] env[62824]: _type = "Task" [ 2270.894248] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2270.903045] env[62824]: DEBUG oslo_vmware.api [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146715, 'name': PowerOnVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2270.910969] env[62824]: DEBUG oslo_vmware.api [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Task: {'id': task-2146717, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2270.914142] env[62824]: DEBUG oslo_vmware.api [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146716, 'name': ReconfigVM_Task, 'duration_secs': 0.993948} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2270.914142] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Reconfigured VM instance instance-0000005d to attach disk [datastore1] 0394d368-c846-4f3b-bfcf-2bc8e858052a/0394d368-c846-4f3b-bfcf-2bc8e858052a.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2270.914931] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f5061822-481e-427b-9f86-bc42429f552c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2270.921011] env[62824]: DEBUG oslo_vmware.api [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2270.921011] env[62824]: value = "task-2146718" [ 2270.921011] env[62824]: _type = "Task" [ 2270.921011] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2270.930532] env[62824]: DEBUG oslo_vmware.api [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146718, 'name': Rename_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2270.982467] env[62824]: DEBUG nova.compute.manager [None req-c5fd25f6-e75a-454a-99ae-93fa0aabd927 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2270.983548] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1609859d-17a2-45b1-a33b-7ec26483e0cd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2270.995806] env[62824]: DEBUG oslo_vmware.api [req-32ee7b84-1112-404a-8337-b83f371c4eef req-6e242573-9123-4732-9c52-7f2b349c2ed1 service nova] Task: {'id': task-2146706, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2271.204099] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146714, 'name': CreateVM_Task, 'duration_secs': 1.05763} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2271.204238] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 815bd7ae-f461-4c74-9401-7c5fe43679ee] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2271.204960] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2271.205179] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2271.205506] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2271.205765] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2e92aa58-7240-4a6f-bd77-5021b3b100d3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2271.210706] env[62824]: DEBUG oslo_vmware.api [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Waiting for the task: (returnval){ [ 2271.210706] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52edda99-29c0-04d0-6724-e10f0ea53d71" [ 2271.210706] env[62824]: _type = "Task" [ 2271.210706] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2271.219715] env[62824]: DEBUG oslo_vmware.api [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52edda99-29c0-04d0-6724-e10f0ea53d71, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2271.391194] env[62824]: DEBUG nova.scheduler.client.report [None req-83ef04f5-6202-449e-9917-20354b7a27cd tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2271.406648] env[62824]: DEBUG oslo_vmware.api [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Task: {'id': task-2146717, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.08138} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2271.409894] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] [instance: 71731b21-c302-4d66-a579-889165673712] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2271.410229] env[62824]: DEBUG oslo_vmware.api [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146715, 'name': PowerOnVM_Task, 'duration_secs': 1.515434} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2271.410922] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acf550fd-1a99-4900-bf98-9e68cfd20ca0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2271.414315] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 3acdf86e-45e9-4353-a282-7272451389e7] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2271.414315] env[62824]: INFO nova.compute.manager [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 3acdf86e-45e9-4353-a282-7272451389e7] Took 15.65 seconds to spawn the instance on the hypervisor. [ 2271.414386] env[62824]: DEBUG nova.compute.manager [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 3acdf86e-45e9-4353-a282-7272451389e7] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2271.415801] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8468a8cb-f19a-4792-ba44-f347e2ea90c5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2271.436635] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] [instance: 71731b21-c302-4d66-a579-889165673712] Reconfiguring VM instance instance-00000062 to attach disk [datastore1] 71731b21-c302-4d66-a579-889165673712/71731b21-c302-4d66-a579-889165673712.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2271.440964] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9c02e8bc-1d63-4512-b393-f151fe8087f3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2271.463800] env[62824]: DEBUG oslo_vmware.api [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146718, 'name': Rename_Task, 'duration_secs': 0.23319} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2271.464959] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2271.465271] env[62824]: DEBUG oslo_vmware.api [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Waiting for the task: (returnval){ [ 2271.465271] env[62824]: value = "task-2146719" [ 2271.465271] env[62824]: _type = "Task" [ 2271.465271] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2271.465447] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2fa7f75f-2346-4917-bc2b-7b5badf01adc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2271.475028] env[62824]: DEBUG oslo_vmware.api [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Task: {'id': task-2146719, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2271.476155] env[62824]: DEBUG oslo_vmware.api [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2271.476155] env[62824]: value = "task-2146720" [ 2271.476155] env[62824]: _type = "Task" [ 2271.476155] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2271.486534] env[62824]: DEBUG oslo_vmware.api [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146720, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2271.489460] env[62824]: DEBUG oslo_vmware.api [req-32ee7b84-1112-404a-8337-b83f371c4eef req-6e242573-9123-4732-9c52-7f2b349c2ed1 service nova] Task: {'id': task-2146706, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2271.721777] env[62824]: DEBUG oslo_vmware.api [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52edda99-29c0-04d0-6724-e10f0ea53d71, 'name': SearchDatastore_Task, 'duration_secs': 0.010345} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2271.722852] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2271.722852] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 815bd7ae-f461-4c74-9401-7c5fe43679ee] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2271.722852] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2271.722852] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2271.723160] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2271.723264] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-af830769-8c50-4f22-a146-fe41d1a978ce {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2271.732096] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2271.732284] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2271.733184] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ba93794f-a383-42b4-ad78-878f9751ee60 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2271.743020] env[62824]: DEBUG oslo_vmware.api [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Waiting for the task: (returnval){ [ 2271.743020] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52e25ee3-5bfd-2df1-b794-698d875abb12" [ 2271.743020] env[62824]: _type = "Task" [ 2271.743020] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2271.749627] env[62824]: DEBUG oslo_vmware.api [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52e25ee3-5bfd-2df1-b794-698d875abb12, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2271.904462] env[62824]: DEBUG oslo_concurrency.lockutils [None req-83ef04f5-6202-449e-9917-20354b7a27cd tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.849s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2271.909879] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 2.937s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2271.910165] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2271.910344] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62824) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 2271.910653] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e89a221d-5c30-48d4-bfc0-a91f11a74aac tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.526s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2271.910900] env[62824]: DEBUG nova.objects.instance [None req-e89a221d-5c30-48d4-bfc0-a91f11a74aac tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lazy-loading 'resources' on Instance uuid 204b866f-e9c5-441d-bf8e-77d324b39ecd {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2271.912490] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37b29ba8-fcaa-4ee4-a2af-e7231c263021 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2271.927674] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e1641bc-b267-4916-b266-b13820af0221 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2271.946811] env[62824]: INFO nova.scheduler.client.report [None req-83ef04f5-6202-449e-9917-20354b7a27cd tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Deleted allocations for instance b9f30f87-3594-4468-9d29-70890d8761e3 [ 2271.949187] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30f041ed-15fe-40b0-9c4e-710bccc95b09 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2271.963632] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-721f6465-154a-417d-9f6e-121101baadce {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2271.969197] env[62824]: INFO nova.compute.manager [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 3acdf86e-45e9-4353-a282-7272451389e7] Took 21.95 seconds to build instance. [ 2271.980655] env[62824]: DEBUG oslo_vmware.api [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Task: {'id': task-2146719, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2272.019418] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179320MB free_disk=174GB free_vcpus=48 pci_devices=None {{(pid=62824) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 2272.019692] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2272.020775] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0528d75c-47d7-4e0c-a2fc-382b4d65e0ef {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2272.030889] env[62824]: DEBUG oslo_vmware.api [req-32ee7b84-1112-404a-8337-b83f371c4eef req-6e242573-9123-4732-9c52-7f2b349c2ed1 service nova] Task: {'id': task-2146706, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2272.035151] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c5fd25f6-e75a-454a-99ae-93fa0aabd927 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Doing hard reboot of VM {{(pid=62824) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 2272.035481] env[62824]: DEBUG oslo_vmware.api [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146720, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2272.035733] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-d7855199-07ef-4aa0-bf0f-1f7e77d26b80 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2272.042144] env[62824]: DEBUG oslo_vmware.api [None req-c5fd25f6-e75a-454a-99ae-93fa0aabd927 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2272.042144] env[62824]: value = "task-2146721" [ 2272.042144] env[62824]: _type = "Task" [ 2272.042144] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2272.050445] env[62824]: DEBUG oslo_vmware.api [None req-c5fd25f6-e75a-454a-99ae-93fa0aabd927 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2146721, 'name': ResetVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2272.254403] env[62824]: DEBUG oslo_vmware.api [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52e25ee3-5bfd-2df1-b794-698d875abb12, 'name': SearchDatastore_Task, 'duration_secs': 0.010232} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2272.255730] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aeb89af0-4850-4d07-bea6-eccd2f62377f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2272.262828] env[62824]: DEBUG oslo_vmware.api [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Waiting for the task: (returnval){ [ 2272.262828] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52ef507b-4b98-2424-e983-7c6a8efa0065" [ 2272.262828] env[62824]: _type = "Task" [ 2272.262828] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2272.272717] env[62824]: DEBUG oslo_vmware.api [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52ef507b-4b98-2424-e983-7c6a8efa0065, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2272.460340] env[62824]: DEBUG oslo_concurrency.lockutils [None req-83ef04f5-6202-449e-9917-20354b7a27cd tempest-VolumesAdminNegativeTest-1643349580 tempest-VolumesAdminNegativeTest-1643349580-project-member] Lock "b9f30f87-3594-4468-9d29-70890d8761e3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.348s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2272.479222] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bb61f7aa-13d0-44f1-a9b8-62d0c5452100 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "3acdf86e-45e9-4353-a282-7272451389e7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.467s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2272.480084] env[62824]: DEBUG oslo_vmware.api [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Task: {'id': task-2146719, 'name': ReconfigVM_Task, 'duration_secs': 0.711636} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2272.486034] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] [instance: 71731b21-c302-4d66-a579-889165673712] Reconfigured VM instance instance-00000062 to attach disk [datastore1] 71731b21-c302-4d66-a579-889165673712/71731b21-c302-4d66-a579-889165673712.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2272.490429] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5e5ba2da-0c94-46d3-8555-e51aecd5049b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2272.497888] env[62824]: DEBUG oslo_vmware.api [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146720, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2272.505501] env[62824]: DEBUG oslo_vmware.api [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Waiting for the task: (returnval){ [ 2272.505501] env[62824]: value = "task-2146722" [ 2272.505501] env[62824]: _type = "Task" [ 2272.505501] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2272.506017] env[62824]: DEBUG oslo_vmware.api [req-32ee7b84-1112-404a-8337-b83f371c4eef req-6e242573-9123-4732-9c52-7f2b349c2ed1 service nova] Task: {'id': task-2146706, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2272.517022] env[62824]: DEBUG oslo_vmware.api [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Task: {'id': task-2146722, 'name': Rename_Task} progress is 10%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2272.554865] env[62824]: DEBUG oslo_vmware.api [None req-c5fd25f6-e75a-454a-99ae-93fa0aabd927 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2146721, 'name': ResetVM_Task, 'duration_secs': 0.100125} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2272.555361] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c5fd25f6-e75a-454a-99ae-93fa0aabd927 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Did hard reboot of VM {{(pid=62824) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 2272.555650] env[62824]: DEBUG nova.compute.manager [None req-c5fd25f6-e75a-454a-99ae-93fa0aabd927 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2272.556433] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-313117a7-cf8f-43f1-ab4c-31baccee706e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2272.668189] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb692ea2-1b7d-43ae-a5f5-f7a7b796b840 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2272.676347] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d06cd47-0bc0-4458-986a-4e2aab692444 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2272.712085] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46236e96-fbf6-4f30-8a0d-b7c88bb0a204 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2272.721820] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3b2ac52-8adb-4f60-8d29-5669ff23f860 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2272.739214] env[62824]: DEBUG nova.compute.provider_tree [None req-e89a221d-5c30-48d4-bfc0-a91f11a74aac tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2272.773085] env[62824]: DEBUG oslo_vmware.api [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52ef507b-4b98-2424-e983-7c6a8efa0065, 'name': SearchDatastore_Task, 'duration_secs': 0.01181} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2272.773594] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2272.773594] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 815bd7ae-f461-4c74-9401-7c5fe43679ee/815bd7ae-f461-4c74-9401-7c5fe43679ee.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2272.773835] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6da1ae9e-2905-445d-8c3a-0ae836916820 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2272.783810] env[62824]: DEBUG oslo_vmware.api [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Waiting for the task: (returnval){ [ 2272.783810] env[62824]: value = "task-2146723" [ 2272.783810] env[62824]: _type = "Task" [ 2272.783810] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2272.792539] env[62824]: DEBUG oslo_vmware.api [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146723, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2272.998111] env[62824]: DEBUG oslo_vmware.api [req-32ee7b84-1112-404a-8337-b83f371c4eef req-6e242573-9123-4732-9c52-7f2b349c2ed1 service nova] Task: {'id': task-2146706, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2273.000289] env[62824]: DEBUG oslo_vmware.api [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146720, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2273.016401] env[62824]: DEBUG oslo_vmware.api [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Task: {'id': task-2146722, 'name': Rename_Task, 'duration_secs': 0.239798} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2273.016722] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] [instance: 71731b21-c302-4d66-a579-889165673712] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2273.017254] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-231733d0-4d7e-488d-8803-1c3d9a805c66 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2273.024435] env[62824]: DEBUG oslo_vmware.api [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Waiting for the task: (returnval){ [ 2273.024435] env[62824]: value = "task-2146724" [ 2273.024435] env[62824]: _type = "Task" [ 2273.024435] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2273.036838] env[62824]: DEBUG oslo_vmware.api [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Task: {'id': task-2146724, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2273.072609] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c5fd25f6-e75a-454a-99ae-93fa0aabd927 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lock "5b3df31f-15fe-473f-992c-ddb272661c53" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 4.408s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2273.089288] env[62824]: DEBUG oslo_concurrency.lockutils [None req-12eec243-fcdd-4df5-9749-a0cadc4b95b3 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquiring lock "3acdf86e-45e9-4353-a282-7272451389e7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2273.092023] env[62824]: DEBUG oslo_concurrency.lockutils [None req-12eec243-fcdd-4df5-9749-a0cadc4b95b3 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "3acdf86e-45e9-4353-a282-7272451389e7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2273.092023] env[62824]: DEBUG oslo_concurrency.lockutils [None req-12eec243-fcdd-4df5-9749-a0cadc4b95b3 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquiring lock "3acdf86e-45e9-4353-a282-7272451389e7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2273.092023] env[62824]: DEBUG oslo_concurrency.lockutils [None req-12eec243-fcdd-4df5-9749-a0cadc4b95b3 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "3acdf86e-45e9-4353-a282-7272451389e7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2273.092023] env[62824]: DEBUG oslo_concurrency.lockutils [None req-12eec243-fcdd-4df5-9749-a0cadc4b95b3 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "3acdf86e-45e9-4353-a282-7272451389e7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2273.094136] env[62824]: INFO nova.compute.manager [None req-12eec243-fcdd-4df5-9749-a0cadc4b95b3 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 3acdf86e-45e9-4353-a282-7272451389e7] Terminating instance [ 2273.243063] env[62824]: DEBUG nova.scheduler.client.report [None req-e89a221d-5c30-48d4-bfc0-a91f11a74aac tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2273.296894] env[62824]: DEBUG oslo_vmware.api [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146723, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2273.498470] env[62824]: DEBUG oslo_vmware.api [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146720, 'name': PowerOnVM_Task, 'duration_secs': 1.985046} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2273.502095] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2273.502322] env[62824]: DEBUG nova.compute.manager [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2273.502635] env[62824]: DEBUG oslo_vmware.api [req-32ee7b84-1112-404a-8337-b83f371c4eef req-6e242573-9123-4732-9c52-7f2b349c2ed1 service nova] Task: {'id': task-2146706, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2273.503377] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d4b1b58-7f03-4062-bff2-87d278bf98d5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2273.538633] env[62824]: DEBUG oslo_vmware.api [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Task: {'id': task-2146724, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2273.602324] env[62824]: DEBUG nova.compute.manager [None req-12eec243-fcdd-4df5-9749-a0cadc4b95b3 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 3acdf86e-45e9-4353-a282-7272451389e7] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2273.602556] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-12eec243-fcdd-4df5-9749-a0cadc4b95b3 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 3acdf86e-45e9-4353-a282-7272451389e7] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2273.603476] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82c35d1f-c972-4efd-8b91-36288aaa38ea {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2273.611432] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-12eec243-fcdd-4df5-9749-a0cadc4b95b3 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 3acdf86e-45e9-4353-a282-7272451389e7] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2273.611688] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c464b468-0dd5-4baa-8c8a-dfeadeb66c89 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2273.618069] env[62824]: DEBUG oslo_vmware.api [None req-12eec243-fcdd-4df5-9749-a0cadc4b95b3 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2273.618069] env[62824]: value = "task-2146725" [ 2273.618069] env[62824]: _type = "Task" [ 2273.618069] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2273.626600] env[62824]: DEBUG oslo_vmware.api [None req-12eec243-fcdd-4df5-9749-a0cadc4b95b3 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146725, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2273.750288] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e89a221d-5c30-48d4-bfc0-a91f11a74aac tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.839s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2273.752804] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 1.732s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2273.791855] env[62824]: INFO nova.scheduler.client.report [None req-e89a221d-5c30-48d4-bfc0-a91f11a74aac tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Deleted allocations for instance 204b866f-e9c5-441d-bf8e-77d324b39ecd [ 2273.810025] env[62824]: DEBUG oslo_vmware.api [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146723, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.537185} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2273.810025] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 815bd7ae-f461-4c74-9401-7c5fe43679ee/815bd7ae-f461-4c74-9401-7c5fe43679ee.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2273.810025] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 815bd7ae-f461-4c74-9401-7c5fe43679ee] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2273.810025] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-661dc73f-6dc4-4606-97ee-cc2573a5c441 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2273.818482] env[62824]: DEBUG oslo_vmware.api [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Waiting for the task: (returnval){ [ 2273.818482] env[62824]: value = "task-2146726" [ 2273.818482] env[62824]: _type = "Task" [ 2273.818482] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2273.832825] env[62824]: DEBUG oslo_vmware.api [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146726, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2273.996516] env[62824]: DEBUG oslo_vmware.api [req-32ee7b84-1112-404a-8337-b83f371c4eef req-6e242573-9123-4732-9c52-7f2b349c2ed1 service nova] Task: {'id': task-2146706, 'name': ReconfigVM_Task, 'duration_secs': 6.177727} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2273.999042] env[62824]: DEBUG oslo_concurrency.lockutils [req-32ee7b84-1112-404a-8337-b83f371c4eef req-6e242573-9123-4732-9c52-7f2b349c2ed1 service nova] Releasing lock "fc77c44a-180c-46ff-9690-9072c6213c91" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2273.999042] env[62824]: DEBUG nova.virt.vmwareapi.vmops [req-32ee7b84-1112-404a-8337-b83f371c4eef req-6e242573-9123-4732-9c52-7f2b349c2ed1 service nova] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Reconfigured VM to detach interface {{(pid=62824) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 2273.999042] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8dadfd7c-4da2-4b7c-97a3-d1039115d4f4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lock "fc77c44a-180c-46ff-9690-9072c6213c91" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 6.538s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2273.999042] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8dadfd7c-4da2-4b7c-97a3-d1039115d4f4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquiring lock "fc77c44a-180c-46ff-9690-9072c6213c91-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2273.999042] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8dadfd7c-4da2-4b7c-97a3-d1039115d4f4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lock "fc77c44a-180c-46ff-9690-9072c6213c91-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2273.999042] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8dadfd7c-4da2-4b7c-97a3-d1039115d4f4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lock "fc77c44a-180c-46ff-9690-9072c6213c91-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2274.000996] env[62824]: INFO nova.compute.manager [None req-8dadfd7c-4da2-4b7c-97a3-d1039115d4f4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Terminating instance [ 2274.017129] env[62824]: INFO nova.compute.manager [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] bringing vm to original state: 'stopped' [ 2274.039555] env[62824]: DEBUG oslo_vmware.api [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Task: {'id': task-2146724, 'name': PowerOnVM_Task, 'duration_secs': 0.873403} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2274.040552] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] [instance: 71731b21-c302-4d66-a579-889165673712] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2274.040863] env[62824]: INFO nova.compute.manager [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] [instance: 71731b21-c302-4d66-a579-889165673712] Took 8.30 seconds to spawn the instance on the hypervisor. [ 2274.041207] env[62824]: DEBUG nova.compute.manager [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] [instance: 71731b21-c302-4d66-a579-889165673712] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2274.043230] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2231a6b2-8ff7-4981-9862-fd6922bd7944 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2274.133232] env[62824]: DEBUG oslo_vmware.api [None req-12eec243-fcdd-4df5-9749-a0cadc4b95b3 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146725, 'name': PowerOffVM_Task, 'duration_secs': 0.210119} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2274.133788] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-12eec243-fcdd-4df5-9749-a0cadc4b95b3 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 3acdf86e-45e9-4353-a282-7272451389e7] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2274.133788] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-12eec243-fcdd-4df5-9749-a0cadc4b95b3 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 3acdf86e-45e9-4353-a282-7272451389e7] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2274.134097] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f81ba9c4-8d70-414f-82b4-100e7a9111da {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2274.300271] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e89a221d-5c30-48d4-bfc0-a91f11a74aac tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "204b866f-e9c5-441d-bf8e-77d324b39ecd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.215s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2274.330143] env[62824]: DEBUG oslo_vmware.api [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146726, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.182346} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2274.330506] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 815bd7ae-f461-4c74-9401-7c5fe43679ee] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2274.331863] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9a939bb-4f15-41c2-9737-5d9b295ff909 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2274.356043] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 815bd7ae-f461-4c74-9401-7c5fe43679ee] Reconfiguring VM instance instance-00000061 to attach disk [datastore1] 815bd7ae-f461-4c74-9401-7c5fe43679ee/815bd7ae-f461-4c74-9401-7c5fe43679ee.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2274.356877] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-95c4275c-470d-420a-b839-1658c956a12d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2274.378175] env[62824]: DEBUG oslo_vmware.api [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Waiting for the task: (returnval){ [ 2274.378175] env[62824]: value = "task-2146728" [ 2274.378175] env[62824]: _type = "Task" [ 2274.378175] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2274.386445] env[62824]: DEBUG oslo_vmware.api [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146728, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2274.505612] env[62824]: DEBUG nova.compute.manager [None req-8dadfd7c-4da2-4b7c-97a3-d1039115d4f4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2274.505857] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-8dadfd7c-4da2-4b7c-97a3-d1039115d4f4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2274.506920] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6ce4beb-1e28-4d24-9740-9b8daeaf0059 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2274.517823] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-8dadfd7c-4da2-4b7c-97a3-d1039115d4f4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2274.519028] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0c5300f3-4782-4070-b072-936d30361b1d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2274.520753] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-12eec243-fcdd-4df5-9749-a0cadc4b95b3 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 3acdf86e-45e9-4353-a282-7272451389e7] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2274.521015] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-12eec243-fcdd-4df5-9749-a0cadc4b95b3 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 3acdf86e-45e9-4353-a282-7272451389e7] Deleting contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2274.521270] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-12eec243-fcdd-4df5-9749-a0cadc4b95b3 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Deleting the datastore file [datastore1] 3acdf86e-45e9-4353-a282-7272451389e7 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2274.523382] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-98c0e970-5b69-4f25-9a07-68d8aef7a5ab {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2274.531575] env[62824]: DEBUG oslo_vmware.api [None req-8dadfd7c-4da2-4b7c-97a3-d1039115d4f4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Waiting for the task: (returnval){ [ 2274.531575] env[62824]: value = "task-2146730" [ 2274.531575] env[62824]: _type = "Task" [ 2274.531575] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2274.533587] env[62824]: DEBUG oslo_vmware.api [None req-12eec243-fcdd-4df5-9749-a0cadc4b95b3 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2274.533587] env[62824]: value = "task-2146729" [ 2274.533587] env[62824]: _type = "Task" [ 2274.533587] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2274.546533] env[62824]: DEBUG oslo_vmware.api [None req-8dadfd7c-4da2-4b7c-97a3-d1039115d4f4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146730, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2274.549735] env[62824]: DEBUG oslo_vmware.api [None req-12eec243-fcdd-4df5-9749-a0cadc4b95b3 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146729, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2274.568896] env[62824]: INFO nova.compute.manager [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] [instance: 71731b21-c302-4d66-a579-889165673712] Took 17.99 seconds to build instance. [ 2274.794981] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance b3003c4b-ae5a-48df-8c12-a915a76253f4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2274.795846] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 6d80ec06-8559-4964-8577-a2512aa366ed actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2274.796067] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance bbfcb2e3-9326-4548-b15b-e054cbfd192e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2274.796765] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 15479062-af75-4925-99b3-77d6a49751ad actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 2, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2274.796765] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance a6cd2032-de60-4f78-bf1e-79801d049df0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2274.797172] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance ca42fef0-1e90-4ab6-9d60-8ef7e4997884 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2274.797381] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 5b3df31f-15fe-473f-992c-ddb272661c53 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2274.797548] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance fc77c44a-180c-46ff-9690-9072c6213c91 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2274.797706] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2274.797856] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance feda7def-7788-4a60-a56a-9353a1475c54 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2274.798015] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 71e9733b-c84b-4501-8faf-a487f089b498 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2274.798347] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 0394d368-c846-4f3b-bfcf-2bc8e858052a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2274.798699] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 3acdf86e-45e9-4353-a282-7272451389e7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2274.798871] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 815bd7ae-f461-4c74-9401-7c5fe43679ee actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2274.799038] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 71731b21-c302-4d66-a579-889165673712 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2274.799291] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Total usable vcpus: 48, total allocated vcpus: 15 {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 2274.799655] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3392MB phys_disk=200GB used_disk=16GB total_vcpus=48 used_vcpus=15 pci_stats=[] {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 2274.887882] env[62824]: DEBUG oslo_vmware.api [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146728, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2275.000440] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6886bc0f-80d6-4460-947a-fac61c87e8ee {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2275.007609] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11cd7ce2-1de1-499b-bdfe-07e46a6c0d6f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2275.039821] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquiring lock "0394d368-c846-4f3b-bfcf-2bc8e858052a" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2275.040148] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "0394d368-c846-4f3b-bfcf-2bc8e858052a" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2275.040410] env[62824]: DEBUG nova.compute.manager [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2275.041407] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f870c47-1f67-4c07-a88e-32296cb2851b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2275.050066] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17f6f20a-a82e-495e-8289-ce31764351ab {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2275.067329] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecd64a78-191b-4795-8ab9-63858e0ed573 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2275.071568] env[62824]: DEBUG oslo_vmware.api [None req-8dadfd7c-4da2-4b7c-97a3-d1039115d4f4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146730, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2275.072756] env[62824]: DEBUG nova.compute.manager [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62824) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3404}} [ 2275.073635] env[62824]: DEBUG oslo_vmware.api [None req-12eec243-fcdd-4df5-9749-a0cadc4b95b3 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146729, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.175395} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2275.074895] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5fc20553-8886-45b1-84d0-2293d451819e tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Lock "71731b21-c302-4d66-a579-889165673712" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.515s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2275.075146] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-12eec243-fcdd-4df5-9749-a0cadc4b95b3 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2275.075330] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-12eec243-fcdd-4df5-9749-a0cadc4b95b3 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 3acdf86e-45e9-4353-a282-7272451389e7] Deleted contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2275.075545] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-12eec243-fcdd-4df5-9749-a0cadc4b95b3 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 3acdf86e-45e9-4353-a282-7272451389e7] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2275.075719] env[62824]: INFO nova.compute.manager [None req-12eec243-fcdd-4df5-9749-a0cadc4b95b3 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 3acdf86e-45e9-4353-a282-7272451389e7] Took 1.47 seconds to destroy the instance on the hypervisor. [ 2275.075958] env[62824]: DEBUG oslo.service.loopingcall [None req-12eec243-fcdd-4df5-9749-a0cadc4b95b3 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2275.076768] env[62824]: DEBUG nova.compute.manager [-] [instance: 3acdf86e-45e9-4353-a282-7272451389e7] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2275.076884] env[62824]: DEBUG nova.network.neutron [-] [instance: 3acdf86e-45e9-4353-a282-7272451389e7] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2275.089146] env[62824]: DEBUG nova.compute.provider_tree [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2275.305568] env[62824]: DEBUG oslo_concurrency.lockutils [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquiring lock "19bcead9-b27c-471a-b217-cf7bfe101ed8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2275.305804] env[62824]: DEBUG oslo_concurrency.lockutils [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "19bcead9-b27c-471a-b217-cf7bfe101ed8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2275.393236] env[62824]: DEBUG oslo_vmware.api [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146728, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2275.396227] env[62824]: DEBUG nova.compute.manager [req-0acce8f6-0b8a-494f-9245-c786e9c7a8ec req-6cb2b8de-e315-4438-baaa-334bd007831f service nova] [instance: 3acdf86e-45e9-4353-a282-7272451389e7] Received event network-vif-deleted-f35780f9-2827-45ec-abd6-0737b92ff5a3 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2275.396728] env[62824]: INFO nova.compute.manager [req-0acce8f6-0b8a-494f-9245-c786e9c7a8ec req-6cb2b8de-e315-4438-baaa-334bd007831f service nova] [instance: 3acdf86e-45e9-4353-a282-7272451389e7] Neutron deleted interface f35780f9-2827-45ec-abd6-0737b92ff5a3; detaching it from the instance and deleting it from the info cache [ 2275.397025] env[62824]: DEBUG nova.network.neutron [req-0acce8f6-0b8a-494f-9245-c786e9c7a8ec req-6cb2b8de-e315-4438-baaa-334bd007831f service nova] [instance: 3acdf86e-45e9-4353-a282-7272451389e7] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2275.551938] env[62824]: DEBUG oslo_vmware.api [None req-8dadfd7c-4da2-4b7c-97a3-d1039115d4f4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146730, 'name': PowerOffVM_Task, 'duration_secs': 0.770709} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2275.551938] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-8dadfd7c-4da2-4b7c-97a3-d1039115d4f4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2275.551938] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-8dadfd7c-4da2-4b7c-97a3-d1039115d4f4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2275.551938] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0d7685da-df52-4f13-821d-347320f3b1d4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2275.577564] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2275.577892] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fbcc456d-ba96-42d2-a201-de8408900199 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2275.584074] env[62824]: DEBUG oslo_vmware.api [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2275.584074] env[62824]: value = "task-2146732" [ 2275.584074] env[62824]: _type = "Task" [ 2275.584074] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2275.591592] env[62824]: DEBUG oslo_vmware.api [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146732, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2275.592607] env[62824]: DEBUG nova.scheduler.client.report [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2275.808930] env[62824]: DEBUG nova.compute.manager [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 19bcead9-b27c-471a-b217-cf7bfe101ed8] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2275.873565] env[62824]: DEBUG nova.network.neutron [-] [instance: 3acdf86e-45e9-4353-a282-7272451389e7] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2275.892059] env[62824]: DEBUG oslo_vmware.api [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146728, 'name': ReconfigVM_Task, 'duration_secs': 1.100297} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2275.896042] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 815bd7ae-f461-4c74-9401-7c5fe43679ee] Reconfigured VM instance instance-00000061 to attach disk [datastore1] 815bd7ae-f461-4c74-9401-7c5fe43679ee/815bd7ae-f461-4c74-9401-7c5fe43679ee.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2275.896042] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-68413207-f4cb-4fd1-9880-4def784a2450 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2275.897209] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-8dadfd7c-4da2-4b7c-97a3-d1039115d4f4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2275.897378] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-8dadfd7c-4da2-4b7c-97a3-d1039115d4f4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Deleting contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2275.897578] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-8dadfd7c-4da2-4b7c-97a3-d1039115d4f4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Deleting the datastore file [datastore2] fc77c44a-180c-46ff-9690-9072c6213c91 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2275.897812] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f1bf9248-a8a7-4fca-bb2e-4cbd07aa0a97 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2275.899984] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-493bf8af-2aca-4748-856d-24e1f2ea12e4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2275.904270] env[62824]: DEBUG oslo_vmware.api [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Waiting for the task: (returnval){ [ 2275.904270] env[62824]: value = "task-2146733" [ 2275.904270] env[62824]: _type = "Task" [ 2275.904270] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2275.910013] env[62824]: DEBUG oslo_vmware.api [None req-8dadfd7c-4da2-4b7c-97a3-d1039115d4f4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Waiting for the task: (returnval){ [ 2275.910013] env[62824]: value = "task-2146734" [ 2275.910013] env[62824]: _type = "Task" [ 2275.910013] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2275.913654] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-803fe285-1def-4605-9629-d29b0c9b208b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2275.929890] env[62824]: INFO nova.compute.manager [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] [instance: 71731b21-c302-4d66-a579-889165673712] Rebuilding instance [ 2275.936914] env[62824]: DEBUG oslo_vmware.api [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146733, 'name': Rename_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2275.946035] env[62824]: DEBUG oslo_vmware.api [None req-8dadfd7c-4da2-4b7c-97a3-d1039115d4f4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146734, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2275.966341] env[62824]: DEBUG nova.compute.manager [req-0acce8f6-0b8a-494f-9245-c786e9c7a8ec req-6cb2b8de-e315-4438-baaa-334bd007831f service nova] [instance: 3acdf86e-45e9-4353-a282-7272451389e7] Detach interface failed, port_id=f35780f9-2827-45ec-abd6-0737b92ff5a3, reason: Instance 3acdf86e-45e9-4353-a282-7272451389e7 could not be found. {{(pid=62824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11482}} [ 2276.011036] env[62824]: DEBUG nova.compute.manager [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] [instance: 71731b21-c302-4d66-a579-889165673712] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2276.012104] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-492defe1-11db-4525-a829-113408651151 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2276.095100] env[62824]: DEBUG oslo_vmware.api [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146732, 'name': PowerOffVM_Task, 'duration_secs': 0.390519} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2276.095434] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2276.095691] env[62824]: DEBUG nova.compute.manager [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2276.096537] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b529a8f-be67-4cfd-91f2-641d6439d178 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2276.100983] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62824) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 2276.101188] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.349s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2276.335953] env[62824]: DEBUG oslo_concurrency.lockutils [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2276.336239] env[62824]: DEBUG oslo_concurrency.lockutils [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2276.337840] env[62824]: INFO nova.compute.claims [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 19bcead9-b27c-471a-b217-cf7bfe101ed8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2276.376606] env[62824]: INFO nova.compute.manager [-] [instance: 3acdf86e-45e9-4353-a282-7272451389e7] Took 1.30 seconds to deallocate network for instance. [ 2276.418974] env[62824]: DEBUG oslo_vmware.api [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146733, 'name': Rename_Task, 'duration_secs': 0.212047} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2276.418974] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 815bd7ae-f461-4c74-9401-7c5fe43679ee] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2276.418974] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-52038af9-1bd4-43df-9dda-cad83ca07620 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2276.422549] env[62824]: DEBUG oslo_vmware.api [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Waiting for the task: (returnval){ [ 2276.422549] env[62824]: value = "task-2146735" [ 2276.422549] env[62824]: _type = "Task" [ 2276.422549] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2276.432377] env[62824]: DEBUG oslo_vmware.api [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146735, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2276.437163] env[62824]: DEBUG oslo_vmware.api [None req-8dadfd7c-4da2-4b7c-97a3-d1039115d4f4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146734, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.203906} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2276.437397] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-8dadfd7c-4da2-4b7c-97a3-d1039115d4f4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2276.437579] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-8dadfd7c-4da2-4b7c-97a3-d1039115d4f4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Deleted contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2276.437746] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-8dadfd7c-4da2-4b7c-97a3-d1039115d4f4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2276.437908] env[62824]: INFO nova.compute.manager [None req-8dadfd7c-4da2-4b7c-97a3-d1039115d4f4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Took 1.93 seconds to destroy the instance on the hypervisor. [ 2276.438251] env[62824]: DEBUG oslo.service.loopingcall [None req-8dadfd7c-4da2-4b7c-97a3-d1039115d4f4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2276.438450] env[62824]: DEBUG nova.compute.manager [-] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2276.438542] env[62824]: DEBUG nova.network.neutron [-] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2276.614858] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "0394d368-c846-4f3b-bfcf-2bc8e858052a" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.575s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2276.884121] env[62824]: DEBUG oslo_concurrency.lockutils [None req-12eec243-fcdd-4df5-9749-a0cadc4b95b3 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2276.933014] env[62824]: DEBUG oslo_vmware.api [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146735, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2277.028440] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] [instance: 71731b21-c302-4d66-a579-889165673712] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2277.028775] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a29d8d26-9ed3-47b5-8049-0ab20103a4c8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2277.036375] env[62824]: DEBUG oslo_vmware.api [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Waiting for the task: (returnval){ [ 2277.036375] env[62824]: value = "task-2146736" [ 2277.036375] env[62824]: _type = "Task" [ 2277.036375] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2277.047867] env[62824]: DEBUG oslo_vmware.api [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Task: {'id': task-2146736, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2277.124956] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2277.429731] env[62824]: DEBUG nova.compute.manager [req-a3d39276-8212-4b50-98fa-315d3f3e4247 req-733f20d8-eaf0-4341-9e70-fff44e55bf64 service nova] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Received event network-vif-deleted-e285a842-d54b-4614-84ea-239e34f5b41f {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2277.429939] env[62824]: INFO nova.compute.manager [req-a3d39276-8212-4b50-98fa-315d3f3e4247 req-733f20d8-eaf0-4341-9e70-fff44e55bf64 service nova] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Neutron deleted interface e285a842-d54b-4614-84ea-239e34f5b41f; detaching it from the instance and deleting it from the info cache [ 2277.430161] env[62824]: DEBUG nova.network.neutron [req-a3d39276-8212-4b50-98fa-315d3f3e4247 req-733f20d8-eaf0-4341-9e70-fff44e55bf64 service nova] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2277.441173] env[62824]: DEBUG oslo_vmware.api [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146735, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2277.525025] env[62824]: DEBUG nova.network.neutron [-] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2277.551817] env[62824]: DEBUG oslo_vmware.api [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Task: {'id': task-2146736, 'name': PowerOffVM_Task, 'duration_secs': 0.223429} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2277.552527] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] [instance: 71731b21-c302-4d66-a579-889165673712] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2277.552527] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] [instance: 71731b21-c302-4d66-a579-889165673712] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2277.553125] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a962d062-c271-4db4-ba99-9531ab099a48 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2277.563187] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] [instance: 71731b21-c302-4d66-a579-889165673712] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2277.563485] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-45e40f8e-9020-4e89-9045-f5085ef4b03f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2277.589205] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] [instance: 71731b21-c302-4d66-a579-889165673712] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2277.589464] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] [instance: 71731b21-c302-4d66-a579-889165673712] Deleting contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2277.589649] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Deleting the datastore file [datastore1] 71731b21-c302-4d66-a579-889165673712 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2277.590528] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e1dceb1-6fe0-4b69-a467-6b980f665726 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2277.593275] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a11ac22b-9cc3-4cc8-8703-9bb6f281098c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2277.601913] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-902241cd-c5e3-423e-af9f-832f5c187665 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2277.605663] env[62824]: DEBUG oslo_vmware.api [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Waiting for the task: (returnval){ [ 2277.605663] env[62824]: value = "task-2146738" [ 2277.605663] env[62824]: _type = "Task" [ 2277.605663] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2277.637317] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4df47693-0fd6-4037-93ea-b6ee677a8786 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2277.644577] env[62824]: DEBUG oslo_vmware.api [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Task: {'id': task-2146738, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2277.650270] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-412ff4a6-a1f7-4e1d-a977-28479b0d3483 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2277.665444] env[62824]: DEBUG nova.compute.provider_tree [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2277.912414] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e2bc6c97-30e3-4527-94a1-f1aa96256c5b tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquiring lock "0394d368-c846-4f3b-bfcf-2bc8e858052a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2277.912830] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e2bc6c97-30e3-4527-94a1-f1aa96256c5b tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "0394d368-c846-4f3b-bfcf-2bc8e858052a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2277.913186] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e2bc6c97-30e3-4527-94a1-f1aa96256c5b tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquiring lock "0394d368-c846-4f3b-bfcf-2bc8e858052a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2277.913499] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e2bc6c97-30e3-4527-94a1-f1aa96256c5b tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "0394d368-c846-4f3b-bfcf-2bc8e858052a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2277.913788] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e2bc6c97-30e3-4527-94a1-f1aa96256c5b tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "0394d368-c846-4f3b-bfcf-2bc8e858052a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2277.917186] env[62824]: INFO nova.compute.manager [None req-e2bc6c97-30e3-4527-94a1-f1aa96256c5b tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Terminating instance [ 2277.939379] env[62824]: DEBUG oslo_vmware.api [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146735, 'name': PowerOnVM_Task, 'duration_secs': 1.286013} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2277.939736] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4524c3f3-3fb7-4bb9-9711-ed7679d7d744 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2277.942593] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 815bd7ae-f461-4c74-9401-7c5fe43679ee] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2277.942938] env[62824]: INFO nova.compute.manager [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 815bd7ae-f461-4c74-9401-7c5fe43679ee] Took 18.30 seconds to spawn the instance on the hypervisor. [ 2277.943263] env[62824]: DEBUG nova.compute.manager [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 815bd7ae-f461-4c74-9401-7c5fe43679ee] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2277.944803] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19d849ef-e81b-476b-bb4e-b1d516a6e94b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2277.966814] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3c5b8d7-153d-4ce1-abdf-5c603d3a1092 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2278.004398] env[62824]: DEBUG nova.compute.manager [req-a3d39276-8212-4b50-98fa-315d3f3e4247 req-733f20d8-eaf0-4341-9e70-fff44e55bf64 service nova] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Detach interface failed, port_id=e285a842-d54b-4614-84ea-239e34f5b41f, reason: Instance fc77c44a-180c-46ff-9690-9072c6213c91 could not be found. {{(pid=62824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11482}} [ 2278.030668] env[62824]: INFO nova.compute.manager [-] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Took 1.59 seconds to deallocate network for instance. [ 2278.116689] env[62824]: DEBUG oslo_vmware.api [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Task: {'id': task-2146738, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.377507} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2278.116917] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2278.117130] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] [instance: 71731b21-c302-4d66-a579-889165673712] Deleted contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2278.117260] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] [instance: 71731b21-c302-4d66-a579-889165673712] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2278.168514] env[62824]: DEBUG nova.scheduler.client.report [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2278.421941] env[62824]: DEBUG nova.compute.manager [None req-e2bc6c97-30e3-4527-94a1-f1aa96256c5b tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2278.422231] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e2bc6c97-30e3-4527-94a1-f1aa96256c5b tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2278.423270] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0697df02-0bac-43e5-9528-44d783462f8a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2278.432174] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e2bc6c97-30e3-4527-94a1-f1aa96256c5b tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2278.432471] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2a4c323c-6a2f-46f7-bd89-10c3f9dbecef {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2278.471734] env[62824]: INFO nova.compute.manager [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 815bd7ae-f461-4c74-9401-7c5fe43679ee] Took 27.12 seconds to build instance. [ 2278.538788] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8dadfd7c-4da2-4b7c-97a3-d1039115d4f4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2278.673973] env[62824]: DEBUG oslo_concurrency.lockutils [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.337s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2278.674835] env[62824]: DEBUG nova.compute.manager [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 19bcead9-b27c-471a-b217-cf7bfe101ed8] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2278.677479] env[62824]: DEBUG oslo_concurrency.lockutils [None req-12eec243-fcdd-4df5-9749-a0cadc4b95b3 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.793s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2278.677929] env[62824]: DEBUG nova.objects.instance [None req-12eec243-fcdd-4df5-9749-a0cadc4b95b3 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lazy-loading 'resources' on Instance uuid 3acdf86e-45e9-4353-a282-7272451389e7 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2278.712055] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e2bc6c97-30e3-4527-94a1-f1aa96256c5b tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2278.712055] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e2bc6c97-30e3-4527-94a1-f1aa96256c5b tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Deleting contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2278.712055] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-e2bc6c97-30e3-4527-94a1-f1aa96256c5b tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Deleting the datastore file [datastore1] 0394d368-c846-4f3b-bfcf-2bc8e858052a {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2278.712055] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-859d21c9-64f1-4ced-8ab9-d738b9f36405 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2278.720296] env[62824]: DEBUG oslo_vmware.api [None req-e2bc6c97-30e3-4527-94a1-f1aa96256c5b tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2278.720296] env[62824]: value = "task-2146740" [ 2278.720296] env[62824]: _type = "Task" [ 2278.720296] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2278.732058] env[62824]: DEBUG oslo_vmware.api [None req-e2bc6c97-30e3-4527-94a1-f1aa96256c5b tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146740, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2278.974033] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bb3b5f7a-a480-412d-9e7b-e0f05ce0f9a2 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Lock "815bd7ae-f461-4c74-9401-7c5fe43679ee" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.632s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2278.994272] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3295493f-8421-4e01-a218-347a2b34fb31 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Acquiring lock "815bd7ae-f461-4c74-9401-7c5fe43679ee" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2278.994648] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3295493f-8421-4e01-a218-347a2b34fb31 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Lock "815bd7ae-f461-4c74-9401-7c5fe43679ee" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2278.994867] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3295493f-8421-4e01-a218-347a2b34fb31 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Acquiring lock "815bd7ae-f461-4c74-9401-7c5fe43679ee-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2278.995074] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3295493f-8421-4e01-a218-347a2b34fb31 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Lock "815bd7ae-f461-4c74-9401-7c5fe43679ee-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2278.995249] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3295493f-8421-4e01-a218-347a2b34fb31 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Lock "815bd7ae-f461-4c74-9401-7c5fe43679ee-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2278.997512] env[62824]: INFO nova.compute.manager [None req-3295493f-8421-4e01-a218-347a2b34fb31 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 815bd7ae-f461-4c74-9401-7c5fe43679ee] Terminating instance [ 2279.158674] env[62824]: DEBUG nova.virt.hardware [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2279.158918] env[62824]: DEBUG nova.virt.hardware [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2279.159211] env[62824]: DEBUG nova.virt.hardware [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2279.159345] env[62824]: DEBUG nova.virt.hardware [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2279.159496] env[62824]: DEBUG nova.virt.hardware [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2279.159618] env[62824]: DEBUG nova.virt.hardware [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2279.159770] env[62824]: DEBUG nova.virt.hardware [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2279.159944] env[62824]: DEBUG nova.virt.hardware [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2279.160486] env[62824]: DEBUG nova.virt.hardware [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2279.160689] env[62824]: DEBUG nova.virt.hardware [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2279.160869] env[62824]: DEBUG nova.virt.hardware [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2279.161773] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edf361b8-ce85-43d4-ac67-31f40c3194a9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2279.170308] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d551312-d4fa-461a-8def-a0f74844243e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2279.184987] env[62824]: DEBUG nova.compute.utils [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2279.192779] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] [instance: 71731b21-c302-4d66-a579-889165673712] Instance VIF info [] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2279.199097] env[62824]: DEBUG oslo.service.loopingcall [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2279.201813] env[62824]: DEBUG nova.compute.manager [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 19bcead9-b27c-471a-b217-cf7bfe101ed8] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2279.202036] env[62824]: DEBUG nova.network.neutron [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 19bcead9-b27c-471a-b217-cf7bfe101ed8] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2279.204206] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 71731b21-c302-4d66-a579-889165673712] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2279.208438] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-135bb5b4-010b-4b4e-ba12-2d9b544f053a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2279.229262] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2279.229262] env[62824]: value = "task-2146741" [ 2279.229262] env[62824]: _type = "Task" [ 2279.229262] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2279.237220] env[62824]: DEBUG oslo_vmware.api [None req-e2bc6c97-30e3-4527-94a1-f1aa96256c5b tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146740, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.226467} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2279.238040] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-e2bc6c97-30e3-4527-94a1-f1aa96256c5b tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2279.238609] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e2bc6c97-30e3-4527-94a1-f1aa96256c5b tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Deleted contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2279.238838] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e2bc6c97-30e3-4527-94a1-f1aa96256c5b tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2279.239335] env[62824]: INFO nova.compute.manager [None req-e2bc6c97-30e3-4527-94a1-f1aa96256c5b tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Took 0.82 seconds to destroy the instance on the hypervisor. [ 2279.239617] env[62824]: DEBUG oslo.service.loopingcall [None req-e2bc6c97-30e3-4527-94a1-f1aa96256c5b tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2279.243055] env[62824]: DEBUG nova.compute.manager [-] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2279.243237] env[62824]: DEBUG nova.network.neutron [-] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2279.244918] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146741, 'name': CreateVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2279.299023] env[62824]: DEBUG nova.policy [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c379c824a178444e99bb121716fc7887', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0c11aef7fb6247cb9b5272a6a063cd12', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2279.463470] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a51a1789-72db-4c0a-951e-14bc8091a1cf {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2279.472691] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24537bf9-1636-4fc8-887b-cbdbd9d3bc1b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2279.512346] env[62824]: DEBUG nova.compute.manager [None req-3295493f-8421-4e01-a218-347a2b34fb31 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 815bd7ae-f461-4c74-9401-7c5fe43679ee] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2279.512748] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-3295493f-8421-4e01-a218-347a2b34fb31 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 815bd7ae-f461-4c74-9401-7c5fe43679ee] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2279.513833] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5eb59ca2-dc15-485e-87e1-364c343cbb76 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2279.517397] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1c51b36-5a2b-4d64-af1f-c35482f8950c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2279.528174] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8932f49-704a-4a29-b4f4-96754e67de4f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2279.532513] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-3295493f-8421-4e01-a218-347a2b34fb31 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 815bd7ae-f461-4c74-9401-7c5fe43679ee] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2279.532788] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-444448a7-9542-43bc-a24a-070ada5dba86 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2279.547896] env[62824]: DEBUG nova.compute.provider_tree [None req-12eec243-fcdd-4df5-9749-a0cadc4b95b3 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2279.552331] env[62824]: DEBUG oslo_vmware.api [None req-3295493f-8421-4e01-a218-347a2b34fb31 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Waiting for the task: (returnval){ [ 2279.552331] env[62824]: value = "task-2146742" [ 2279.552331] env[62824]: _type = "Task" [ 2279.552331] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2279.562024] env[62824]: DEBUG oslo_vmware.api [None req-3295493f-8421-4e01-a218-347a2b34fb31 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146742, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2279.568675] env[62824]: DEBUG nova.compute.manager [req-cd8f0dfd-ead4-402d-a52d-cd5737b8ca27 req-6738d6c8-3adc-4fdd-bec1-4cf7f2d59c1e service nova] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Received event network-vif-deleted-2c7a5b60-534e-457c-a27b-29a489877783 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2279.568884] env[62824]: INFO nova.compute.manager [req-cd8f0dfd-ead4-402d-a52d-cd5737b8ca27 req-6738d6c8-3adc-4fdd-bec1-4cf7f2d59c1e service nova] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Neutron deleted interface 2c7a5b60-534e-457c-a27b-29a489877783; detaching it from the instance and deleting it from the info cache [ 2279.569081] env[62824]: DEBUG nova.network.neutron [req-cd8f0dfd-ead4-402d-a52d-cd5737b8ca27 req-6738d6c8-3adc-4fdd-bec1-4cf7f2d59c1e service nova] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2279.701389] env[62824]: DEBUG nova.compute.manager [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 19bcead9-b27c-471a-b217-cf7bfe101ed8] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2279.739269] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146741, 'name': CreateVM_Task, 'duration_secs': 0.291658} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2279.739449] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 71731b21-c302-4d66-a579-889165673712] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2279.739899] env[62824]: DEBUG oslo_concurrency.lockutils [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2279.740072] env[62824]: DEBUG oslo_concurrency.lockutils [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2279.740399] env[62824]: DEBUG oslo_concurrency.lockutils [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2279.740659] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b61f5c35-e266-4a8c-9b3b-9d90c2ab1f8a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2279.745366] env[62824]: DEBUG oslo_vmware.api [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Waiting for the task: (returnval){ [ 2279.745366] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5267e3c3-aef2-fd16-8d5b-55b1ee2747a7" [ 2279.745366] env[62824]: _type = "Task" [ 2279.745366] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2279.753581] env[62824]: DEBUG oslo_vmware.api [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5267e3c3-aef2-fd16-8d5b-55b1ee2747a7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2279.846969] env[62824]: DEBUG nova.network.neutron [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 19bcead9-b27c-471a-b217-cf7bfe101ed8] Successfully created port: f94cb6aa-5577-481c-aee8-c9399a84ba6e {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2280.038086] env[62824]: DEBUG nova.network.neutron [-] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2280.053321] env[62824]: DEBUG nova.scheduler.client.report [None req-12eec243-fcdd-4df5-9749-a0cadc4b95b3 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2280.074688] env[62824]: DEBUG oslo_vmware.api [None req-3295493f-8421-4e01-a218-347a2b34fb31 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146742, 'name': PowerOffVM_Task, 'duration_secs': 0.187225} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2280.074688] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-3295493f-8421-4e01-a218-347a2b34fb31 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 815bd7ae-f461-4c74-9401-7c5fe43679ee] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2280.074688] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-3295493f-8421-4e01-a218-347a2b34fb31 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 815bd7ae-f461-4c74-9401-7c5fe43679ee] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2280.074928] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1156e87b-e8f7-4b10-bbca-4558fe31cd40 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2280.079110] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e672add3-7dc9-4126-a11d-1bf8167c9f5c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2280.088917] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd52b547-24d0-4dde-8be6-6dfd99cca5e8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2280.128903] env[62824]: DEBUG nova.compute.manager [req-cd8f0dfd-ead4-402d-a52d-cd5737b8ca27 req-6738d6c8-3adc-4fdd-bec1-4cf7f2d59c1e service nova] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Detach interface failed, port_id=2c7a5b60-534e-457c-a27b-29a489877783, reason: Instance 0394d368-c846-4f3b-bfcf-2bc8e858052a could not be found. {{(pid=62824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11482}} [ 2280.251210] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-3295493f-8421-4e01-a218-347a2b34fb31 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 815bd7ae-f461-4c74-9401-7c5fe43679ee] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2280.252451] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-3295493f-8421-4e01-a218-347a2b34fb31 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 815bd7ae-f461-4c74-9401-7c5fe43679ee] Deleting contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2280.253451] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-3295493f-8421-4e01-a218-347a2b34fb31 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Deleting the datastore file [datastore1] 815bd7ae-f461-4c74-9401-7c5fe43679ee {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2280.254458] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-226b3bcd-73d9-45c2-b355-a4fe06361e8d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2280.263156] env[62824]: DEBUG oslo_vmware.api [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5267e3c3-aef2-fd16-8d5b-55b1ee2747a7, 'name': SearchDatastore_Task, 'duration_secs': 0.010133} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2280.264227] env[62824]: DEBUG oslo_concurrency.lockutils [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2280.264691] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] [instance: 71731b21-c302-4d66-a579-889165673712] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2280.265079] env[62824]: DEBUG oslo_concurrency.lockutils [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2280.265416] env[62824]: DEBUG oslo_concurrency.lockutils [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2280.265712] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2280.271042] env[62824]: DEBUG oslo_vmware.api [None req-3295493f-8421-4e01-a218-347a2b34fb31 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Waiting for the task: (returnval){ [ 2280.271042] env[62824]: value = "task-2146744" [ 2280.271042] env[62824]: _type = "Task" [ 2280.271042] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2280.271042] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2bc27293-9bf1-4f6d-9612-8a60fe8ca11e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2280.281992] env[62824]: DEBUG oslo_vmware.api [None req-3295493f-8421-4e01-a218-347a2b34fb31 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146744, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2280.283518] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2280.283844] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2280.284819] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8996ca80-4908-4d43-bf74-2478bd1df90f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2280.291064] env[62824]: DEBUG oslo_vmware.api [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Waiting for the task: (returnval){ [ 2280.291064] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]521e540e-945b-c22e-c392-c491a35e70ea" [ 2280.291064] env[62824]: _type = "Task" [ 2280.291064] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2280.301862] env[62824]: DEBUG oslo_vmware.api [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]521e540e-945b-c22e-c392-c491a35e70ea, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2280.544592] env[62824]: INFO nova.compute.manager [-] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Took 1.30 seconds to deallocate network for instance. [ 2280.561487] env[62824]: DEBUG oslo_concurrency.lockutils [None req-12eec243-fcdd-4df5-9749-a0cadc4b95b3 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.884s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2280.563940] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 3.439s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2280.564210] env[62824]: DEBUG nova.objects.instance [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62824) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 2280.588342] env[62824]: INFO nova.scheduler.client.report [None req-12eec243-fcdd-4df5-9749-a0cadc4b95b3 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Deleted allocations for instance 3acdf86e-45e9-4353-a282-7272451389e7 [ 2280.711080] env[62824]: DEBUG nova.compute.manager [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 19bcead9-b27c-471a-b217-cf7bfe101ed8] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2280.734169] env[62824]: DEBUG nova.virt.hardware [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2280.734414] env[62824]: DEBUG nova.virt.hardware [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2280.734572] env[62824]: DEBUG nova.virt.hardware [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2280.734758] env[62824]: DEBUG nova.virt.hardware [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2280.734907] env[62824]: DEBUG nova.virt.hardware [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2280.735070] env[62824]: DEBUG nova.virt.hardware [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2280.735284] env[62824]: DEBUG nova.virt.hardware [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2280.735447] env[62824]: DEBUG nova.virt.hardware [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2280.735614] env[62824]: DEBUG nova.virt.hardware [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2280.735780] env[62824]: DEBUG nova.virt.hardware [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2280.735955] env[62824]: DEBUG nova.virt.hardware [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2280.736842] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf849dfd-6d60-4969-bbd8-57872f78a490 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2280.745174] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ec9f8b7-c348-4fb2-9a98-e02d0d7acdaf {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2280.777578] env[62824]: DEBUG oslo_vmware.api [None req-3295493f-8421-4e01-a218-347a2b34fb31 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146744, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.146726} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2280.777823] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-3295493f-8421-4e01-a218-347a2b34fb31 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2280.778011] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-3295493f-8421-4e01-a218-347a2b34fb31 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 815bd7ae-f461-4c74-9401-7c5fe43679ee] Deleted contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2280.778203] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-3295493f-8421-4e01-a218-347a2b34fb31 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 815bd7ae-f461-4c74-9401-7c5fe43679ee] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2280.778373] env[62824]: INFO nova.compute.manager [None req-3295493f-8421-4e01-a218-347a2b34fb31 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 815bd7ae-f461-4c74-9401-7c5fe43679ee] Took 1.27 seconds to destroy the instance on the hypervisor. [ 2280.778649] env[62824]: DEBUG oslo.service.loopingcall [None req-3295493f-8421-4e01-a218-347a2b34fb31 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2280.778925] env[62824]: DEBUG nova.compute.manager [-] [instance: 815bd7ae-f461-4c74-9401-7c5fe43679ee] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2280.779045] env[62824]: DEBUG nova.network.neutron [-] [instance: 815bd7ae-f461-4c74-9401-7c5fe43679ee] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2280.801149] env[62824]: DEBUG oslo_vmware.api [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]521e540e-945b-c22e-c392-c491a35e70ea, 'name': SearchDatastore_Task, 'duration_secs': 0.009276} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2280.801984] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c2a1cdb7-473a-4d3e-8839-148358f0a2da {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2280.807430] env[62824]: DEBUG oslo_vmware.api [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Waiting for the task: (returnval){ [ 2280.807430] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5227e92f-8156-49fe-5efa-79df44b5e77e" [ 2280.807430] env[62824]: _type = "Task" [ 2280.807430] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2280.815225] env[62824]: DEBUG oslo_vmware.api [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5227e92f-8156-49fe-5efa-79df44b5e77e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2281.051186] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e2bc6c97-30e3-4527-94a1-f1aa96256c5b tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2281.095849] env[62824]: DEBUG oslo_concurrency.lockutils [None req-12eec243-fcdd-4df5-9749-a0cadc4b95b3 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "3acdf86e-45e9-4353-a282-7272451389e7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.006s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2281.318412] env[62824]: DEBUG oslo_vmware.api [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5227e92f-8156-49fe-5efa-79df44b5e77e, 'name': SearchDatastore_Task, 'duration_secs': 0.010361} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2281.318731] env[62824]: DEBUG oslo_concurrency.lockutils [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2281.318931] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 71731b21-c302-4d66-a579-889165673712/71731b21-c302-4d66-a579-889165673712.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2281.319220] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a840fd70-9e63-4856-8dd5-89104edb501b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2281.326113] env[62824]: DEBUG oslo_vmware.api [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Waiting for the task: (returnval){ [ 2281.326113] env[62824]: value = "task-2146745" [ 2281.326113] env[62824]: _type = "Task" [ 2281.326113] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2281.333713] env[62824]: DEBUG oslo_vmware.api [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Task: {'id': task-2146745, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2281.539920] env[62824]: DEBUG nova.network.neutron [-] [instance: 815bd7ae-f461-4c74-9401-7c5fe43679ee] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2281.574649] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c7ee61a5-9558-4fa1-aaca-92fdd19e91aa tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.010s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2281.575556] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8dadfd7c-4da2-4b7c-97a3-d1039115d4f4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.037s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2281.576160] env[62824]: DEBUG nova.objects.instance [None req-8dadfd7c-4da2-4b7c-97a3-d1039115d4f4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lazy-loading 'resources' on Instance uuid fc77c44a-180c-46ff-9690-9072c6213c91 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2281.581898] env[62824]: DEBUG nova.compute.manager [req-dc530000-3f3a-4313-9754-a0fbc2b124c4 req-da0068a9-f60e-441b-a219-6581a0ed462d service nova] [instance: 19bcead9-b27c-471a-b217-cf7bfe101ed8] Received event network-vif-plugged-f94cb6aa-5577-481c-aee8-c9399a84ba6e {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2281.582136] env[62824]: DEBUG oslo_concurrency.lockutils [req-dc530000-3f3a-4313-9754-a0fbc2b124c4 req-da0068a9-f60e-441b-a219-6581a0ed462d service nova] Acquiring lock "19bcead9-b27c-471a-b217-cf7bfe101ed8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2281.582352] env[62824]: DEBUG oslo_concurrency.lockutils [req-dc530000-3f3a-4313-9754-a0fbc2b124c4 req-da0068a9-f60e-441b-a219-6581a0ed462d service nova] Lock "19bcead9-b27c-471a-b217-cf7bfe101ed8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2281.582570] env[62824]: DEBUG oslo_concurrency.lockutils [req-dc530000-3f3a-4313-9754-a0fbc2b124c4 req-da0068a9-f60e-441b-a219-6581a0ed462d service nova] Lock "19bcead9-b27c-471a-b217-cf7bfe101ed8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2281.582729] env[62824]: DEBUG nova.compute.manager [req-dc530000-3f3a-4313-9754-a0fbc2b124c4 req-da0068a9-f60e-441b-a219-6581a0ed462d service nova] [instance: 19bcead9-b27c-471a-b217-cf7bfe101ed8] No waiting events found dispatching network-vif-plugged-f94cb6aa-5577-481c-aee8-c9399a84ba6e {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2281.582906] env[62824]: WARNING nova.compute.manager [req-dc530000-3f3a-4313-9754-a0fbc2b124c4 req-da0068a9-f60e-441b-a219-6581a0ed462d service nova] [instance: 19bcead9-b27c-471a-b217-cf7bfe101ed8] Received unexpected event network-vif-plugged-f94cb6aa-5577-481c-aee8-c9399a84ba6e for instance with vm_state building and task_state spawning. [ 2281.606226] env[62824]: DEBUG nova.compute.manager [req-85738f37-514d-4ed4-9da7-d6ff03f854ec req-8bf1c567-a905-488b-abb9-58ac2bc56d94 service nova] [instance: 815bd7ae-f461-4c74-9401-7c5fe43679ee] Received event network-vif-deleted-3cb13172-9b03-4150-8070-b5b5f504d942 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2281.690490] env[62824]: DEBUG oslo_concurrency.lockutils [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquiring lock "1805a6e7-48f2-47b4-b097-0644ba8217fa" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2281.690490] env[62824]: DEBUG oslo_concurrency.lockutils [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "1805a6e7-48f2-47b4-b097-0644ba8217fa" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2281.694995] env[62824]: DEBUG nova.network.neutron [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 19bcead9-b27c-471a-b217-cf7bfe101ed8] Successfully updated port: f94cb6aa-5577-481c-aee8-c9399a84ba6e {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2281.837194] env[62824]: DEBUG oslo_vmware.api [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Task: {'id': task-2146745, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.46921} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2281.837492] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 71731b21-c302-4d66-a579-889165673712/71731b21-c302-4d66-a579-889165673712.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2281.837721] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] [instance: 71731b21-c302-4d66-a579-889165673712] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2281.838277] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8ad31573-254a-4ab3-b773-c1ba04e57c2b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2281.844500] env[62824]: DEBUG oslo_vmware.api [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Waiting for the task: (returnval){ [ 2281.844500] env[62824]: value = "task-2146746" [ 2281.844500] env[62824]: _type = "Task" [ 2281.844500] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2281.852428] env[62824]: DEBUG oslo_vmware.api [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Task: {'id': task-2146746, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2282.042595] env[62824]: INFO nova.compute.manager [-] [instance: 815bd7ae-f461-4c74-9401-7c5fe43679ee] Took 1.26 seconds to deallocate network for instance. [ 2282.179963] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5009cb31-7660-4a2f-869d-7d502893a00d tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquiring lock "feda7def-7788-4a60-a56a-9353a1475c54" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2282.180233] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5009cb31-7660-4a2f-869d-7d502893a00d tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "feda7def-7788-4a60-a56a-9353a1475c54" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2282.180575] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5009cb31-7660-4a2f-869d-7d502893a00d tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquiring lock "feda7def-7788-4a60-a56a-9353a1475c54-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2282.180739] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5009cb31-7660-4a2f-869d-7d502893a00d tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "feda7def-7788-4a60-a56a-9353a1475c54-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2282.180958] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5009cb31-7660-4a2f-869d-7d502893a00d tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "feda7def-7788-4a60-a56a-9353a1475c54-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2282.183166] env[62824]: INFO nova.compute.manager [None req-5009cb31-7660-4a2f-869d-7d502893a00d tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: feda7def-7788-4a60-a56a-9353a1475c54] Terminating instance [ 2282.194107] env[62824]: DEBUG nova.compute.manager [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2282.200398] env[62824]: DEBUG oslo_concurrency.lockutils [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquiring lock "refresh_cache-19bcead9-b27c-471a-b217-cf7bfe101ed8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2282.200524] env[62824]: DEBUG oslo_concurrency.lockutils [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquired lock "refresh_cache-19bcead9-b27c-471a-b217-cf7bfe101ed8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2282.200662] env[62824]: DEBUG nova.network.neutron [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 19bcead9-b27c-471a-b217-cf7bfe101ed8] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2282.323885] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3459e0ea-d9ff-456f-ad45-2d68fbaeeeb6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2282.331989] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20294f6c-9fb7-4993-bc05-8a331ce4309e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2282.366636] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca33c3cd-6853-4d32-86e6-352d6fb28283 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2282.375806] env[62824]: DEBUG oslo_vmware.api [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Task: {'id': task-2146746, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061536} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2282.376676] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af744b28-4612-4a38-95e2-6a0f78a69043 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2282.380234] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] [instance: 71731b21-c302-4d66-a579-889165673712] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2282.381017] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-609f2ab7-2b0f-426b-bf8a-7a28ece0ca35 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2282.392575] env[62824]: DEBUG nova.compute.provider_tree [None req-8dadfd7c-4da2-4b7c-97a3-d1039115d4f4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2282.409963] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] [instance: 71731b21-c302-4d66-a579-889165673712] Reconfiguring VM instance instance-00000062 to attach disk [datastore1] 71731b21-c302-4d66-a579-889165673712/71731b21-c302-4d66-a579-889165673712.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2282.410901] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4618eda9-4ca7-4e69-b7d1-a68ad7f99ad5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2282.430305] env[62824]: DEBUG oslo_vmware.api [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Waiting for the task: (returnval){ [ 2282.430305] env[62824]: value = "task-2146747" [ 2282.430305] env[62824]: _type = "Task" [ 2282.430305] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2282.438103] env[62824]: DEBUG oslo_vmware.api [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Task: {'id': task-2146747, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2282.549189] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3295493f-8421-4e01-a218-347a2b34fb31 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2282.688905] env[62824]: DEBUG nova.compute.manager [None req-5009cb31-7660-4a2f-869d-7d502893a00d tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: feda7def-7788-4a60-a56a-9353a1475c54] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2282.689211] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5009cb31-7660-4a2f-869d-7d502893a00d tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: feda7def-7788-4a60-a56a-9353a1475c54] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2282.690210] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c26aa79-b3ce-4c82-89af-14ff3863de3a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2282.699772] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-5009cb31-7660-4a2f-869d-7d502893a00d tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: feda7def-7788-4a60-a56a-9353a1475c54] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2282.700015] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-008c6d29-1975-4a8e-ada2-49c464e0b8fa {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2282.709342] env[62824]: DEBUG oslo_vmware.api [None req-5009cb31-7660-4a2f-869d-7d502893a00d tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2282.709342] env[62824]: value = "task-2146748" [ 2282.709342] env[62824]: _type = "Task" [ 2282.709342] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2282.717652] env[62824]: DEBUG oslo_vmware.api [None req-5009cb31-7660-4a2f-869d-7d502893a00d tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146748, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2282.718617] env[62824]: DEBUG oslo_concurrency.lockutils [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2282.736495] env[62824]: DEBUG nova.network.neutron [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 19bcead9-b27c-471a-b217-cf7bfe101ed8] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2282.896375] env[62824]: DEBUG nova.scheduler.client.report [None req-8dadfd7c-4da2-4b7c-97a3-d1039115d4f4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2282.904763] env[62824]: DEBUG nova.network.neutron [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 19bcead9-b27c-471a-b217-cf7bfe101ed8] Updating instance_info_cache with network_info: [{"id": "f94cb6aa-5577-481c-aee8-c9399a84ba6e", "address": "fa:16:3e:a2:47:e9", "network": {"id": "b995934c-4934-4469-884c-73cb4de1796f", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1892307918-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c11aef7fb6247cb9b5272a6a063cd12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15165046-2de9-4ada-9e99-0126e20854a9", "external-id": "nsx-vlan-transportzone-974", "segmentation_id": 974, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf94cb6aa-55", "ovs_interfaceid": "f94cb6aa-5577-481c-aee8-c9399a84ba6e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2282.941775] env[62824]: DEBUG oslo_vmware.api [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Task: {'id': task-2146747, 'name': ReconfigVM_Task, 'duration_secs': 0.269993} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2282.942014] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] [instance: 71731b21-c302-4d66-a579-889165673712] Reconfigured VM instance instance-00000062 to attach disk [datastore1] 71731b21-c302-4d66-a579-889165673712/71731b21-c302-4d66-a579-889165673712.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2282.942587] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c01cc3d7-6180-4309-ac16-371192207c20 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2282.948800] env[62824]: DEBUG oslo_vmware.api [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Waiting for the task: (returnval){ [ 2282.948800] env[62824]: value = "task-2146749" [ 2282.948800] env[62824]: _type = "Task" [ 2282.948800] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2282.956345] env[62824]: DEBUG oslo_vmware.api [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Task: {'id': task-2146749, 'name': Rename_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2283.218656] env[62824]: DEBUG oslo_vmware.api [None req-5009cb31-7660-4a2f-869d-7d502893a00d tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146748, 'name': PowerOffVM_Task, 'duration_secs': 0.181294} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2283.218885] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-5009cb31-7660-4a2f-869d-7d502893a00d tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: feda7def-7788-4a60-a56a-9353a1475c54] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2283.219071] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5009cb31-7660-4a2f-869d-7d502893a00d tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: feda7def-7788-4a60-a56a-9353a1475c54] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2283.219331] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-57416eb0-06de-4e90-af5b-773554bdebd7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2283.402281] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8dadfd7c-4da2-4b7c-97a3-d1039115d4f4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.827s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2283.405348] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e2bc6c97-30e3-4527-94a1-f1aa96256c5b tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.354s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2283.405636] env[62824]: DEBUG nova.objects.instance [None req-e2bc6c97-30e3-4527-94a1-f1aa96256c5b tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lazy-loading 'resources' on Instance uuid 0394d368-c846-4f3b-bfcf-2bc8e858052a {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2283.410654] env[62824]: DEBUG oslo_concurrency.lockutils [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Releasing lock "refresh_cache-19bcead9-b27c-471a-b217-cf7bfe101ed8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2283.411009] env[62824]: DEBUG nova.compute.manager [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 19bcead9-b27c-471a-b217-cf7bfe101ed8] Instance network_info: |[{"id": "f94cb6aa-5577-481c-aee8-c9399a84ba6e", "address": "fa:16:3e:a2:47:e9", "network": {"id": "b995934c-4934-4469-884c-73cb4de1796f", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1892307918-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c11aef7fb6247cb9b5272a6a063cd12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15165046-2de9-4ada-9e99-0126e20854a9", "external-id": "nsx-vlan-transportzone-974", "segmentation_id": 974, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf94cb6aa-55", "ovs_interfaceid": "f94cb6aa-5577-481c-aee8-c9399a84ba6e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2283.411426] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 19bcead9-b27c-471a-b217-cf7bfe101ed8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a2:47:e9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '15165046-2de9-4ada-9e99-0126e20854a9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f94cb6aa-5577-481c-aee8-c9399a84ba6e', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2283.418967] env[62824]: DEBUG oslo.service.loopingcall [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2283.419754] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 19bcead9-b27c-471a-b217-cf7bfe101ed8] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2283.420036] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0cac3e3e-573e-432f-929d-001824046e4f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2283.434871] env[62824]: INFO nova.scheduler.client.report [None req-8dadfd7c-4da2-4b7c-97a3-d1039115d4f4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Deleted allocations for instance fc77c44a-180c-46ff-9690-9072c6213c91 [ 2283.442818] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2283.442818] env[62824]: value = "task-2146751" [ 2283.442818] env[62824]: _type = "Task" [ 2283.442818] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2283.450722] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146751, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2283.458163] env[62824]: DEBUG oslo_vmware.api [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Task: {'id': task-2146749, 'name': Rename_Task, 'duration_secs': 0.138585} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2283.458410] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] [instance: 71731b21-c302-4d66-a579-889165673712] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2283.458638] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a037d887-91ba-425c-9a4c-7051f30752d2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2283.464721] env[62824]: DEBUG oslo_vmware.api [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Waiting for the task: (returnval){ [ 2283.464721] env[62824]: value = "task-2146752" [ 2283.464721] env[62824]: _type = "Task" [ 2283.464721] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2283.474854] env[62824]: DEBUG oslo_vmware.api [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Task: {'id': task-2146752, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2283.612263] env[62824]: DEBUG nova.compute.manager [req-e7c71a01-4fd6-41df-a685-1e887f2bc95e req-3e49c394-6dc6-42da-b1b7-aa969a54937c service nova] [instance: 19bcead9-b27c-471a-b217-cf7bfe101ed8] Received event network-changed-f94cb6aa-5577-481c-aee8-c9399a84ba6e {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2283.612588] env[62824]: DEBUG nova.compute.manager [req-e7c71a01-4fd6-41df-a685-1e887f2bc95e req-3e49c394-6dc6-42da-b1b7-aa969a54937c service nova] [instance: 19bcead9-b27c-471a-b217-cf7bfe101ed8] Refreshing instance network info cache due to event network-changed-f94cb6aa-5577-481c-aee8-c9399a84ba6e. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2283.612676] env[62824]: DEBUG oslo_concurrency.lockutils [req-e7c71a01-4fd6-41df-a685-1e887f2bc95e req-3e49c394-6dc6-42da-b1b7-aa969a54937c service nova] Acquiring lock "refresh_cache-19bcead9-b27c-471a-b217-cf7bfe101ed8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2283.612839] env[62824]: DEBUG oslo_concurrency.lockutils [req-e7c71a01-4fd6-41df-a685-1e887f2bc95e req-3e49c394-6dc6-42da-b1b7-aa969a54937c service nova] Acquired lock "refresh_cache-19bcead9-b27c-471a-b217-cf7bfe101ed8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2283.613047] env[62824]: DEBUG nova.network.neutron [req-e7c71a01-4fd6-41df-a685-1e887f2bc95e req-3e49c394-6dc6-42da-b1b7-aa969a54937c service nova] [instance: 19bcead9-b27c-471a-b217-cf7bfe101ed8] Refreshing network info cache for port f94cb6aa-5577-481c-aee8-c9399a84ba6e {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2283.943745] env[62824]: DEBUG oslo_concurrency.lockutils [None req-8dadfd7c-4da2-4b7c-97a3-d1039115d4f4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lock "fc77c44a-180c-46ff-9690-9072c6213c91" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.946s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2283.944766] env[62824]: DEBUG oslo_concurrency.lockutils [req-173a28c0-4b70-47fe-a51f-32b96501007c req-855fc076-b8d3-449d-ad77-3a00b5b589f2 service nova] Acquired lock "fc77c44a-180c-46ff-9690-9072c6213c91" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2283.948509] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-827c63bc-6a36-4e1a-9931-3ae5939ef57f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2283.958941] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146751, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2283.960203] env[62824]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 2283.960358] env[62824]: DEBUG oslo_vmware.api [-] Fault list: [ManagedObjectNotFound] {{(pid=62824) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 2283.961284] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-863b59f8-c2b2-4cad-93be-9aa275d39988 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2283.973479] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7307c6a5-1a42-40a4-be41-5405cb497fb2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2283.992897] env[62824]: DEBUG oslo_vmware.api [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Task: {'id': task-2146752, 'name': PowerOnVM_Task, 'duration_secs': 0.436608} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2283.993246] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] [instance: 71731b21-c302-4d66-a579-889165673712] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2283.993463] env[62824]: DEBUG nova.compute.manager [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] [instance: 71731b21-c302-4d66-a579-889165673712] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2283.994292] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fbecfa7-f375-44d8-9317-5b09d59e9206 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2284.014463] env[62824]: ERROR root [req-173a28c0-4b70-47fe-a51f-32b96501007c req-855fc076-b8d3-449d-ad77-3a00b5b589f2 service nova] Original exception being dropped: ['Traceback (most recent call last):\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py", line 377, in request_handler\n response = request(managed_object, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 586, in __call__\n return client.invoke(args, kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 728, in invoke\n result = self.send(soapenv, timeout=timeout)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 777, in send\n return self.process_reply(reply.message, None, None)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 840, in process_reply\n raise WebFault(fault, replyroot)\n', "suds.WebFault: Server raised fault: 'The object 'vim.VirtualMachine:vm-438699' has already been deleted or has not been completely created'\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 301, in _invoke_api\n return api_method(*args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/vim_util.py", line 480, in get_object_property\n props = get_object_properties(vim, moref, [property_name],\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/vim_util.py", line 360, in get_object_properties\n retrieve_result = vim.RetrievePropertiesEx(\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py", line 413, in request_handler\n raise exceptions.VimFaultException(fault_list, fault_string,\n', "oslo_vmware.exceptions.VimFaultException: The object 'vim.VirtualMachine:vm-438699' has already been deleted or has not been completely created\nCause: Server raised fault: 'The object 'vim.VirtualMachine:vm-438699' has already been deleted or has not been completely created'\nFaults: [ManagedObjectNotFound]\nDetails: {'obj': 'vm-438699'}\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/session.py", line 123, in _call_method\n return self.invoke_api(module, method, self.vim, *args,\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 358, in invoke_api\n return _invoke_api(module, method, *args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 122, in func\n return evt.wait()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait\n result = hub.switch()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch\n return self.greenlet.switch()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/common/loopingcall.py", line 122, in _inner\n idle = self.f(*self.args, **self.kw)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 96, in _func\n result = f(*args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 341, in _invoke_api\n raise clazz(str(excep),\n', "oslo_vmware.exceptions.ManagedObjectNotFoundException: The object 'vim.VirtualMachine:vm-438699' has already been deleted or has not been completely created\nCause: Server raised fault: 'The object 'vim.VirtualMachine:vm-438699' has already been deleted or has not been completely created'\nFaults: [ManagedObjectNotFound]\nDetails: {'obj': 'vm-438699'}\n"]: nova.exception.InstanceNotFound: Instance fc77c44a-180c-46ff-9690-9072c6213c91 could not be found. [ 2284.014933] env[62824]: DEBUG oslo_concurrency.lockutils [req-173a28c0-4b70-47fe-a51f-32b96501007c req-855fc076-b8d3-449d-ad77-3a00b5b589f2 service nova] Releasing lock "fc77c44a-180c-46ff-9690-9072c6213c91" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2284.015177] env[62824]: DEBUG nova.compute.manager [req-173a28c0-4b70-47fe-a51f-32b96501007c req-855fc076-b8d3-449d-ad77-3a00b5b589f2 service nova] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Detach interface failed, port_id=2b6381b4-f3c7-4e22-b698-de190c8199ee, reason: Instance fc77c44a-180c-46ff-9690-9072c6213c91 could not be found. {{(pid=62824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11482}} [ 2284.015375] env[62824]: DEBUG nova.compute.manager [req-173a28c0-4b70-47fe-a51f-32b96501007c req-855fc076-b8d3-449d-ad77-3a00b5b589f2 service nova] [instance: c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9] Received event network-vif-deleted-78cd208b-52d0-44fb-af4b-7012963b8b85 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2284.015546] env[62824]: INFO nova.compute.manager [req-173a28c0-4b70-47fe-a51f-32b96501007c req-855fc076-b8d3-449d-ad77-3a00b5b589f2 service nova] [instance: c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9] Neutron deleted interface 78cd208b-52d0-44fb-af4b-7012963b8b85; detaching it from the instance and deleting it from the info cache [ 2284.015717] env[62824]: DEBUG nova.network.neutron [req-173a28c0-4b70-47fe-a51f-32b96501007c req-855fc076-b8d3-449d-ad77-3a00b5b589f2 service nova] [instance: c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2284.123359] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2eef23e-6e36-4f63-978a-9cd2f7aa3cc4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2284.131471] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c8a1bc9-442c-4358-9d0b-e70e23829de6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2284.163562] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9c2bdf1-5fde-4e74-a12d-23a907a9fcd8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2284.171473] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1edf429b-b0bb-4753-8e25-e1fd7aae0464 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2284.188541] env[62824]: DEBUG nova.compute.provider_tree [None req-e2bc6c97-30e3-4527-94a1-f1aa96256c5b tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2284.374566] env[62824]: DEBUG nova.network.neutron [req-e7c71a01-4fd6-41df-a685-1e887f2bc95e req-3e49c394-6dc6-42da-b1b7-aa969a54937c service nova] [instance: 19bcead9-b27c-471a-b217-cf7bfe101ed8] Updated VIF entry in instance network info cache for port f94cb6aa-5577-481c-aee8-c9399a84ba6e. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2284.374947] env[62824]: DEBUG nova.network.neutron [req-e7c71a01-4fd6-41df-a685-1e887f2bc95e req-3e49c394-6dc6-42da-b1b7-aa969a54937c service nova] [instance: 19bcead9-b27c-471a-b217-cf7bfe101ed8] Updating instance_info_cache with network_info: [{"id": "f94cb6aa-5577-481c-aee8-c9399a84ba6e", "address": "fa:16:3e:a2:47:e9", "network": {"id": "b995934c-4934-4469-884c-73cb4de1796f", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1892307918-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c11aef7fb6247cb9b5272a6a063cd12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15165046-2de9-4ada-9e99-0126e20854a9", "external-id": "nsx-vlan-transportzone-974", "segmentation_id": 974, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf94cb6aa-55", "ovs_interfaceid": "f94cb6aa-5577-481c-aee8-c9399a84ba6e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2284.455804] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146751, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2284.521739] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d0d4fd91-5c50-4c07-874a-8483d8adf29d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2284.526991] env[62824]: DEBUG oslo_concurrency.lockutils [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2284.531317] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d866b444-b122-4c38-a61a-2ec7915c77cc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2284.562131] env[62824]: DEBUG nova.compute.manager [req-173a28c0-4b70-47fe-a51f-32b96501007c req-855fc076-b8d3-449d-ad77-3a00b5b589f2 service nova] [instance: c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9] Detach interface failed, port_id=78cd208b-52d0-44fb-af4b-7012963b8b85, reason: Instance c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9 could not be found. {{(pid=62824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11482}} [ 2284.562305] env[62824]: DEBUG nova.compute.manager [req-173a28c0-4b70-47fe-a51f-32b96501007c req-855fc076-b8d3-449d-ad77-3a00b5b589f2 service nova] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] Received event network-vif-deleted-2d7ab792-8a41-4c8b-91a4-e4a677466a12 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2284.562535] env[62824]: INFO nova.compute.manager [req-173a28c0-4b70-47fe-a51f-32b96501007c req-855fc076-b8d3-449d-ad77-3a00b5b589f2 service nova] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] Neutron deleted interface 2d7ab792-8a41-4c8b-91a4-e4a677466a12; detaching it from the instance and deleting it from the info cache [ 2284.562784] env[62824]: DEBUG nova.network.neutron [req-173a28c0-4b70-47fe-a51f-32b96501007c req-855fc076-b8d3-449d-ad77-3a00b5b589f2 service nova] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2284.691996] env[62824]: DEBUG nova.scheduler.client.report [None req-e2bc6c97-30e3-4527-94a1-f1aa96256c5b tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2284.878480] env[62824]: DEBUG oslo_concurrency.lockutils [req-e7c71a01-4fd6-41df-a685-1e887f2bc95e req-3e49c394-6dc6-42da-b1b7-aa969a54937c service nova] Releasing lock "refresh_cache-19bcead9-b27c-471a-b217-cf7bfe101ed8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2284.935846] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7fd912ef-d4ab-43f7-9196-f9d4a6f5c085 tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Acquiring lock "71731b21-c302-4d66-a579-889165673712" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2284.936141] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7fd912ef-d4ab-43f7-9196-f9d4a6f5c085 tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Lock "71731b21-c302-4d66-a579-889165673712" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2284.936367] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7fd912ef-d4ab-43f7-9196-f9d4a6f5c085 tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Acquiring lock "71731b21-c302-4d66-a579-889165673712-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2284.936553] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7fd912ef-d4ab-43f7-9196-f9d4a6f5c085 tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Lock "71731b21-c302-4d66-a579-889165673712-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2284.936747] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7fd912ef-d4ab-43f7-9196-f9d4a6f5c085 tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Lock "71731b21-c302-4d66-a579-889165673712-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2284.939101] env[62824]: INFO nova.compute.manager [None req-7fd912ef-d4ab-43f7-9196-f9d4a6f5c085 tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] [instance: 71731b21-c302-4d66-a579-889165673712] Terminating instance [ 2284.955801] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146751, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2285.066282] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0d716224-b2ff-47ab-8253-84034828b5eb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2285.076925] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfe92621-7d32-4b96-a413-3ef0ec2c3749 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2285.116599] env[62824]: DEBUG nova.compute.manager [req-173a28c0-4b70-47fe-a51f-32b96501007c req-855fc076-b8d3-449d-ad77-3a00b5b589f2 service nova] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] Detach interface failed, port_id=2d7ab792-8a41-4c8b-91a4-e4a677466a12, reason: Instance b9f30f87-3594-4468-9d29-70890d8761e3 could not be found. {{(pid=62824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11482}} [ 2285.197646] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e2bc6c97-30e3-4527-94a1-f1aa96256c5b tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.792s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2285.200272] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3295493f-8421-4e01-a218-347a2b34fb31 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.651s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2285.200498] env[62824]: DEBUG nova.objects.instance [None req-3295493f-8421-4e01-a218-347a2b34fb31 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Lazy-loading 'resources' on Instance uuid 815bd7ae-f461-4c74-9401-7c5fe43679ee {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2285.223557] env[62824]: INFO nova.scheduler.client.report [None req-e2bc6c97-30e3-4527-94a1-f1aa96256c5b tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Deleted allocations for instance 0394d368-c846-4f3b-bfcf-2bc8e858052a [ 2285.443064] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7fd912ef-d4ab-43f7-9196-f9d4a6f5c085 tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Acquiring lock "refresh_cache-71731b21-c302-4d66-a579-889165673712" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2285.443335] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7fd912ef-d4ab-43f7-9196-f9d4a6f5c085 tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Acquired lock "refresh_cache-71731b21-c302-4d66-a579-889165673712" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2285.443524] env[62824]: DEBUG nova.network.neutron [None req-7fd912ef-d4ab-43f7-9196-f9d4a6f5c085 tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] [instance: 71731b21-c302-4d66-a579-889165673712] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2285.457710] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146751, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2285.734714] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e2bc6c97-30e3-4527-94a1-f1aa96256c5b tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "0394d368-c846-4f3b-bfcf-2bc8e858052a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.822s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2285.865074] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73e3709f-a0a9-4203-a909-fb27565f5983 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2285.873033] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80a1f381-5e66-4a5a-a46c-a6142714844a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2285.903044] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a210db1-6450-4819-80fd-edab70ae938a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2285.909373] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8d6da39-a50f-4b57-affb-a4af029f7b12 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2285.923225] env[62824]: DEBUG nova.compute.provider_tree [None req-3295493f-8421-4e01-a218-347a2b34fb31 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2285.955489] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146751, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2285.963803] env[62824]: DEBUG nova.network.neutron [None req-7fd912ef-d4ab-43f7-9196-f9d4a6f5c085 tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] [instance: 71731b21-c302-4d66-a579-889165673712] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2286.010627] env[62824]: DEBUG nova.network.neutron [None req-7fd912ef-d4ab-43f7-9196-f9d4a6f5c085 tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] [instance: 71731b21-c302-4d66-a579-889165673712] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2286.292553] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquiring lock "e1574266-e53b-4169-9db6-c66b84895edb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2286.292795] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lock "e1574266-e53b-4169-9db6-c66b84895edb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2286.426199] env[62824]: DEBUG nova.scheduler.client.report [None req-3295493f-8421-4e01-a218-347a2b34fb31 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2286.458700] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146751, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2286.513176] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7fd912ef-d4ab-43f7-9196-f9d4a6f5c085 tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Releasing lock "refresh_cache-71731b21-c302-4d66-a579-889165673712" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2286.513643] env[62824]: DEBUG nova.compute.manager [None req-7fd912ef-d4ab-43f7-9196-f9d4a6f5c085 tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] [instance: 71731b21-c302-4d66-a579-889165673712] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2286.513840] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-7fd912ef-d4ab-43f7-9196-f9d4a6f5c085 tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] [instance: 71731b21-c302-4d66-a579-889165673712] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2286.514804] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b58d0f65-99c9-473d-a34f-98707fc3e86c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2286.522427] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-7fd912ef-d4ab-43f7-9196-f9d4a6f5c085 tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] [instance: 71731b21-c302-4d66-a579-889165673712] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2286.524453] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-759054ba-22dd-432c-aace-ccd472ade9ed {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2286.529895] env[62824]: DEBUG oslo_vmware.api [None req-7fd912ef-d4ab-43f7-9196-f9d4a6f5c085 tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Waiting for the task: (returnval){ [ 2286.529895] env[62824]: value = "task-2146753" [ 2286.529895] env[62824]: _type = "Task" [ 2286.529895] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2286.537889] env[62824]: DEBUG oslo_vmware.api [None req-7fd912ef-d4ab-43f7-9196-f9d4a6f5c085 tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Task: {'id': task-2146753, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2286.795678] env[62824]: DEBUG nova.compute.manager [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2286.933012] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3295493f-8421-4e01-a218-347a2b34fb31 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.733s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2286.936626] env[62824]: DEBUG oslo_concurrency.lockutils [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.218s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2286.939095] env[62824]: INFO nova.compute.claims [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2286.963086] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146751, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2286.964796] env[62824]: INFO nova.scheduler.client.report [None req-3295493f-8421-4e01-a218-347a2b34fb31 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Deleted allocations for instance 815bd7ae-f461-4c74-9401-7c5fe43679ee [ 2287.040917] env[62824]: DEBUG oslo_vmware.api [None req-7fd912ef-d4ab-43f7-9196-f9d4a6f5c085 tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Task: {'id': task-2146753, 'name': PowerOffVM_Task, 'duration_secs': 0.185179} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2287.041176] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-7fd912ef-d4ab-43f7-9196-f9d4a6f5c085 tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] [instance: 71731b21-c302-4d66-a579-889165673712] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2287.041386] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-7fd912ef-d4ab-43f7-9196-f9d4a6f5c085 tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] [instance: 71731b21-c302-4d66-a579-889165673712] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2287.041656] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bb80dcbe-3aa3-4cd4-8bfe-51bb84f5f1e2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2287.067555] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-7fd912ef-d4ab-43f7-9196-f9d4a6f5c085 tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] [instance: 71731b21-c302-4d66-a579-889165673712] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2287.067831] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-7fd912ef-d4ab-43f7-9196-f9d4a6f5c085 tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] [instance: 71731b21-c302-4d66-a579-889165673712] Deleting contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2287.068057] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-7fd912ef-d4ab-43f7-9196-f9d4a6f5c085 tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Deleting the datastore file [datastore1] 71731b21-c302-4d66-a579-889165673712 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2287.068348] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f3d2e4fe-38cf-4fa2-9477-2de86ac83124 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2287.075827] env[62824]: DEBUG oslo_vmware.api [None req-7fd912ef-d4ab-43f7-9196-f9d4a6f5c085 tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Waiting for the task: (returnval){ [ 2287.075827] env[62824]: value = "task-2146755" [ 2287.075827] env[62824]: _type = "Task" [ 2287.075827] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2287.084263] env[62824]: DEBUG oslo_vmware.api [None req-7fd912ef-d4ab-43f7-9196-f9d4a6f5c085 tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Task: {'id': task-2146755, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2287.324305] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2287.459783] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146751, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2287.474626] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3295493f-8421-4e01-a218-347a2b34fb31 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Lock "815bd7ae-f461-4c74-9401-7c5fe43679ee" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.480s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2287.590025] env[62824]: DEBUG oslo_vmware.api [None req-7fd912ef-d4ab-43f7-9196-f9d4a6f5c085 tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Task: {'id': task-2146755, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.094114} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2287.590025] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-7fd912ef-d4ab-43f7-9196-f9d4a6f5c085 tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2287.590025] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-7fd912ef-d4ab-43f7-9196-f9d4a6f5c085 tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] [instance: 71731b21-c302-4d66-a579-889165673712] Deleted contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2287.590025] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-7fd912ef-d4ab-43f7-9196-f9d4a6f5c085 tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] [instance: 71731b21-c302-4d66-a579-889165673712] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2287.590025] env[62824]: INFO nova.compute.manager [None req-7fd912ef-d4ab-43f7-9196-f9d4a6f5c085 tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] [instance: 71731b21-c302-4d66-a579-889165673712] Took 1.07 seconds to destroy the instance on the hypervisor. [ 2287.590025] env[62824]: DEBUG oslo.service.loopingcall [None req-7fd912ef-d4ab-43f7-9196-f9d4a6f5c085 tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2287.590025] env[62824]: DEBUG nova.compute.manager [-] [instance: 71731b21-c302-4d66-a579-889165673712] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2287.590025] env[62824]: DEBUG nova.network.neutron [-] [instance: 71731b21-c302-4d66-a579-889165673712] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2287.686600] env[62824]: DEBUG nova.network.neutron [-] [instance: 71731b21-c302-4d66-a579-889165673712] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2287.960059] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146751, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2288.129096] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7f85451-6461-42c8-8764-a524b7cb7813 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2288.135766] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d83c6e6-58d4-4357-b253-48fdc8edcef4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2288.166396] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b887c980-7e7b-4091-bce0-51ca5220d469 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2288.174177] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70d8aaa1-e062-4683-8332-d52cf7bc9227 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2288.187841] env[62824]: DEBUG nova.compute.provider_tree [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2288.189265] env[62824]: DEBUG nova.network.neutron [-] [instance: 71731b21-c302-4d66-a579-889165673712] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2288.460812] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146751, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2288.504013] env[62824]: DEBUG oslo_concurrency.lockutils [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquiring lock "3d294ee9-92f6-4115-95db-9d4b66562b56" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2288.504401] env[62824]: DEBUG oslo_concurrency.lockutils [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "3d294ee9-92f6-4115-95db-9d4b66562b56" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2288.692488] env[62824]: DEBUG nova.scheduler.client.report [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2288.696026] env[62824]: INFO nova.compute.manager [-] [instance: 71731b21-c302-4d66-a579-889165673712] Took 1.11 seconds to deallocate network for instance. [ 2288.961362] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146751, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2289.006802] env[62824]: DEBUG nova.compute.manager [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2289.197119] env[62824]: DEBUG oslo_concurrency.lockutils [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.261s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2289.197672] env[62824]: DEBUG nova.compute.manager [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2289.200106] env[62824]: DEBUG oslo_concurrency.lockutils [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 4.673s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2289.200290] env[62824]: DEBUG nova.objects.instance [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] [instance: 71731b21-c302-4d66-a579-889165673712] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62824) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 2289.202966] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7fd912ef-d4ab-43f7-9196-f9d4a6f5c085 tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2289.461836] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146751, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2289.529661] env[62824]: DEBUG oslo_concurrency.lockutils [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2289.703969] env[62824]: DEBUG nova.compute.utils [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2289.707899] env[62824]: DEBUG nova.compute.manager [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2289.708149] env[62824]: DEBUG nova.network.neutron [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2289.764028] env[62824]: DEBUG nova.policy [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e8c10fcbfe69448fa71cfad4a7c8e179', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4479921caf5f405b8fc49baad390a0e9', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2289.962731] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146751, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2290.180489] env[62824]: DEBUG nova.network.neutron [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Successfully created port: 37654518-0a1a-44e2-8297-2169c44d0aad {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2290.212075] env[62824]: DEBUG oslo_concurrency.lockutils [None req-17a590fd-509c-4ddd-9353-e9a90393ed7c tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.012s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2290.213387] env[62824]: DEBUG nova.compute.manager [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2290.216436] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.893s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2290.217993] env[62824]: INFO nova.compute.claims [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2290.458429] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5009cb31-7660-4a2f-869d-7d502893a00d tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: feda7def-7788-4a60-a56a-9353a1475c54] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2290.458652] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5009cb31-7660-4a2f-869d-7d502893a00d tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: feda7def-7788-4a60-a56a-9353a1475c54] Deleting contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2290.458827] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-5009cb31-7660-4a2f-869d-7d502893a00d tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Deleting the datastore file [datastore2] feda7def-7788-4a60-a56a-9353a1475c54 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2290.459530] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a903f0b2-45d2-4eb1-a6c0-df3af1ead3fc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2290.464438] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146751, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2290.468611] env[62824]: DEBUG oslo_vmware.api [None req-5009cb31-7660-4a2f-869d-7d502893a00d tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2290.468611] env[62824]: value = "task-2146756" [ 2290.468611] env[62824]: _type = "Task" [ 2290.468611] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2290.475715] env[62824]: DEBUG oslo_vmware.api [None req-5009cb31-7660-4a2f-869d-7d502893a00d tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146756, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2290.963819] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146751, 'name': CreateVM_Task, 'duration_secs': 7.235466} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2290.964020] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 19bcead9-b27c-471a-b217-cf7bfe101ed8] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2290.964777] env[62824]: DEBUG oslo_concurrency.lockutils [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2290.964937] env[62824]: DEBUG oslo_concurrency.lockutils [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2290.965276] env[62824]: DEBUG oslo_concurrency.lockutils [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2290.965551] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e3ad36f3-6c56-4d85-bf24-9a09fe6ff6c9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2290.970093] env[62824]: DEBUG oslo_vmware.api [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 2290.970093] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]521aba37-6f9e-e83a-1e95-fcb489c62512" [ 2290.970093] env[62824]: _type = "Task" [ 2290.970093] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2290.981807] env[62824]: DEBUG oslo_vmware.api [None req-5009cb31-7660-4a2f-869d-7d502893a00d tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146756, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.142698} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2290.985161] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-5009cb31-7660-4a2f-869d-7d502893a00d tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2290.985321] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5009cb31-7660-4a2f-869d-7d502893a00d tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: feda7def-7788-4a60-a56a-9353a1475c54] Deleted contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2290.985508] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5009cb31-7660-4a2f-869d-7d502893a00d tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: feda7def-7788-4a60-a56a-9353a1475c54] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2290.985751] env[62824]: INFO nova.compute.manager [None req-5009cb31-7660-4a2f-869d-7d502893a00d tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: feda7def-7788-4a60-a56a-9353a1475c54] Took 8.30 seconds to destroy the instance on the hypervisor. [ 2290.986011] env[62824]: DEBUG oslo.service.loopingcall [None req-5009cb31-7660-4a2f-869d-7d502893a00d tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2290.986224] env[62824]: DEBUG oslo_vmware.api [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]521aba37-6f9e-e83a-1e95-fcb489c62512, 'name': SearchDatastore_Task, 'duration_secs': 0.009742} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2290.986413] env[62824]: DEBUG nova.compute.manager [-] [instance: feda7def-7788-4a60-a56a-9353a1475c54] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2290.986510] env[62824]: DEBUG nova.network.neutron [-] [instance: feda7def-7788-4a60-a56a-9353a1475c54] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2290.988106] env[62824]: DEBUG oslo_concurrency.lockutils [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2290.988338] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 19bcead9-b27c-471a-b217-cf7bfe101ed8] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2290.988590] env[62824]: DEBUG oslo_concurrency.lockutils [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2290.988744] env[62824]: DEBUG oslo_concurrency.lockutils [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2290.988931] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2290.989182] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2d3551f1-ef05-4997-89d1-3f9991b02693 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2290.997175] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2290.997258] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2290.997956] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-39f9cf97-4196-4aa6-8ddf-73bc6dc0c4f0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2291.003161] env[62824]: DEBUG oslo_vmware.api [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 2291.003161] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52b2b146-9e29-dab4-b3c9-81a49450967d" [ 2291.003161] env[62824]: _type = "Task" [ 2291.003161] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2291.010520] env[62824]: DEBUG oslo_vmware.api [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52b2b146-9e29-dab4-b3c9-81a49450967d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2291.223976] env[62824]: DEBUG nova.compute.manager [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2291.253743] env[62824]: DEBUG nova.virt.hardware [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2291.254119] env[62824]: DEBUG nova.virt.hardware [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2291.254173] env[62824]: DEBUG nova.virt.hardware [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2291.254334] env[62824]: DEBUG nova.virt.hardware [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2291.255104] env[62824]: DEBUG nova.virt.hardware [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2291.255104] env[62824]: DEBUG nova.virt.hardware [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2291.255104] env[62824]: DEBUG nova.virt.hardware [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2291.255104] env[62824]: DEBUG nova.virt.hardware [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2291.255272] env[62824]: DEBUG nova.virt.hardware [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2291.255315] env[62824]: DEBUG nova.virt.hardware [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2291.255489] env[62824]: DEBUG nova.virt.hardware [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2291.256647] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d2648bc-a441-4b1d-8964-a87779f0116b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2291.269158] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9ebad46-2877-431e-83e4-2d362087569b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2291.298793] env[62824]: DEBUG nova.compute.manager [req-112e615f-4b07-469e-8414-9b5ab2f35a01 req-daf38108-57c4-441f-819b-25be51a68493 service nova] [instance: feda7def-7788-4a60-a56a-9353a1475c54] Received event network-vif-deleted-b133d2a7-e5fb-44fd-b716-20381faecc3d {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2291.298989] env[62824]: INFO nova.compute.manager [req-112e615f-4b07-469e-8414-9b5ab2f35a01 req-daf38108-57c4-441f-819b-25be51a68493 service nova] [instance: feda7def-7788-4a60-a56a-9353a1475c54] Neutron deleted interface b133d2a7-e5fb-44fd-b716-20381faecc3d; detaching it from the instance and deleting it from the info cache [ 2291.299178] env[62824]: DEBUG nova.network.neutron [req-112e615f-4b07-469e-8414-9b5ab2f35a01 req-daf38108-57c4-441f-819b-25be51a68493 service nova] [instance: feda7def-7788-4a60-a56a-9353a1475c54] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2291.446989] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ecf0742-f0eb-4f69-868a-648a8f66750a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2291.454763] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-601837e2-9593-45fe-b487-9a0571374af3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2291.487113] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45b2e748-4541-4dfa-86f6-1f73be61b5de {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2291.494682] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afe1985e-4198-4e3c-947a-94ff2663c786 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2291.508477] env[62824]: DEBUG nova.compute.provider_tree [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2291.517584] env[62824]: DEBUG oslo_vmware.api [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52b2b146-9e29-dab4-b3c9-81a49450967d, 'name': SearchDatastore_Task, 'duration_secs': 0.007907} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2291.518810] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5f1c2c31-8a3b-4598-8cc7-b7c4fe54a613 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2291.523979] env[62824]: DEBUG oslo_vmware.api [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 2291.523979] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5239ee27-a54a-523a-967a-065b0c3a7006" [ 2291.523979] env[62824]: _type = "Task" [ 2291.523979] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2291.534273] env[62824]: DEBUG oslo_vmware.api [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5239ee27-a54a-523a-967a-065b0c3a7006, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2291.706990] env[62824]: DEBUG nova.compute.manager [req-944a0b2b-cecd-4400-a0ec-24c312103dd4 req-a70cdd0d-9fac-4b4f-98a6-396d42241e4a service nova] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Received event network-vif-plugged-37654518-0a1a-44e2-8297-2169c44d0aad {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2291.707228] env[62824]: DEBUG oslo_concurrency.lockutils [req-944a0b2b-cecd-4400-a0ec-24c312103dd4 req-a70cdd0d-9fac-4b4f-98a6-396d42241e4a service nova] Acquiring lock "1805a6e7-48f2-47b4-b097-0644ba8217fa-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2291.707473] env[62824]: DEBUG oslo_concurrency.lockutils [req-944a0b2b-cecd-4400-a0ec-24c312103dd4 req-a70cdd0d-9fac-4b4f-98a6-396d42241e4a service nova] Lock "1805a6e7-48f2-47b4-b097-0644ba8217fa-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2291.707652] env[62824]: DEBUG oslo_concurrency.lockutils [req-944a0b2b-cecd-4400-a0ec-24c312103dd4 req-a70cdd0d-9fac-4b4f-98a6-396d42241e4a service nova] Lock "1805a6e7-48f2-47b4-b097-0644ba8217fa-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2291.707792] env[62824]: DEBUG nova.compute.manager [req-944a0b2b-cecd-4400-a0ec-24c312103dd4 req-a70cdd0d-9fac-4b4f-98a6-396d42241e4a service nova] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] No waiting events found dispatching network-vif-plugged-37654518-0a1a-44e2-8297-2169c44d0aad {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2291.707959] env[62824]: WARNING nova.compute.manager [req-944a0b2b-cecd-4400-a0ec-24c312103dd4 req-a70cdd0d-9fac-4b4f-98a6-396d42241e4a service nova] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Received unexpected event network-vif-plugged-37654518-0a1a-44e2-8297-2169c44d0aad for instance with vm_state building and task_state spawning. [ 2291.744332] env[62824]: DEBUG nova.network.neutron [-] [instance: feda7def-7788-4a60-a56a-9353a1475c54] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2291.802157] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-06d67375-ff14-44a9-b9aa-4c9319a67341 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2291.812498] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c701795-b649-450a-ae4d-3407b4a4c00f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2291.831378] env[62824]: DEBUG nova.network.neutron [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Successfully updated port: 37654518-0a1a-44e2-8297-2169c44d0aad {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2291.855780] env[62824]: DEBUG nova.compute.manager [req-112e615f-4b07-469e-8414-9b5ab2f35a01 req-daf38108-57c4-441f-819b-25be51a68493 service nova] [instance: feda7def-7788-4a60-a56a-9353a1475c54] Detach interface failed, port_id=b133d2a7-e5fb-44fd-b716-20381faecc3d, reason: Instance feda7def-7788-4a60-a56a-9353a1475c54 could not be found. {{(pid=62824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11482}} [ 2292.013953] env[62824]: DEBUG nova.scheduler.client.report [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2292.035617] env[62824]: DEBUG oslo_vmware.api [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5239ee27-a54a-523a-967a-065b0c3a7006, 'name': SearchDatastore_Task, 'duration_secs': 0.008815} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2292.035881] env[62824]: DEBUG oslo_concurrency.lockutils [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2292.036158] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 19bcead9-b27c-471a-b217-cf7bfe101ed8/19bcead9-b27c-471a-b217-cf7bfe101ed8.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2292.036412] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-16c8f352-f4c8-4352-90b0-c19ee11365d5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2292.042619] env[62824]: DEBUG oslo_vmware.api [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 2292.042619] env[62824]: value = "task-2146757" [ 2292.042619] env[62824]: _type = "Task" [ 2292.042619] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2292.050698] env[62824]: DEBUG oslo_vmware.api [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146757, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2292.246889] env[62824]: INFO nova.compute.manager [-] [instance: feda7def-7788-4a60-a56a-9353a1475c54] Took 1.26 seconds to deallocate network for instance. [ 2292.337413] env[62824]: DEBUG oslo_concurrency.lockutils [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquiring lock "refresh_cache-1805a6e7-48f2-47b4-b097-0644ba8217fa" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2292.337571] env[62824]: DEBUG oslo_concurrency.lockutils [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquired lock "refresh_cache-1805a6e7-48f2-47b4-b097-0644ba8217fa" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2292.337802] env[62824]: DEBUG nova.network.neutron [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2292.520112] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.304s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2292.520750] env[62824]: DEBUG nova.compute.manager [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2292.524610] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7fd912ef-d4ab-43f7-9196-f9d4a6f5c085 tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.322s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2292.524891] env[62824]: DEBUG nova.objects.instance [None req-7fd912ef-d4ab-43f7-9196-f9d4a6f5c085 tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Lazy-loading 'resources' on Instance uuid 71731b21-c302-4d66-a579-889165673712 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2292.553642] env[62824]: DEBUG oslo_vmware.api [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146757, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.44041} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2292.553722] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 19bcead9-b27c-471a-b217-cf7bfe101ed8/19bcead9-b27c-471a-b217-cf7bfe101ed8.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2292.553881] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 19bcead9-b27c-471a-b217-cf7bfe101ed8] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2292.554144] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7d878b48-cecf-4e0a-b3ec-9c177b958afa {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2292.560703] env[62824]: DEBUG oslo_vmware.api [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 2292.560703] env[62824]: value = "task-2146758" [ 2292.560703] env[62824]: _type = "Task" [ 2292.560703] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2292.568700] env[62824]: DEBUG oslo_vmware.api [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146758, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2292.753977] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5009cb31-7660-4a2f-869d-7d502893a00d tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2292.870163] env[62824]: DEBUG nova.network.neutron [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2293.025736] env[62824]: DEBUG nova.compute.utils [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2293.027549] env[62824]: DEBUG nova.compute.manager [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2293.027741] env[62824]: DEBUG nova.network.neutron [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: e1574266-e53b-4169-9db6-c66b84895edb] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2293.071028] env[62824]: DEBUG oslo_vmware.api [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146758, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060889} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2293.071938] env[62824]: DEBUG nova.network.neutron [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Updating instance_info_cache with network_info: [{"id": "37654518-0a1a-44e2-8297-2169c44d0aad", "address": "fa:16:3e:e9:a5:9f", "network": {"id": "12411632-b7f6-4544-b2e4-b04ae7337547", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-917106830-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4479921caf5f405b8fc49baad390a0e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "de7fa486-5f28-44ae-b0cf-72234ff87546", "external-id": "nsx-vlan-transportzone-229", "segmentation_id": 229, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap37654518-0a", "ovs_interfaceid": "37654518-0a1a-44e2-8297-2169c44d0aad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2293.075910] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 19bcead9-b27c-471a-b217-cf7bfe101ed8] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2293.077150] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74410aef-0e76-4b82-9101-7758cc39e548 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2293.101872] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 19bcead9-b27c-471a-b217-cf7bfe101ed8] Reconfiguring VM instance instance-00000063 to attach disk [datastore1] 19bcead9-b27c-471a-b217-cf7bfe101ed8/19bcead9-b27c-471a-b217-cf7bfe101ed8.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2293.103424] env[62824]: DEBUG nova.policy [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b989eb89498b45dab6604116870d2fa1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c987c5f844be42c99cc9d57a9ca84c31', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2293.107565] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-357c067c-90ff-455d-a492-1af8173687f3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2293.127692] env[62824]: DEBUG oslo_vmware.api [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 2293.127692] env[62824]: value = "task-2146759" [ 2293.127692] env[62824]: _type = "Task" [ 2293.127692] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2293.139062] env[62824]: DEBUG oslo_vmware.api [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146759, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2293.277644] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d36811b5-0bcc-4be7-81e4-fe7e628b5b5e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2293.287124] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d029aa61-1a24-4033-9b4a-e763f786a1e9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2293.319257] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-924ea1a7-63a7-4d0b-94aa-bfd52246b87d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2293.327257] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cf4abf7-ad74-4e03-8517-05e5dca91e7a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2293.340855] env[62824]: DEBUG nova.compute.provider_tree [None req-7fd912ef-d4ab-43f7-9196-f9d4a6f5c085 tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2293.535183] env[62824]: DEBUG nova.compute.manager [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2293.577013] env[62824]: DEBUG oslo_concurrency.lockutils [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Releasing lock "refresh_cache-1805a6e7-48f2-47b4-b097-0644ba8217fa" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2293.577395] env[62824]: DEBUG nova.compute.manager [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Instance network_info: |[{"id": "37654518-0a1a-44e2-8297-2169c44d0aad", "address": "fa:16:3e:e9:a5:9f", "network": {"id": "12411632-b7f6-4544-b2e4-b04ae7337547", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-917106830-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4479921caf5f405b8fc49baad390a0e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "de7fa486-5f28-44ae-b0cf-72234ff87546", "external-id": "nsx-vlan-transportzone-229", "segmentation_id": 229, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap37654518-0a", "ovs_interfaceid": "37654518-0a1a-44e2-8297-2169c44d0aad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2293.577821] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e9:a5:9f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'de7fa486-5f28-44ae-b0cf-72234ff87546', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '37654518-0a1a-44e2-8297-2169c44d0aad', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2293.585877] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Creating folder: Project (4479921caf5f405b8fc49baad390a0e9). Parent ref: group-v438503. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 2293.587877] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-249826d8-4adf-48fa-a3bb-55149cefb017 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2293.598165] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Created folder: Project (4479921caf5f405b8fc49baad390a0e9) in parent group-v438503. [ 2293.598165] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Creating folder: Instances. Parent ref: group-v438745. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 2293.598165] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3ea5b95a-9e97-4b2c-a8ec-e351ef7b5164 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2293.608323] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Created folder: Instances in parent group-v438745. [ 2293.608323] env[62824]: DEBUG oslo.service.loopingcall [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2293.608323] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2293.608323] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-38ee1c31-5ad3-47a2-9ebf-af02fe2e8c03 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2293.625583] env[62824]: DEBUG nova.network.neutron [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Successfully created port: d5c6ba87-e0c7-49c2-bbaf-b7c64f84cdf8 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2293.628872] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2293.628872] env[62824]: value = "task-2146762" [ 2293.628872] env[62824]: _type = "Task" [ 2293.628872] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2293.640421] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146762, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2293.643967] env[62824]: DEBUG oslo_vmware.api [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146759, 'name': ReconfigVM_Task, 'duration_secs': 0.400733} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2293.644310] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 19bcead9-b27c-471a-b217-cf7bfe101ed8] Reconfigured VM instance instance-00000063 to attach disk [datastore1] 19bcead9-b27c-471a-b217-cf7bfe101ed8/19bcead9-b27c-471a-b217-cf7bfe101ed8.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2293.644999] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6642ef2f-cdce-48f9-a6a6-504d475640b3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2293.651652] env[62824]: DEBUG oslo_vmware.api [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 2293.651652] env[62824]: value = "task-2146763" [ 2293.651652] env[62824]: _type = "Task" [ 2293.651652] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2293.664543] env[62824]: DEBUG oslo_vmware.api [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146763, 'name': Rename_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2293.757218] env[62824]: DEBUG nova.compute.manager [req-ff6d1a39-2c86-4d5b-9ca0-417c1ab38f5e req-d2503641-75a5-4587-9bc8-ddd5a197abc9 service nova] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Received event network-changed-37654518-0a1a-44e2-8297-2169c44d0aad {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2293.758187] env[62824]: DEBUG nova.compute.manager [req-ff6d1a39-2c86-4d5b-9ca0-417c1ab38f5e req-d2503641-75a5-4587-9bc8-ddd5a197abc9 service nova] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Refreshing instance network info cache due to event network-changed-37654518-0a1a-44e2-8297-2169c44d0aad. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2293.758270] env[62824]: DEBUG oslo_concurrency.lockutils [req-ff6d1a39-2c86-4d5b-9ca0-417c1ab38f5e req-d2503641-75a5-4587-9bc8-ddd5a197abc9 service nova] Acquiring lock "refresh_cache-1805a6e7-48f2-47b4-b097-0644ba8217fa" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2293.758548] env[62824]: DEBUG oslo_concurrency.lockutils [req-ff6d1a39-2c86-4d5b-9ca0-417c1ab38f5e req-d2503641-75a5-4587-9bc8-ddd5a197abc9 service nova] Acquired lock "refresh_cache-1805a6e7-48f2-47b4-b097-0644ba8217fa" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2293.758864] env[62824]: DEBUG nova.network.neutron [req-ff6d1a39-2c86-4d5b-9ca0-417c1ab38f5e req-d2503641-75a5-4587-9bc8-ddd5a197abc9 service nova] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Refreshing network info cache for port 37654518-0a1a-44e2-8297-2169c44d0aad {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2293.843953] env[62824]: DEBUG nova.scheduler.client.report [None req-7fd912ef-d4ab-43f7-9196-f9d4a6f5c085 tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2294.142490] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146762, 'name': CreateVM_Task, 'duration_secs': 0.506093} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2294.142673] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2294.143398] env[62824]: DEBUG oslo_concurrency.lockutils [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2294.143571] env[62824]: DEBUG oslo_concurrency.lockutils [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2294.143899] env[62824]: DEBUG oslo_concurrency.lockutils [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2294.144176] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-629dfa32-fca2-4edd-ae8b-7be6dcbd4aae {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2294.148993] env[62824]: DEBUG oslo_vmware.api [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2294.148993] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]527781cb-aa01-4b59-505d-c846ec6be7b7" [ 2294.148993] env[62824]: _type = "Task" [ 2294.148993] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2294.159754] env[62824]: DEBUG oslo_vmware.api [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]527781cb-aa01-4b59-505d-c846ec6be7b7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2294.163252] env[62824]: DEBUG oslo_vmware.api [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146763, 'name': Rename_Task, 'duration_secs': 0.342341} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2294.163522] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 19bcead9-b27c-471a-b217-cf7bfe101ed8] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2294.163788] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1ea6d841-569b-498a-aa36-27afee29851b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2294.169818] env[62824]: DEBUG oslo_vmware.api [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 2294.169818] env[62824]: value = "task-2146764" [ 2294.169818] env[62824]: _type = "Task" [ 2294.169818] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2294.177204] env[62824]: DEBUG oslo_vmware.api [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146764, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2294.353071] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7fd912ef-d4ab-43f7-9196-f9d4a6f5c085 tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.828s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2294.355443] env[62824]: DEBUG oslo_concurrency.lockutils [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.826s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2294.357378] env[62824]: INFO nova.compute.claims [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2294.382666] env[62824]: INFO nova.scheduler.client.report [None req-7fd912ef-d4ab-43f7-9196-f9d4a6f5c085 tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Deleted allocations for instance 71731b21-c302-4d66-a579-889165673712 [ 2294.546262] env[62824]: DEBUG nova.compute.manager [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2294.572342] env[62824]: DEBUG nova.virt.hardware [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2294.572585] env[62824]: DEBUG nova.virt.hardware [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2294.572742] env[62824]: DEBUG nova.virt.hardware [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2294.575310] env[62824]: DEBUG nova.virt.hardware [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2294.575310] env[62824]: DEBUG nova.virt.hardware [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2294.575310] env[62824]: DEBUG nova.virt.hardware [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2294.575310] env[62824]: DEBUG nova.virt.hardware [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2294.575310] env[62824]: DEBUG nova.virt.hardware [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2294.575310] env[62824]: DEBUG nova.virt.hardware [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2294.575310] env[62824]: DEBUG nova.virt.hardware [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2294.575310] env[62824]: DEBUG nova.virt.hardware [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2294.575310] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50f0197b-cc64-45f0-8f0a-2aba762d645c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2294.583517] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-165cb807-8396-4d63-a2bc-c31b0c47d7e6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2294.598500] env[62824]: DEBUG nova.network.neutron [req-ff6d1a39-2c86-4d5b-9ca0-417c1ab38f5e req-d2503641-75a5-4587-9bc8-ddd5a197abc9 service nova] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Updated VIF entry in instance network info cache for port 37654518-0a1a-44e2-8297-2169c44d0aad. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2294.598851] env[62824]: DEBUG nova.network.neutron [req-ff6d1a39-2c86-4d5b-9ca0-417c1ab38f5e req-d2503641-75a5-4587-9bc8-ddd5a197abc9 service nova] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Updating instance_info_cache with network_info: [{"id": "37654518-0a1a-44e2-8297-2169c44d0aad", "address": "fa:16:3e:e9:a5:9f", "network": {"id": "12411632-b7f6-4544-b2e4-b04ae7337547", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-917106830-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4479921caf5f405b8fc49baad390a0e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "de7fa486-5f28-44ae-b0cf-72234ff87546", "external-id": "nsx-vlan-transportzone-229", "segmentation_id": 229, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap37654518-0a", "ovs_interfaceid": "37654518-0a1a-44e2-8297-2169c44d0aad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2294.659955] env[62824]: DEBUG oslo_vmware.api [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]527781cb-aa01-4b59-505d-c846ec6be7b7, 'name': SearchDatastore_Task, 'duration_secs': 0.012623} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2294.660276] env[62824]: DEBUG oslo_concurrency.lockutils [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2294.660511] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2294.660750] env[62824]: DEBUG oslo_concurrency.lockutils [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2294.660892] env[62824]: DEBUG oslo_concurrency.lockutils [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2294.661087] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2294.661421] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e73c519c-6f2e-4855-b01b-a7e6bed633bb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2294.671323] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2294.671323] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2294.676359] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5c41adc0-f85c-48bb-8d3f-9e418f161220 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2294.682654] env[62824]: DEBUG oslo_vmware.api [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2294.682654] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]525dbcac-faf0-73aa-7539-aa45f71cb8b7" [ 2294.682654] env[62824]: _type = "Task" [ 2294.682654] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2294.685447] env[62824]: DEBUG oslo_vmware.api [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146764, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2294.692884] env[62824]: DEBUG oslo_vmware.api [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]525dbcac-faf0-73aa-7539-aa45f71cb8b7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2294.793047] env[62824]: INFO nova.compute.manager [None req-a0775373-e67f-4d6b-9ff2-c9ba99f3f454 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 15479062-af75-4925-99b3-77d6a49751ad] Get console output [ 2294.793047] env[62824]: WARNING nova.virt.vmwareapi.driver [None req-a0775373-e67f-4d6b-9ff2-c9ba99f3f454 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 15479062-af75-4925-99b3-77d6a49751ad] The console log is missing. Check your VSPC configuration [ 2294.890760] env[62824]: DEBUG oslo_concurrency.lockutils [None req-7fd912ef-d4ab-43f7-9196-f9d4a6f5c085 tempest-ServerShowV257Test-153047016 tempest-ServerShowV257Test-153047016-project-member] Lock "71731b21-c302-4d66-a579-889165673712" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.954s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2295.102063] env[62824]: DEBUG oslo_concurrency.lockutils [req-ff6d1a39-2c86-4d5b-9ca0-417c1ab38f5e req-d2503641-75a5-4587-9bc8-ddd5a197abc9 service nova] Releasing lock "refresh_cache-1805a6e7-48f2-47b4-b097-0644ba8217fa" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2295.169054] env[62824]: DEBUG nova.compute.manager [req-4c3c4ad6-50b4-4fc3-b2b0-e503c9c68510 req-777c8a65-2db4-4a71-b3a0-d6ad47115447 service nova] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Received event network-vif-plugged-d5c6ba87-e0c7-49c2-bbaf-b7c64f84cdf8 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2295.169054] env[62824]: DEBUG oslo_concurrency.lockutils [req-4c3c4ad6-50b4-4fc3-b2b0-e503c9c68510 req-777c8a65-2db4-4a71-b3a0-d6ad47115447 service nova] Acquiring lock "e1574266-e53b-4169-9db6-c66b84895edb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2295.169054] env[62824]: DEBUG oslo_concurrency.lockutils [req-4c3c4ad6-50b4-4fc3-b2b0-e503c9c68510 req-777c8a65-2db4-4a71-b3a0-d6ad47115447 service nova] Lock "e1574266-e53b-4169-9db6-c66b84895edb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2295.169054] env[62824]: DEBUG oslo_concurrency.lockutils [req-4c3c4ad6-50b4-4fc3-b2b0-e503c9c68510 req-777c8a65-2db4-4a71-b3a0-d6ad47115447 service nova] Lock "e1574266-e53b-4169-9db6-c66b84895edb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2295.169956] env[62824]: DEBUG nova.compute.manager [req-4c3c4ad6-50b4-4fc3-b2b0-e503c9c68510 req-777c8a65-2db4-4a71-b3a0-d6ad47115447 service nova] [instance: e1574266-e53b-4169-9db6-c66b84895edb] No waiting events found dispatching network-vif-plugged-d5c6ba87-e0c7-49c2-bbaf-b7c64f84cdf8 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2295.170352] env[62824]: WARNING nova.compute.manager [req-4c3c4ad6-50b4-4fc3-b2b0-e503c9c68510 req-777c8a65-2db4-4a71-b3a0-d6ad47115447 service nova] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Received unexpected event network-vif-plugged-d5c6ba87-e0c7-49c2-bbaf-b7c64f84cdf8 for instance with vm_state building and task_state spawning. [ 2295.182863] env[62824]: DEBUG oslo_vmware.api [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146764, 'name': PowerOnVM_Task, 'duration_secs': 0.842493} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2295.182863] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 19bcead9-b27c-471a-b217-cf7bfe101ed8] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2295.183184] env[62824]: INFO nova.compute.manager [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 19bcead9-b27c-471a-b217-cf7bfe101ed8] Took 14.47 seconds to spawn the instance on the hypervisor. [ 2295.183413] env[62824]: DEBUG nova.compute.manager [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 19bcead9-b27c-471a-b217-cf7bfe101ed8] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2295.184411] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97a5d076-200d-4d8e-8775-851db6cff0f4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2295.201488] env[62824]: DEBUG oslo_vmware.api [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]525dbcac-faf0-73aa-7539-aa45f71cb8b7, 'name': SearchDatastore_Task, 'duration_secs': 0.01218} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2295.206021] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-81401b65-c9fd-4934-9eda-95691037b704 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2295.208196] env[62824]: DEBUG oslo_vmware.api [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2295.208196] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52d9feb0-9bb0-007e-4935-b7b40b0afc23" [ 2295.208196] env[62824]: _type = "Task" [ 2295.208196] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2295.216026] env[62824]: DEBUG nova.network.neutron [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Successfully updated port: d5c6ba87-e0c7-49c2-bbaf-b7c64f84cdf8 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2295.224725] env[62824]: DEBUG oslo_vmware.api [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52d9feb0-9bb0-007e-4935-b7b40b0afc23, 'name': SearchDatastore_Task, 'duration_secs': 0.010752} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2295.224806] env[62824]: DEBUG oslo_concurrency.lockutils [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2295.226193] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 1805a6e7-48f2-47b4-b097-0644ba8217fa/1805a6e7-48f2-47b4-b097-0644ba8217fa.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2295.226966] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-86ad1d30-852a-4033-b399-d9425bc37279 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2295.235692] env[62824]: DEBUG oslo_vmware.api [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2295.235692] env[62824]: value = "task-2146765" [ 2295.235692] env[62824]: _type = "Task" [ 2295.235692] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2295.244998] env[62824]: DEBUG oslo_vmware.api [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146765, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2295.584209] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb67dd4c-bd8d-4db1-8fc8-e1600a70f7e1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2295.594212] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9679c00c-b292-474a-9913-504b5b48f96f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2295.633786] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cecb7af0-356d-4f86-9377-4733851dd342 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2295.642523] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e0c3569-d96f-4b47-96c7-b72d8c22f4e5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2295.658551] env[62824]: DEBUG nova.compute.provider_tree [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2295.712731] env[62824]: INFO nova.compute.manager [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 19bcead9-b27c-471a-b217-cf7bfe101ed8] Took 19.40 seconds to build instance. [ 2295.716417] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquiring lock "refresh_cache-e1574266-e53b-4169-9db6-c66b84895edb" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2295.716667] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquired lock "refresh_cache-e1574266-e53b-4169-9db6-c66b84895edb" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2295.716787] env[62824]: DEBUG nova.network.neutron [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2295.745940] env[62824]: DEBUG oslo_vmware.api [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146765, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2295.872558] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e6d89d95-2eeb-4b1c-b1c1-6130619cae05 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Acquiring lock "15479062-af75-4925-99b3-77d6a49751ad" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2295.874155] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e6d89d95-2eeb-4b1c-b1c1-6130619cae05 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Lock "15479062-af75-4925-99b3-77d6a49751ad" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2295.874155] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e6d89d95-2eeb-4b1c-b1c1-6130619cae05 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Acquiring lock "15479062-af75-4925-99b3-77d6a49751ad-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2295.874155] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e6d89d95-2eeb-4b1c-b1c1-6130619cae05 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Lock "15479062-af75-4925-99b3-77d6a49751ad-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2295.874155] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e6d89d95-2eeb-4b1c-b1c1-6130619cae05 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Lock "15479062-af75-4925-99b3-77d6a49751ad-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2295.875911] env[62824]: INFO nova.compute.manager [None req-e6d89d95-2eeb-4b1c-b1c1-6130619cae05 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 15479062-af75-4925-99b3-77d6a49751ad] Terminating instance [ 2296.165243] env[62824]: DEBUG nova.scheduler.client.report [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2296.215396] env[62824]: DEBUG oslo_concurrency.lockutils [None req-af6a4bf7-5ad2-469a-aba3-12ffa3502c24 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "19bcead9-b27c-471a-b217-cf7bfe101ed8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.909s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2296.249215] env[62824]: DEBUG oslo_vmware.api [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146765, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.523812} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2296.249215] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 1805a6e7-48f2-47b4-b097-0644ba8217fa/1805a6e7-48f2-47b4-b097-0644ba8217fa.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2296.249215] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2296.249215] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1de481fa-c82f-44de-b86e-b3ad24145a1d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2296.256130] env[62824]: DEBUG nova.network.neutron [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2296.257603] env[62824]: DEBUG oslo_vmware.api [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2296.257603] env[62824]: value = "task-2146766" [ 2296.257603] env[62824]: _type = "Task" [ 2296.257603] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2296.265651] env[62824]: DEBUG oslo_vmware.api [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146766, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2296.380076] env[62824]: DEBUG nova.compute.manager [None req-e6d89d95-2eeb-4b1c-b1c1-6130619cae05 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 15479062-af75-4925-99b3-77d6a49751ad] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2296.380348] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e6d89d95-2eeb-4b1c-b1c1-6130619cae05 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 15479062-af75-4925-99b3-77d6a49751ad] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2296.381241] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4a7e94e-e84f-4bc2-909f-31352ba62111 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2296.390622] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6d89d95-2eeb-4b1c-b1c1-6130619cae05 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 15479062-af75-4925-99b3-77d6a49751ad] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2296.390881] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7a2ad0b4-0702-4f7d-a0eb-4f478b16555f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2296.396639] env[62824]: DEBUG oslo_vmware.api [None req-e6d89d95-2eeb-4b1c-b1c1-6130619cae05 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Waiting for the task: (returnval){ [ 2296.396639] env[62824]: value = "task-2146767" [ 2296.396639] env[62824]: _type = "Task" [ 2296.396639] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2296.404542] env[62824]: DEBUG oslo_vmware.api [None req-e6d89d95-2eeb-4b1c-b1c1-6130619cae05 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Task: {'id': task-2146767, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2296.405457] env[62824]: DEBUG nova.network.neutron [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Updating instance_info_cache with network_info: [{"id": "d5c6ba87-e0c7-49c2-bbaf-b7c64f84cdf8", "address": "fa:16:3e:c4:9c:f4", "network": {"id": "aa1c09e5-8d72-43ad-a903-a0f5e711da80", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1643616572-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c987c5f844be42c99cc9d57a9ca84c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9bc2632-36f9-4912-8782-8bbb789f909d", "external-id": "nsx-vlan-transportzone-897", "segmentation_id": 897, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd5c6ba87-e0", "ovs_interfaceid": "d5c6ba87-e0c7-49c2-bbaf-b7c64f84cdf8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2296.651446] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dc3c20b-303f-41de-8f0e-6f479b7149ec {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2296.660058] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-fee4870d-ea55-443f-91a4-3b6544589b80 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 19bcead9-b27c-471a-b217-cf7bfe101ed8] Suspending the VM {{(pid=62824) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 2296.660275] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-a80f9731-5d64-4be3-ae4b-784421bf82ef {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2296.667139] env[62824]: DEBUG oslo_concurrency.lockutils [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.312s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2296.667668] env[62824]: DEBUG nova.compute.manager [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2296.670447] env[62824]: DEBUG oslo_vmware.api [None req-fee4870d-ea55-443f-91a4-3b6544589b80 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 2296.670447] env[62824]: value = "task-2146768" [ 2296.670447] env[62824]: _type = "Task" [ 2296.670447] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2296.670907] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5009cb31-7660-4a2f-869d-7d502893a00d tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.917s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2296.671174] env[62824]: DEBUG nova.objects.instance [None req-5009cb31-7660-4a2f-869d-7d502893a00d tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lazy-loading 'resources' on Instance uuid feda7def-7788-4a60-a56a-9353a1475c54 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2296.682046] env[62824]: DEBUG oslo_vmware.api [None req-fee4870d-ea55-443f-91a4-3b6544589b80 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146768, 'name': SuspendVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2296.767380] env[62824]: DEBUG oslo_vmware.api [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146766, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066446} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2296.767699] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2296.768536] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-210d132b-2de9-4011-8c9f-8b4e1eda1f3b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2296.792957] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Reconfiguring VM instance instance-00000064 to attach disk [datastore1] 1805a6e7-48f2-47b4-b097-0644ba8217fa/1805a6e7-48f2-47b4-b097-0644ba8217fa.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2296.793326] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c77f5deb-5221-4807-9ac5-698356c2ed7a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2296.813528] env[62824]: DEBUG oslo_vmware.api [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2296.813528] env[62824]: value = "task-2146769" [ 2296.813528] env[62824]: _type = "Task" [ 2296.813528] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2296.824239] env[62824]: DEBUG oslo_vmware.api [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146769, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2296.906166] env[62824]: DEBUG oslo_vmware.api [None req-e6d89d95-2eeb-4b1c-b1c1-6130619cae05 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Task: {'id': task-2146767, 'name': PowerOffVM_Task, 'duration_secs': 0.238045} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2296.906465] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6d89d95-2eeb-4b1c-b1c1-6130619cae05 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 15479062-af75-4925-99b3-77d6a49751ad] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2296.906638] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e6d89d95-2eeb-4b1c-b1c1-6130619cae05 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 15479062-af75-4925-99b3-77d6a49751ad] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2296.906897] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9383b67b-f3ce-4e79-8219-90f996d48156 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2296.908762] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Releasing lock "refresh_cache-e1574266-e53b-4169-9db6-c66b84895edb" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2296.909079] env[62824]: DEBUG nova.compute.manager [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Instance network_info: |[{"id": "d5c6ba87-e0c7-49c2-bbaf-b7c64f84cdf8", "address": "fa:16:3e:c4:9c:f4", "network": {"id": "aa1c09e5-8d72-43ad-a903-a0f5e711da80", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1643616572-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c987c5f844be42c99cc9d57a9ca84c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9bc2632-36f9-4912-8782-8bbb789f909d", "external-id": "nsx-vlan-transportzone-897", "segmentation_id": 897, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd5c6ba87-e0", "ovs_interfaceid": "d5c6ba87-e0c7-49c2-bbaf-b7c64f84cdf8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2296.909458] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c4:9c:f4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c9bc2632-36f9-4912-8782-8bbb789f909d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd5c6ba87-e0c7-49c2-bbaf-b7c64f84cdf8', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2296.916871] env[62824]: DEBUG oslo.service.loopingcall [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2296.917541] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2296.917798] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-76d69519-a315-4256-9e48-69448b414c08 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2296.937503] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2296.937503] env[62824]: value = "task-2146771" [ 2296.937503] env[62824]: _type = "Task" [ 2296.937503] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2296.945454] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146771, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2297.173052] env[62824]: DEBUG nova.compute.utils [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2297.174871] env[62824]: DEBUG nova.compute.manager [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2297.174997] env[62824]: DEBUG nova.network.neutron [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2297.188238] env[62824]: DEBUG oslo_vmware.api [None req-fee4870d-ea55-443f-91a4-3b6544589b80 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146768, 'name': SuspendVM_Task} progress is 70%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2297.196626] env[62824]: DEBUG nova.compute.manager [req-ad225858-f4a2-4cfc-a75d-4c031fa441f4 req-515e97be-8a9a-47b9-95d0-a7426f272320 service nova] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Received event network-changed-d5c6ba87-e0c7-49c2-bbaf-b7c64f84cdf8 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2297.196626] env[62824]: DEBUG nova.compute.manager [req-ad225858-f4a2-4cfc-a75d-4c031fa441f4 req-515e97be-8a9a-47b9-95d0-a7426f272320 service nova] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Refreshing instance network info cache due to event network-changed-d5c6ba87-e0c7-49c2-bbaf-b7c64f84cdf8. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2297.196755] env[62824]: DEBUG oslo_concurrency.lockutils [req-ad225858-f4a2-4cfc-a75d-4c031fa441f4 req-515e97be-8a9a-47b9-95d0-a7426f272320 service nova] Acquiring lock "refresh_cache-e1574266-e53b-4169-9db6-c66b84895edb" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2297.196869] env[62824]: DEBUG oslo_concurrency.lockutils [req-ad225858-f4a2-4cfc-a75d-4c031fa441f4 req-515e97be-8a9a-47b9-95d0-a7426f272320 service nova] Acquired lock "refresh_cache-e1574266-e53b-4169-9db6-c66b84895edb" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2297.196963] env[62824]: DEBUG nova.network.neutron [req-ad225858-f4a2-4cfc-a75d-4c031fa441f4 req-515e97be-8a9a-47b9-95d0-a7426f272320 service nova] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Refreshing network info cache for port d5c6ba87-e0c7-49c2-bbaf-b7c64f84cdf8 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2297.233344] env[62824]: DEBUG nova.policy [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9cc01135021d4a90a718dc605ce98af2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '815ecf9340934733a83d4dcb26612fd4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2297.324950] env[62824]: DEBUG oslo_vmware.api [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146769, 'name': ReconfigVM_Task, 'duration_secs': 0.381586} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2297.325284] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Reconfigured VM instance instance-00000064 to attach disk [datastore1] 1805a6e7-48f2-47b4-b097-0644ba8217fa/1805a6e7-48f2-47b4-b097-0644ba8217fa.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2297.325880] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2944d459-66b0-499f-a390-6cd2d989688e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2297.332349] env[62824]: DEBUG oslo_vmware.api [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2297.332349] env[62824]: value = "task-2146772" [ 2297.332349] env[62824]: _type = "Task" [ 2297.332349] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2297.343616] env[62824]: DEBUG oslo_vmware.api [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146772, 'name': Rename_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2297.385665] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5fe6f0f-5049-44d4-8910-9356745f3974 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2297.394056] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-958bb03e-21c6-41ea-9e01-52632873af44 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2297.432780] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02c7bfcb-725d-4bf2-bc13-a0fe47f93f6b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2297.443382] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee988dd6-760d-42f8-bdcc-4bc999f5ce77 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2297.451970] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146771, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2297.459709] env[62824]: DEBUG nova.compute.provider_tree [None req-5009cb31-7660-4a2f-869d-7d502893a00d tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2297.513865] env[62824]: DEBUG nova.network.neutron [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Successfully created port: f1800205-8a92-425e-9f64-1dabd5809221 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2297.606348] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e6d89d95-2eeb-4b1c-b1c1-6130619cae05 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 15479062-af75-4925-99b3-77d6a49751ad] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2297.606531] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e6d89d95-2eeb-4b1c-b1c1-6130619cae05 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 15479062-af75-4925-99b3-77d6a49751ad] Deleting contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2297.606717] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6d89d95-2eeb-4b1c-b1c1-6130619cae05 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Deleting the datastore file [datastore2] 15479062-af75-4925-99b3-77d6a49751ad {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2297.606977] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3cb334c4-6e4c-4d14-abbb-698e4a6f0bd2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2297.614221] env[62824]: DEBUG oslo_vmware.api [None req-e6d89d95-2eeb-4b1c-b1c1-6130619cae05 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Waiting for the task: (returnval){ [ 2297.614221] env[62824]: value = "task-2146773" [ 2297.614221] env[62824]: _type = "Task" [ 2297.614221] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2297.622591] env[62824]: DEBUG oslo_vmware.api [None req-e6d89d95-2eeb-4b1c-b1c1-6130619cae05 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Task: {'id': task-2146773, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2297.678825] env[62824]: DEBUG nova.compute.manager [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2297.689063] env[62824]: DEBUG oslo_vmware.api [None req-fee4870d-ea55-443f-91a4-3b6544589b80 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146768, 'name': SuspendVM_Task, 'duration_secs': 0.609048} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2297.689338] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-fee4870d-ea55-443f-91a4-3b6544589b80 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 19bcead9-b27c-471a-b217-cf7bfe101ed8] Suspended the VM {{(pid=62824) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 2297.689520] env[62824]: DEBUG nova.compute.manager [None req-fee4870d-ea55-443f-91a4-3b6544589b80 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 19bcead9-b27c-471a-b217-cf7bfe101ed8] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2297.690426] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8367c427-c5a3-484f-ac4b-439c9a30f7cf {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2297.842289] env[62824]: DEBUG oslo_vmware.api [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146772, 'name': Rename_Task, 'duration_secs': 0.26291} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2297.842564] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2297.842812] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a10ec2ad-78a5-4f06-88d2-d2d3bbdad2c1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2297.848472] env[62824]: DEBUG oslo_vmware.api [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2297.848472] env[62824]: value = "task-2146774" [ 2297.848472] env[62824]: _type = "Task" [ 2297.848472] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2297.856095] env[62824]: DEBUG oslo_vmware.api [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146774, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2297.948800] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146771, 'name': CreateVM_Task, 'duration_secs': 0.725219} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2297.948989] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2297.949642] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2297.949833] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2297.950169] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2297.950418] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f0a6c473-850f-46bd-a46e-6e8587c240d6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2297.954753] env[62824]: DEBUG oslo_vmware.api [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Waiting for the task: (returnval){ [ 2297.954753] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5226809c-2fe7-5c5a-35fe-b39e973e8f76" [ 2297.954753] env[62824]: _type = "Task" [ 2297.954753] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2297.962939] env[62824]: DEBUG nova.scheduler.client.report [None req-5009cb31-7660-4a2f-869d-7d502893a00d tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2297.965955] env[62824]: DEBUG oslo_vmware.api [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5226809c-2fe7-5c5a-35fe-b39e973e8f76, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2298.123700] env[62824]: DEBUG oslo_vmware.api [None req-e6d89d95-2eeb-4b1c-b1c1-6130619cae05 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Task: {'id': task-2146773, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.199292} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2298.124071] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6d89d95-2eeb-4b1c-b1c1-6130619cae05 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2298.124203] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e6d89d95-2eeb-4b1c-b1c1-6130619cae05 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 15479062-af75-4925-99b3-77d6a49751ad] Deleted contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2298.124396] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e6d89d95-2eeb-4b1c-b1c1-6130619cae05 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 15479062-af75-4925-99b3-77d6a49751ad] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2298.124570] env[62824]: INFO nova.compute.manager [None req-e6d89d95-2eeb-4b1c-b1c1-6130619cae05 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] [instance: 15479062-af75-4925-99b3-77d6a49751ad] Took 1.74 seconds to destroy the instance on the hypervisor. [ 2298.124910] env[62824]: DEBUG oslo.service.loopingcall [None req-e6d89d95-2eeb-4b1c-b1c1-6130619cae05 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2298.124989] env[62824]: DEBUG nova.compute.manager [-] [instance: 15479062-af75-4925-99b3-77d6a49751ad] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2298.125100] env[62824]: DEBUG nova.network.neutron [-] [instance: 15479062-af75-4925-99b3-77d6a49751ad] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2298.358864] env[62824]: DEBUG oslo_vmware.api [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146774, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2298.432194] env[62824]: DEBUG nova.network.neutron [req-ad225858-f4a2-4cfc-a75d-4c031fa441f4 req-515e97be-8a9a-47b9-95d0-a7426f272320 service nova] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Updated VIF entry in instance network info cache for port d5c6ba87-e0c7-49c2-bbaf-b7c64f84cdf8. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2298.432625] env[62824]: DEBUG nova.network.neutron [req-ad225858-f4a2-4cfc-a75d-4c031fa441f4 req-515e97be-8a9a-47b9-95d0-a7426f272320 service nova] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Updating instance_info_cache with network_info: [{"id": "d5c6ba87-e0c7-49c2-bbaf-b7c64f84cdf8", "address": "fa:16:3e:c4:9c:f4", "network": {"id": "aa1c09e5-8d72-43ad-a903-a0f5e711da80", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1643616572-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c987c5f844be42c99cc9d57a9ca84c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9bc2632-36f9-4912-8782-8bbb789f909d", "external-id": "nsx-vlan-transportzone-897", "segmentation_id": 897, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd5c6ba87-e0", "ovs_interfaceid": "d5c6ba87-e0c7-49c2-bbaf-b7c64f84cdf8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2298.467595] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5009cb31-7660-4a2f-869d-7d502893a00d tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.797s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2298.469834] env[62824]: DEBUG oslo_vmware.api [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5226809c-2fe7-5c5a-35fe-b39e973e8f76, 'name': SearchDatastore_Task, 'duration_secs': 0.009773} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2298.472411] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2298.472411] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2298.472411] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2298.472411] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2298.472411] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2298.472411] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-79d1c7e0-91fa-49e2-8ae2-ae433827ca60 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2298.481039] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2298.481255] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2298.482063] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4ca844c3-0156-45a2-9e06-2337c23f16b8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2298.488859] env[62824]: DEBUG oslo_vmware.api [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Waiting for the task: (returnval){ [ 2298.488859] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52745782-e2ec-d0a6-9060-246053a6a98c" [ 2298.488859] env[62824]: _type = "Task" [ 2298.488859] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2298.499481] env[62824]: DEBUG oslo_vmware.api [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52745782-e2ec-d0a6-9060-246053a6a98c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2298.500887] env[62824]: INFO nova.scheduler.client.report [None req-5009cb31-7660-4a2f-869d-7d502893a00d tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Deleted allocations for instance feda7def-7788-4a60-a56a-9353a1475c54 [ 2298.691447] env[62824]: DEBUG nova.compute.manager [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2298.726438] env[62824]: DEBUG nova.virt.hardware [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2298.726438] env[62824]: DEBUG nova.virt.hardware [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2298.726671] env[62824]: DEBUG nova.virt.hardware [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2298.726671] env[62824]: DEBUG nova.virt.hardware [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2298.726886] env[62824]: DEBUG nova.virt.hardware [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2298.727163] env[62824]: DEBUG nova.virt.hardware [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2298.727438] env[62824]: DEBUG nova.virt.hardware [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2298.727776] env[62824]: DEBUG nova.virt.hardware [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2298.728015] env[62824]: DEBUG nova.virt.hardware [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2298.728322] env[62824]: DEBUG nova.virt.hardware [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2298.728550] env[62824]: DEBUG nova.virt.hardware [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2298.729708] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d22c0d50-9f4a-495a-87c4-2e85b5bdd196 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2298.737816] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8ecd21a-43e7-4731-a885-76a35238f854 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2298.859911] env[62824]: DEBUG oslo_vmware.api [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146774, 'name': PowerOnVM_Task, 'duration_secs': 0.657506} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2298.860204] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2298.860412] env[62824]: INFO nova.compute.manager [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Took 7.64 seconds to spawn the instance on the hypervisor. [ 2298.860592] env[62824]: DEBUG nova.compute.manager [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2298.862117] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b779c6fd-1a21-45f9-826b-f2198bc3e77c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2298.934947] env[62824]: DEBUG oslo_concurrency.lockutils [req-ad225858-f4a2-4cfc-a75d-4c031fa441f4 req-515e97be-8a9a-47b9-95d0-a7426f272320 service nova] Releasing lock "refresh_cache-e1574266-e53b-4169-9db6-c66b84895edb" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2299.000193] env[62824]: DEBUG oslo_vmware.api [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52745782-e2ec-d0a6-9060-246053a6a98c, 'name': SearchDatastore_Task, 'duration_secs': 0.015433} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2299.002454] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c421b460-fd54-4dc9-b880-eedd4c8630f2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2299.009238] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5009cb31-7660-4a2f-869d-7d502893a00d tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "feda7def-7788-4a60-a56a-9353a1475c54" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.828s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2299.012837] env[62824]: DEBUG oslo_vmware.api [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Waiting for the task: (returnval){ [ 2299.012837] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52136a1d-982e-ff68-c200-8a188204136a" [ 2299.012837] env[62824]: _type = "Task" [ 2299.012837] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2299.027671] env[62824]: DEBUG oslo_vmware.api [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52136a1d-982e-ff68-c200-8a188204136a, 'name': SearchDatastore_Task, 'duration_secs': 0.009499} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2299.029158] env[62824]: DEBUG nova.compute.manager [req-ea872802-902c-48ae-b59a-bf50c5fe0b7b req-3e16ae56-e170-403e-b627-0eb87540bbc9 service nova] [instance: 15479062-af75-4925-99b3-77d6a49751ad] Received event network-vif-deleted-6b8b3a2d-f55e-4aa8-b995-4880c1adea60 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2299.029412] env[62824]: INFO nova.compute.manager [req-ea872802-902c-48ae-b59a-bf50c5fe0b7b req-3e16ae56-e170-403e-b627-0eb87540bbc9 service nova] [instance: 15479062-af75-4925-99b3-77d6a49751ad] Neutron deleted interface 6b8b3a2d-f55e-4aa8-b995-4880c1adea60; detaching it from the instance and deleting it from the info cache [ 2299.029592] env[62824]: DEBUG nova.network.neutron [req-ea872802-902c-48ae-b59a-bf50c5fe0b7b req-3e16ae56-e170-403e-b627-0eb87540bbc9 service nova] [instance: 15479062-af75-4925-99b3-77d6a49751ad] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2299.031119] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2299.031373] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] e1574266-e53b-4169-9db6-c66b84895edb/e1574266-e53b-4169-9db6-c66b84895edb.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2299.032007] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e5c49d81-6018-4030-848c-ef239dc41bd3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2299.042515] env[62824]: DEBUG oslo_vmware.api [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Waiting for the task: (returnval){ [ 2299.042515] env[62824]: value = "task-2146775" [ 2299.042515] env[62824]: _type = "Task" [ 2299.042515] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2299.053343] env[62824]: DEBUG oslo_vmware.api [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146775, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2299.270453] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d9f6b187-61c9-454d-8fbe-17f8e63decd3 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquiring lock "19bcead9-b27c-471a-b217-cf7bfe101ed8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2299.270810] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d9f6b187-61c9-454d-8fbe-17f8e63decd3 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "19bcead9-b27c-471a-b217-cf7bfe101ed8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2299.271060] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d9f6b187-61c9-454d-8fbe-17f8e63decd3 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquiring lock "19bcead9-b27c-471a-b217-cf7bfe101ed8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2299.271261] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d9f6b187-61c9-454d-8fbe-17f8e63decd3 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "19bcead9-b27c-471a-b217-cf7bfe101ed8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2299.271482] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d9f6b187-61c9-454d-8fbe-17f8e63decd3 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "19bcead9-b27c-471a-b217-cf7bfe101ed8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2299.274077] env[62824]: INFO nova.compute.manager [None req-d9f6b187-61c9-454d-8fbe-17f8e63decd3 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 19bcead9-b27c-471a-b217-cf7bfe101ed8] Terminating instance [ 2299.379423] env[62824]: INFO nova.compute.manager [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Took 16.68 seconds to build instance. [ 2299.494262] env[62824]: DEBUG nova.network.neutron [-] [instance: 15479062-af75-4925-99b3-77d6a49751ad] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2299.532580] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-809c32f8-a5ba-4b11-be6a-3fa3a07fec54 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2299.543201] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fee070a1-1244-44b9-8fdd-0115e38d8f8a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2299.573020] env[62824]: DEBUG oslo_vmware.api [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146775, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.448397} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2299.573020] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] e1574266-e53b-4169-9db6-c66b84895edb/e1574266-e53b-4169-9db6-c66b84895edb.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2299.573020] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2299.573020] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-eea10a4d-4347-4b69-8b65-055cfa938f22 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2299.585673] env[62824]: DEBUG nova.compute.manager [req-ea872802-902c-48ae-b59a-bf50c5fe0b7b req-3e16ae56-e170-403e-b627-0eb87540bbc9 service nova] [instance: 15479062-af75-4925-99b3-77d6a49751ad] Detach interface failed, port_id=6b8b3a2d-f55e-4aa8-b995-4880c1adea60, reason: Instance 15479062-af75-4925-99b3-77d6a49751ad could not be found. {{(pid=62824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11482}} [ 2299.590033] env[62824]: DEBUG nova.network.neutron [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Successfully updated port: f1800205-8a92-425e-9f64-1dabd5809221 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2299.596670] env[62824]: DEBUG oslo_vmware.api [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Waiting for the task: (returnval){ [ 2299.596670] env[62824]: value = "task-2146776" [ 2299.596670] env[62824]: _type = "Task" [ 2299.596670] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2299.607029] env[62824]: DEBUG oslo_vmware.api [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146776, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2299.716474] env[62824]: DEBUG nova.compute.manager [req-f99fcea1-37ad-46dc-9727-7aee7b843b30 req-07607571-46d3-4862-bf74-71b97008cb0b service nova] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Received event network-vif-plugged-f1800205-8a92-425e-9f64-1dabd5809221 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2299.716716] env[62824]: DEBUG oslo_concurrency.lockutils [req-f99fcea1-37ad-46dc-9727-7aee7b843b30 req-07607571-46d3-4862-bf74-71b97008cb0b service nova] Acquiring lock "3d294ee9-92f6-4115-95db-9d4b66562b56-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2299.716898] env[62824]: DEBUG oslo_concurrency.lockutils [req-f99fcea1-37ad-46dc-9727-7aee7b843b30 req-07607571-46d3-4862-bf74-71b97008cb0b service nova] Lock "3d294ee9-92f6-4115-95db-9d4b66562b56-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2299.717066] env[62824]: DEBUG oslo_concurrency.lockutils [req-f99fcea1-37ad-46dc-9727-7aee7b843b30 req-07607571-46d3-4862-bf74-71b97008cb0b service nova] Lock "3d294ee9-92f6-4115-95db-9d4b66562b56-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2299.717243] env[62824]: DEBUG nova.compute.manager [req-f99fcea1-37ad-46dc-9727-7aee7b843b30 req-07607571-46d3-4862-bf74-71b97008cb0b service nova] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] No waiting events found dispatching network-vif-plugged-f1800205-8a92-425e-9f64-1dabd5809221 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2299.717409] env[62824]: WARNING nova.compute.manager [req-f99fcea1-37ad-46dc-9727-7aee7b843b30 req-07607571-46d3-4862-bf74-71b97008cb0b service nova] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Received unexpected event network-vif-plugged-f1800205-8a92-425e-9f64-1dabd5809221 for instance with vm_state building and task_state spawning. [ 2299.782487] env[62824]: DEBUG nova.compute.manager [None req-d9f6b187-61c9-454d-8fbe-17f8e63decd3 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 19bcead9-b27c-471a-b217-cf7bfe101ed8] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2299.782656] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-d9f6b187-61c9-454d-8fbe-17f8e63decd3 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 19bcead9-b27c-471a-b217-cf7bfe101ed8] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2299.783833] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e20cc89e-dd5e-4f81-afbe-bc1f2790bee3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2299.791692] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-d9f6b187-61c9-454d-8fbe-17f8e63decd3 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 19bcead9-b27c-471a-b217-cf7bfe101ed8] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2299.791939] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8398732f-9c95-4803-973e-612e52306d78 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2299.885416] env[62824]: DEBUG oslo_concurrency.lockutils [None req-31950ce1-53bb-489d-8197-86e9da6876f0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "1805a6e7-48f2-47b4-b097-0644ba8217fa" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.196s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2299.991784] env[62824]: INFO nova.compute.manager [-] [instance: 15479062-af75-4925-99b3-77d6a49751ad] Took 1.87 seconds to deallocate network for instance. [ 2300.049086] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-d9f6b187-61c9-454d-8fbe-17f8e63decd3 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 19bcead9-b27c-471a-b217-cf7bfe101ed8] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2300.049261] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-d9f6b187-61c9-454d-8fbe-17f8e63decd3 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 19bcead9-b27c-471a-b217-cf7bfe101ed8] Deleting contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2300.049568] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-d9f6b187-61c9-454d-8fbe-17f8e63decd3 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Deleting the datastore file [datastore1] 19bcead9-b27c-471a-b217-cf7bfe101ed8 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2300.049973] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-590cee70-063e-4efc-9b1a-61743840e1be {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2300.058167] env[62824]: DEBUG oslo_vmware.api [None req-d9f6b187-61c9-454d-8fbe-17f8e63decd3 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 2300.058167] env[62824]: value = "task-2146778" [ 2300.058167] env[62824]: _type = "Task" [ 2300.058167] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2300.070834] env[62824]: DEBUG oslo_vmware.api [None req-d9f6b187-61c9-454d-8fbe-17f8e63decd3 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146778, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2300.091808] env[62824]: DEBUG oslo_concurrency.lockutils [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquiring lock "refresh_cache-3d294ee9-92f6-4115-95db-9d4b66562b56" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2300.092052] env[62824]: DEBUG oslo_concurrency.lockutils [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquired lock "refresh_cache-3d294ee9-92f6-4115-95db-9d4b66562b56" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2300.092249] env[62824]: DEBUG nova.network.neutron [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2300.108253] env[62824]: DEBUG oslo_vmware.api [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146776, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067458} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2300.109019] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2300.109916] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09f1b53f-943c-4e54-bf99-2c3c436a21ec {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2300.135121] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Reconfiguring VM instance instance-00000065 to attach disk [datastore2] e1574266-e53b-4169-9db6-c66b84895edb/e1574266-e53b-4169-9db6-c66b84895edb.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2300.135847] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-76980ceb-5b86-4e44-ba0f-3abc6347c761 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2300.155771] env[62824]: DEBUG oslo_vmware.api [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Waiting for the task: (returnval){ [ 2300.155771] env[62824]: value = "task-2146779" [ 2300.155771] env[62824]: _type = "Task" [ 2300.155771] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2300.165684] env[62824]: DEBUG oslo_vmware.api [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146779, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2300.405240] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquiring lock "68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2300.405742] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2300.497673] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e6d89d95-2eeb-4b1c-b1c1-6130619cae05 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2300.498038] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e6d89d95-2eeb-4b1c-b1c1-6130619cae05 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2300.498277] env[62824]: DEBUG nova.objects.instance [None req-e6d89d95-2eeb-4b1c-b1c1-6130619cae05 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Lazy-loading 'resources' on Instance uuid 15479062-af75-4925-99b3-77d6a49751ad {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2300.503173] env[62824]: INFO nova.compute.manager [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Rebuilding instance [ 2300.546981] env[62824]: DEBUG nova.compute.manager [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2300.547965] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c0b750e-2370-4e79-894a-d87b334c1ffb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2300.566342] env[62824]: DEBUG oslo_vmware.api [None req-d9f6b187-61c9-454d-8fbe-17f8e63decd3 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146778, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.142266} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2300.566584] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-d9f6b187-61c9-454d-8fbe-17f8e63decd3 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2300.566764] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-d9f6b187-61c9-454d-8fbe-17f8e63decd3 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 19bcead9-b27c-471a-b217-cf7bfe101ed8] Deleted contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2300.566943] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-d9f6b187-61c9-454d-8fbe-17f8e63decd3 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 19bcead9-b27c-471a-b217-cf7bfe101ed8] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2300.567139] env[62824]: INFO nova.compute.manager [None req-d9f6b187-61c9-454d-8fbe-17f8e63decd3 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: 19bcead9-b27c-471a-b217-cf7bfe101ed8] Took 0.78 seconds to destroy the instance on the hypervisor. [ 2300.567371] env[62824]: DEBUG oslo.service.loopingcall [None req-d9f6b187-61c9-454d-8fbe-17f8e63decd3 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2300.567559] env[62824]: DEBUG nova.compute.manager [-] [instance: 19bcead9-b27c-471a-b217-cf7bfe101ed8] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2300.567656] env[62824]: DEBUG nova.network.neutron [-] [instance: 19bcead9-b27c-471a-b217-cf7bfe101ed8] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2300.640574] env[62824]: DEBUG nova.network.neutron [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2300.667140] env[62824]: DEBUG oslo_vmware.api [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146779, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2300.802273] env[62824]: DEBUG nova.network.neutron [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Updating instance_info_cache with network_info: [{"id": "f1800205-8a92-425e-9f64-1dabd5809221", "address": "fa:16:3e:17:9c:53", "network": {"id": "b4ed285c-26a7-461b-8a71-d18e07582ab8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-997150974-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "815ecf9340934733a83d4dcb26612fd4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69f65356-c85e-4b7f-ad28-7c7b5e8cf50c", "external-id": "nsx-vlan-transportzone-281", "segmentation_id": 281, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf1800205-8a", "ovs_interfaceid": "f1800205-8a92-425e-9f64-1dabd5809221", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2300.909091] env[62824]: DEBUG nova.compute.manager [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2300.952778] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bf852af9-823d-4ea1-8526-8b1688e7a098 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Acquiring lock "bbfcb2e3-9326-4548-b15b-e054cbfd192e" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2300.953238] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bf852af9-823d-4ea1-8526-8b1688e7a098 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Lock "bbfcb2e3-9326-4548-b15b-e054cbfd192e" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2301.169851] env[62824]: DEBUG oslo_vmware.api [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146779, 'name': ReconfigVM_Task, 'duration_secs': 0.90545} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2301.172539] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Reconfigured VM instance instance-00000065 to attach disk [datastore2] e1574266-e53b-4169-9db6-c66b84895edb/e1574266-e53b-4169-9db6-c66b84895edb.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2301.174846] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bd4b17d7-f7bc-44a1-ad09-4d46a79fbf68 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2301.180379] env[62824]: DEBUG oslo_vmware.api [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Waiting for the task: (returnval){ [ 2301.180379] env[62824]: value = "task-2146780" [ 2301.180379] env[62824]: _type = "Task" [ 2301.180379] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2301.190771] env[62824]: DEBUG oslo_vmware.api [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146780, 'name': Rename_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2301.213136] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c4cb545-45f2-40fb-8a51-a36d6e3244a9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2301.220632] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6210dcb0-ab44-43db-875f-0915720e9a9c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2301.254109] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ab12872-a4d3-472e-9669-d116258d1edc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2301.262334] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ddd0ee2-9773-459b-8d24-c40d6e6d1a8e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2301.275892] env[62824]: DEBUG nova.compute.provider_tree [None req-e6d89d95-2eeb-4b1c-b1c1-6130619cae05 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2301.305077] env[62824]: DEBUG oslo_concurrency.lockutils [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Releasing lock "refresh_cache-3d294ee9-92f6-4115-95db-9d4b66562b56" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2301.305345] env[62824]: DEBUG nova.compute.manager [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Instance network_info: |[{"id": "f1800205-8a92-425e-9f64-1dabd5809221", "address": "fa:16:3e:17:9c:53", "network": {"id": "b4ed285c-26a7-461b-8a71-d18e07582ab8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-997150974-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "815ecf9340934733a83d4dcb26612fd4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69f65356-c85e-4b7f-ad28-7c7b5e8cf50c", "external-id": "nsx-vlan-transportzone-281", "segmentation_id": 281, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf1800205-8a", "ovs_interfaceid": "f1800205-8a92-425e-9f64-1dabd5809221", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2301.305784] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:17:9c:53', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '69f65356-c85e-4b7f-ad28-7c7b5e8cf50c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f1800205-8a92-425e-9f64-1dabd5809221', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2301.313725] env[62824]: DEBUG oslo.service.loopingcall [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2301.313939] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2301.314204] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8a96ba6d-e703-4932-9eae-c133e4ed06e8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2301.331616] env[62824]: DEBUG nova.network.neutron [-] [instance: 19bcead9-b27c-471a-b217-cf7bfe101ed8] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2301.334248] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2301.334248] env[62824]: value = "task-2146781" [ 2301.334248] env[62824]: _type = "Task" [ 2301.334248] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2301.343275] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146781, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2301.434358] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2301.456554] env[62824]: INFO nova.compute.manager [None req-bf852af9-823d-4ea1-8526-8b1688e7a098 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Detaching volume 10a6ba42-78a0-4e92-b936-f88a95560707 [ 2301.492927] env[62824]: INFO nova.virt.block_device [None req-bf852af9-823d-4ea1-8526-8b1688e7a098 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Attempting to driver detach volume 10a6ba42-78a0-4e92-b936-f88a95560707 from mountpoint /dev/sdb [ 2301.493158] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-bf852af9-823d-4ea1-8526-8b1688e7a098 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Volume detach. Driver type: vmdk {{(pid=62824) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 2301.493348] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-bf852af9-823d-4ea1-8526-8b1688e7a098 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-438727', 'volume_id': '10a6ba42-78a0-4e92-b936-f88a95560707', 'name': 'volume-10a6ba42-78a0-4e92-b936-f88a95560707', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'bbfcb2e3-9326-4548-b15b-e054cbfd192e', 'attached_at': '', 'detached_at': '', 'volume_id': '10a6ba42-78a0-4e92-b936-f88a95560707', 'serial': '10a6ba42-78a0-4e92-b936-f88a95560707'} {{(pid=62824) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 2301.494206] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aeab6209-1fc1-472d-a4d7-a3b89ee090a0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2301.517157] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1f7f830-13cb-4d66-9098-d7e55285c723 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2301.524081] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-735669bc-b5a8-4f38-a4e7-4b6f7807cefb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2301.543871] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84976e85-5586-4f49-9f36-d897800dbb59 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2301.558580] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-bf852af9-823d-4ea1-8526-8b1688e7a098 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] The volume has not been displaced from its original location: [datastore2] volume-10a6ba42-78a0-4e92-b936-f88a95560707/volume-10a6ba42-78a0-4e92-b936-f88a95560707.vmdk. No consolidation needed. {{(pid=62824) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 2301.563895] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-bf852af9-823d-4ea1-8526-8b1688e7a098 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Reconfiguring VM instance instance-00000039 to detach disk 2001 {{(pid=62824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 2301.564434] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2301.564656] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c09afb29-201f-40bf-bd8f-130bce1db1bd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2301.582283] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-03af2d48-7e5e-49a7-9ef8-07db0d798fed {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2301.590174] env[62824]: DEBUG oslo_vmware.api [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2301.590174] env[62824]: value = "task-2146783" [ 2301.590174] env[62824]: _type = "Task" [ 2301.590174] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2301.591453] env[62824]: DEBUG oslo_vmware.api [None req-bf852af9-823d-4ea1-8526-8b1688e7a098 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for the task: (returnval){ [ 2301.591453] env[62824]: value = "task-2146782" [ 2301.591453] env[62824]: _type = "Task" [ 2301.591453] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2301.603601] env[62824]: DEBUG oslo_vmware.api [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146783, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2301.607416] env[62824]: DEBUG oslo_vmware.api [None req-bf852af9-823d-4ea1-8526-8b1688e7a098 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2146782, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2301.690654] env[62824]: DEBUG oslo_vmware.api [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146780, 'name': Rename_Task, 'duration_secs': 0.196485} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2301.691102] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2301.691474] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c894e1ab-7df8-48ad-beda-816d3c1f8788 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2301.698904] env[62824]: DEBUG oslo_vmware.api [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Waiting for the task: (returnval){ [ 2301.698904] env[62824]: value = "task-2146784" [ 2301.698904] env[62824]: _type = "Task" [ 2301.698904] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2301.708400] env[62824]: DEBUG oslo_vmware.api [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146784, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2301.748136] env[62824]: DEBUG nova.compute.manager [req-8f5c262b-6e8c-4b93-bf11-1dfc4eed1822 req-7a59d460-70c4-4b4a-be74-c0b85d3ae04a service nova] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Received event network-changed-f1800205-8a92-425e-9f64-1dabd5809221 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2301.748279] env[62824]: DEBUG nova.compute.manager [req-8f5c262b-6e8c-4b93-bf11-1dfc4eed1822 req-7a59d460-70c4-4b4a-be74-c0b85d3ae04a service nova] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Refreshing instance network info cache due to event network-changed-f1800205-8a92-425e-9f64-1dabd5809221. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2301.748612] env[62824]: DEBUG oslo_concurrency.lockutils [req-8f5c262b-6e8c-4b93-bf11-1dfc4eed1822 req-7a59d460-70c4-4b4a-be74-c0b85d3ae04a service nova] Acquiring lock "refresh_cache-3d294ee9-92f6-4115-95db-9d4b66562b56" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2301.749066] env[62824]: DEBUG oslo_concurrency.lockutils [req-8f5c262b-6e8c-4b93-bf11-1dfc4eed1822 req-7a59d460-70c4-4b4a-be74-c0b85d3ae04a service nova] Acquired lock "refresh_cache-3d294ee9-92f6-4115-95db-9d4b66562b56" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2301.749066] env[62824]: DEBUG nova.network.neutron [req-8f5c262b-6e8c-4b93-bf11-1dfc4eed1822 req-7a59d460-70c4-4b4a-be74-c0b85d3ae04a service nova] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Refreshing network info cache for port f1800205-8a92-425e-9f64-1dabd5809221 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2301.780173] env[62824]: DEBUG nova.scheduler.client.report [None req-e6d89d95-2eeb-4b1c-b1c1-6130619cae05 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2301.840880] env[62824]: INFO nova.compute.manager [-] [instance: 19bcead9-b27c-471a-b217-cf7bfe101ed8] Took 1.27 seconds to deallocate network for instance. [ 2301.850192] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146781, 'name': CreateVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2302.593845] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Acquiring lock "71e9733b-c84b-4501-8faf-a487f089b498" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2302.594147] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Lock "71e9733b-c84b-4501-8faf-a487f089b498" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2302.594320] env[62824]: INFO nova.compute.manager [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Shelving [ 2302.598035] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e6d89d95-2eeb-4b1c-b1c1-6130619cae05 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.100s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2302.600422] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d9f6b187-61c9-454d-8fbe-17f8e63decd3 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2302.608909] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.175s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2302.610778] env[62824]: INFO nova.compute.claims [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2302.628011] env[62824]: DEBUG oslo_vmware.api [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146784, 'name': PowerOnVM_Task} progress is 78%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2302.628288] env[62824]: DEBUG oslo_vmware.api [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146783, 'name': PowerOffVM_Task, 'duration_secs': 0.269225} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2302.628491] env[62824]: DEBUG oslo_vmware.api [None req-bf852af9-823d-4ea1-8526-8b1688e7a098 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2146782, 'name': ReconfigVM_Task, 'duration_secs': 0.280306} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2302.629391] env[62824]: INFO nova.scheduler.client.report [None req-e6d89d95-2eeb-4b1c-b1c1-6130619cae05 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Deleted allocations for instance 15479062-af75-4925-99b3-77d6a49751ad [ 2302.633152] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2302.633401] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2302.633720] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-bf852af9-823d-4ea1-8526-8b1688e7a098 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Reconfigured VM instance instance-00000039 to detach disk 2001 {{(pid=62824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 2302.638488] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146781, 'name': CreateVM_Task, 'duration_secs': 0.55689} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2302.641549] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b81f6f1-c0b6-463b-a700-ebff6f282646 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2302.644577] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e4c1b0ee-af76-4ecd-ac1e-03f69d5bf227 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2302.656320] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2302.659913] env[62824]: DEBUG oslo_concurrency.lockutils [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2302.660108] env[62824]: DEBUG oslo_concurrency.lockutils [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2302.660447] env[62824]: DEBUG oslo_concurrency.lockutils [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2302.661679] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f8f3059f-0d38-4f76-8679-ab6b9ad067d1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2302.666386] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2302.667624] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-39b5419c-d0ca-4cd3-a787-c65c9c17a15b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2302.669266] env[62824]: DEBUG oslo_vmware.api [None req-bf852af9-823d-4ea1-8526-8b1688e7a098 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for the task: (returnval){ [ 2302.669266] env[62824]: value = "task-2146785" [ 2302.669266] env[62824]: _type = "Task" [ 2302.669266] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2302.670761] env[62824]: DEBUG oslo_vmware.api [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2302.670761] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5295cdbd-0cc5-4517-c03a-64bda9c067b2" [ 2302.670761] env[62824]: _type = "Task" [ 2302.670761] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2302.683396] env[62824]: DEBUG oslo_vmware.api [None req-bf852af9-823d-4ea1-8526-8b1688e7a098 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2146785, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2302.686360] env[62824]: DEBUG oslo_vmware.api [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5295cdbd-0cc5-4517-c03a-64bda9c067b2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2302.845814] env[62824]: DEBUG nova.network.neutron [req-8f5c262b-6e8c-4b93-bf11-1dfc4eed1822 req-7a59d460-70c4-4b4a-be74-c0b85d3ae04a service nova] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Updated VIF entry in instance network info cache for port f1800205-8a92-425e-9f64-1dabd5809221. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2302.846213] env[62824]: DEBUG nova.network.neutron [req-8f5c262b-6e8c-4b93-bf11-1dfc4eed1822 req-7a59d460-70c4-4b4a-be74-c0b85d3ae04a service nova] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Updating instance_info_cache with network_info: [{"id": "f1800205-8a92-425e-9f64-1dabd5809221", "address": "fa:16:3e:17:9c:53", "network": {"id": "b4ed285c-26a7-461b-8a71-d18e07582ab8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-997150974-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "815ecf9340934733a83d4dcb26612fd4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69f65356-c85e-4b7f-ad28-7c7b5e8cf50c", "external-id": "nsx-vlan-transportzone-281", "segmentation_id": 281, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf1800205-8a", "ovs_interfaceid": "f1800205-8a92-425e-9f64-1dabd5809221", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2303.104625] env[62824]: DEBUG oslo_vmware.api [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146784, 'name': PowerOnVM_Task, 'duration_secs': 1.10034} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2303.104916] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2303.105135] env[62824]: INFO nova.compute.manager [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Took 8.56 seconds to spawn the instance on the hypervisor. [ 2303.105312] env[62824]: DEBUG nova.compute.manager [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2303.106064] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69411526-293e-42bd-af28-b8458cc30143 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2303.163739] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e6d89d95-2eeb-4b1c-b1c1-6130619cae05 tempest-ServersWithSpecificFlavorTestJSON-793074610 tempest-ServersWithSpecificFlavorTestJSON-793074610-project-member] Lock "15479062-af75-4925-99b3-77d6a49751ad" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.291s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2303.191399] env[62824]: DEBUG oslo_vmware.api [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5295cdbd-0cc5-4517-c03a-64bda9c067b2, 'name': SearchDatastore_Task, 'duration_secs': 0.013399} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2303.191720] env[62824]: DEBUG oslo_vmware.api [None req-bf852af9-823d-4ea1-8526-8b1688e7a098 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2146785, 'name': ReconfigVM_Task, 'duration_secs': 0.17126} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2303.192084] env[62824]: DEBUG oslo_concurrency.lockutils [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2303.192394] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2303.192721] env[62824]: DEBUG oslo_concurrency.lockutils [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2303.192929] env[62824]: DEBUG oslo_concurrency.lockutils [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2303.193195] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2303.193603] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-bf852af9-823d-4ea1-8526-8b1688e7a098 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-438727', 'volume_id': '10a6ba42-78a0-4e92-b936-f88a95560707', 'name': 'volume-10a6ba42-78a0-4e92-b936-f88a95560707', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'bbfcb2e3-9326-4548-b15b-e054cbfd192e', 'attached_at': '', 'detached_at': '', 'volume_id': '10a6ba42-78a0-4e92-b936-f88a95560707', 'serial': '10a6ba42-78a0-4e92-b936-f88a95560707'} {{(pid=62824) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 2303.196467] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2a19d48a-00a2-4674-abe2-6630a5a319e4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2303.205952] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2303.206215] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2303.207590] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3dc3a2d8-a787-44dc-8e71-310a5f1b592d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2303.214583] env[62824]: DEBUG oslo_vmware.api [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2303.214583] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]526f33e7-40d1-8637-d9b7-6a7c246048dd" [ 2303.214583] env[62824]: _type = "Task" [ 2303.214583] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2303.225076] env[62824]: DEBUG oslo_vmware.api [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]526f33e7-40d1-8637-d9b7-6a7c246048dd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2303.349145] env[62824]: DEBUG oslo_concurrency.lockutils [req-8f5c262b-6e8c-4b93-bf11-1dfc4eed1822 req-7a59d460-70c4-4b4a-be74-c0b85d3ae04a service nova] Releasing lock "refresh_cache-3d294ee9-92f6-4115-95db-9d4b66562b56" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2303.349309] env[62824]: DEBUG nova.compute.manager [req-8f5c262b-6e8c-4b93-bf11-1dfc4eed1822 req-7a59d460-70c4-4b4a-be74-c0b85d3ae04a service nova] [instance: 19bcead9-b27c-471a-b217-cf7bfe101ed8] Received event network-vif-deleted-f94cb6aa-5577-481c-aee8-c9399a84ba6e {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2303.614414] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2303.614414] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b0338e6f-d2fa-46f5-9b82-8d8c9c8bd5c9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2303.631582] env[62824]: DEBUG oslo_vmware.api [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Waiting for the task: (returnval){ [ 2303.631582] env[62824]: value = "task-2146787" [ 2303.631582] env[62824]: _type = "Task" [ 2303.631582] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2303.640409] env[62824]: INFO nova.compute.manager [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Took 16.34 seconds to build instance. [ 2303.648699] env[62824]: DEBUG oslo_vmware.api [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146787, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2303.724826] env[62824]: DEBUG oslo_vmware.api [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]526f33e7-40d1-8637-d9b7-6a7c246048dd, 'name': SearchDatastore_Task, 'duration_secs': 0.008103} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2303.728122] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-27d441e5-8f63-45a9-8fb3-85b6a6fc8d6b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2303.733660] env[62824]: DEBUG oslo_vmware.api [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2303.733660] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52e6fbab-9eec-0481-a8cc-213f3cb2b585" [ 2303.733660] env[62824]: _type = "Task" [ 2303.733660] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2303.741584] env[62824]: DEBUG nova.objects.instance [None req-bf852af9-823d-4ea1-8526-8b1688e7a098 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Lazy-loading 'flavor' on Instance uuid bbfcb2e3-9326-4548-b15b-e054cbfd192e {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2303.748330] env[62824]: DEBUG oslo_vmware.api [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52e6fbab-9eec-0481-a8cc-213f3cb2b585, 'name': SearchDatastore_Task, 'duration_secs': 0.010839} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2303.748818] env[62824]: DEBUG oslo_concurrency.lockutils [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2303.749754] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 3d294ee9-92f6-4115-95db-9d4b66562b56/3d294ee9-92f6-4115-95db-9d4b66562b56.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2303.749754] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-49951423-d750-474c-a69c-a99bfc142dbc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2303.756488] env[62824]: DEBUG oslo_vmware.api [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2303.756488] env[62824]: value = "task-2146788" [ 2303.756488] env[62824]: _type = "Task" [ 2303.756488] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2303.770219] env[62824]: DEBUG oslo_vmware.api [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146788, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2303.843587] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1b85ab7-27b6-4593-9618-5bf7a6d3b16a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2303.852268] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6643fe69-1492-46f4-be62-d28df8be2bec {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2303.891838] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9abbd378-5d37-4925-b14e-9b3a9debe3df {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2303.901164] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83f98a59-348b-414c-b7a3-51198af54e47 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2303.916664] env[62824]: DEBUG nova.compute.provider_tree [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2304.142735] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1435c9d5-a749-4e19-b828-2b540305d1b1 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lock "e1574266-e53b-4169-9db6-c66b84895edb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.850s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2304.150126] env[62824]: DEBUG oslo_vmware.api [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146787, 'name': PowerOffVM_Task, 'duration_secs': 0.225664} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2304.150557] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2304.151866] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3aa9f3f5-74f6-4f43-a8a7-81c69c362bc1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2304.179707] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bcbab22-9910-42a8-8601-1ae2d56c55ed {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2304.244399] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2304.244651] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Deleting contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2304.244848] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Deleting the datastore file [datastore1] 1805a6e7-48f2-47b4-b097-0644ba8217fa {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2304.245468] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fd78da55-41ee-450e-80b6-8d6a57a9cf1b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2304.252226] env[62824]: DEBUG oslo_vmware.api [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2304.252226] env[62824]: value = "task-2146789" [ 2304.252226] env[62824]: _type = "Task" [ 2304.252226] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2304.264440] env[62824]: DEBUG oslo_vmware.api [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146789, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2304.269596] env[62824]: DEBUG oslo_vmware.api [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146788, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.445311} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2304.269818] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 3d294ee9-92f6-4115-95db-9d4b66562b56/3d294ee9-92f6-4115-95db-9d4b66562b56.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2304.270103] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2304.270365] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2e81424a-f599-444b-8755-fc2c79108510 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2304.276330] env[62824]: DEBUG oslo_vmware.api [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2304.276330] env[62824]: value = "task-2146790" [ 2304.276330] env[62824]: _type = "Task" [ 2304.276330] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2304.284861] env[62824]: DEBUG oslo_vmware.api [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146790, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2304.423172] env[62824]: DEBUG nova.scheduler.client.report [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2304.693883] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Creating Snapshot of the VM instance {{(pid=62824) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 2304.694291] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-47447ff2-e04a-40e6-b51e-dbbc26486990 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2304.702405] env[62824]: DEBUG oslo_vmware.api [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Waiting for the task: (returnval){ [ 2304.702405] env[62824]: value = "task-2146791" [ 2304.702405] env[62824]: _type = "Task" [ 2304.702405] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2304.710932] env[62824]: DEBUG oslo_vmware.api [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146791, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2304.757639] env[62824]: DEBUG oslo_concurrency.lockutils [None req-bf852af9-823d-4ea1-8526-8b1688e7a098 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Lock "bbfcb2e3-9326-4548-b15b-e054cbfd192e" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.804s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2304.765863] env[62824]: DEBUG oslo_vmware.api [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146789, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.148063} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2304.765948] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2304.767527] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Deleted contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2304.767527] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2304.789377] env[62824]: DEBUG oslo_vmware.api [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146790, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07527} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2304.789734] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2304.790709] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05e5cfdc-0cb4-4c69-80a8-1dd9b7eb7cd6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2304.821375] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Reconfiguring VM instance instance-00000066 to attach disk [datastore2] 3d294ee9-92f6-4115-95db-9d4b66562b56/3d294ee9-92f6-4115-95db-9d4b66562b56.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2304.822095] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b3517109-c02d-494d-9b1f-dee8fae78b0a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2304.850206] env[62824]: DEBUG oslo_vmware.api [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2304.850206] env[62824]: value = "task-2146792" [ 2304.850206] env[62824]: _type = "Task" [ 2304.850206] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2304.858208] env[62824]: DEBUG oslo_vmware.api [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146792, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2304.931589] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.323s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2304.932110] env[62824]: DEBUG nova.compute.manager [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2304.934714] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d9f6b187-61c9-454d-8fbe-17f8e63decd3 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.334s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2304.934870] env[62824]: DEBUG nova.objects.instance [None req-d9f6b187-61c9-454d-8fbe-17f8e63decd3 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lazy-loading 'resources' on Instance uuid 19bcead9-b27c-471a-b217-cf7bfe101ed8 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2305.081970] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquiring lock "1db9ff97-e59c-43f6-984e-6405d8e993a4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2305.082308] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lock "1db9ff97-e59c-43f6-984e-6405d8e993a4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2305.215666] env[62824]: DEBUG oslo_vmware.api [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146791, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2305.313523] env[62824]: DEBUG nova.compute.manager [req-ebaa183f-e97c-432c-b898-5755fae30120 req-e826b108-b942-4d27-ad69-3813372b12b1 service nova] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Received event network-changed-d5c6ba87-e0c7-49c2-bbaf-b7c64f84cdf8 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2305.313523] env[62824]: DEBUG nova.compute.manager [req-ebaa183f-e97c-432c-b898-5755fae30120 req-e826b108-b942-4d27-ad69-3813372b12b1 service nova] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Refreshing instance network info cache due to event network-changed-d5c6ba87-e0c7-49c2-bbaf-b7c64f84cdf8. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2305.313523] env[62824]: DEBUG oslo_concurrency.lockutils [req-ebaa183f-e97c-432c-b898-5755fae30120 req-e826b108-b942-4d27-ad69-3813372b12b1 service nova] Acquiring lock "refresh_cache-e1574266-e53b-4169-9db6-c66b84895edb" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2305.313884] env[62824]: DEBUG oslo_concurrency.lockutils [req-ebaa183f-e97c-432c-b898-5755fae30120 req-e826b108-b942-4d27-ad69-3813372b12b1 service nova] Acquired lock "refresh_cache-e1574266-e53b-4169-9db6-c66b84895edb" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2305.313940] env[62824]: DEBUG nova.network.neutron [req-ebaa183f-e97c-432c-b898-5755fae30120 req-e826b108-b942-4d27-ad69-3813372b12b1 service nova] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Refreshing network info cache for port d5c6ba87-e0c7-49c2-bbaf-b7c64f84cdf8 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2305.348454] env[62824]: DEBUG oslo_concurrency.lockutils [None req-96907f0c-6087-4a41-a36e-34211547ef1e tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Acquiring lock "bbfcb2e3-9326-4548-b15b-e054cbfd192e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2305.348770] env[62824]: DEBUG oslo_concurrency.lockutils [None req-96907f0c-6087-4a41-a36e-34211547ef1e tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Lock "bbfcb2e3-9326-4548-b15b-e054cbfd192e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2305.349023] env[62824]: DEBUG oslo_concurrency.lockutils [None req-96907f0c-6087-4a41-a36e-34211547ef1e tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Acquiring lock "bbfcb2e3-9326-4548-b15b-e054cbfd192e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2305.349230] env[62824]: DEBUG oslo_concurrency.lockutils [None req-96907f0c-6087-4a41-a36e-34211547ef1e tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Lock "bbfcb2e3-9326-4548-b15b-e054cbfd192e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2305.349407] env[62824]: DEBUG oslo_concurrency.lockutils [None req-96907f0c-6087-4a41-a36e-34211547ef1e tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Lock "bbfcb2e3-9326-4548-b15b-e054cbfd192e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2305.351879] env[62824]: INFO nova.compute.manager [None req-96907f0c-6087-4a41-a36e-34211547ef1e tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Terminating instance [ 2305.364049] env[62824]: DEBUG oslo_vmware.api [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146792, 'name': ReconfigVM_Task, 'duration_secs': 0.279156} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2305.364049] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Reconfigured VM instance instance-00000066 to attach disk [datastore2] 3d294ee9-92f6-4115-95db-9d4b66562b56/3d294ee9-92f6-4115-95db-9d4b66562b56.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2305.365639] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-19b770b7-a390-4e5d-940a-6a0c25629656 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2305.372529] env[62824]: DEBUG oslo_vmware.api [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2305.372529] env[62824]: value = "task-2146793" [ 2305.372529] env[62824]: _type = "Task" [ 2305.372529] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2305.387346] env[62824]: DEBUG oslo_vmware.api [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146793, 'name': Rename_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2305.438434] env[62824]: DEBUG nova.compute.utils [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2305.447226] env[62824]: DEBUG nova.compute.manager [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2305.447226] env[62824]: DEBUG nova.network.neutron [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2305.502545] env[62824]: DEBUG nova.policy [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '959d127a7d144b33a0cae94db5c11846', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dfff08982dad4790bf4d555e2b4db5e4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2305.585616] env[62824]: DEBUG nova.compute.manager [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2305.704859] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-faf45f3e-16f1-46b2-9d1a-a3677e548245 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2305.719737] env[62824]: DEBUG oslo_vmware.api [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146791, 'name': CreateSnapshot_Task, 'duration_secs': 0.930284} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2305.720310] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Created Snapshot of the VM instance {{(pid=62824) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 2305.721388] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6797429e-b894-42b9-8fb0-4c8aa9f0df1d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2305.725319] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ae1ec03-fc3d-4af1-b261-a8b6e621794f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2305.768614] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bfec055-2fe9-4766-9a2e-891b9de3c982 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2305.782382] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a68db8a9-ca14-4c24-8e39-ac0061b019c2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2305.801024] env[62824]: DEBUG nova.compute.provider_tree [None req-d9f6b187-61c9-454d-8fbe-17f8e63decd3 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2305.831624] env[62824]: DEBUG nova.virt.hardware [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2305.831879] env[62824]: DEBUG nova.virt.hardware [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2305.832048] env[62824]: DEBUG nova.virt.hardware [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2305.832253] env[62824]: DEBUG nova.virt.hardware [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2305.832413] env[62824]: DEBUG nova.virt.hardware [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2305.832563] env[62824]: DEBUG nova.virt.hardware [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2305.832786] env[62824]: DEBUG nova.virt.hardware [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2305.832952] env[62824]: DEBUG nova.virt.hardware [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2305.833161] env[62824]: DEBUG nova.virt.hardware [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2305.833331] env[62824]: DEBUG nova.virt.hardware [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2305.833582] env[62824]: DEBUG nova.virt.hardware [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2305.835255] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94c91619-131f-4b73-b67a-24393b5d3e71 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2305.847352] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5c6c646-7d43-4092-9d57-8e18d8eb0a59 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2305.863684] env[62824]: DEBUG nova.compute.manager [None req-96907f0c-6087-4a41-a36e-34211547ef1e tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2305.863996] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-96907f0c-6087-4a41-a36e-34211547ef1e tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2305.864608] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e9:a5:9f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'de7fa486-5f28-44ae-b0cf-72234ff87546', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '37654518-0a1a-44e2-8297-2169c44d0aad', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2305.872591] env[62824]: DEBUG oslo.service.loopingcall [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2305.874929] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-016fd74e-f481-4d83-bfce-0ed0a7e9c621 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2305.878390] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2305.878742] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-eb3d88be-5951-4e78-8684-c4d7c1c161d3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2305.902793] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-96907f0c-6087-4a41-a36e-34211547ef1e tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2305.906067] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0f400dc5-e854-4230-85f8-edcf2c6473ac {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2305.907736] env[62824]: DEBUG oslo_vmware.api [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146793, 'name': Rename_Task, 'duration_secs': 0.153286} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2305.908964] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2305.909270] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2305.909270] env[62824]: value = "task-2146794" [ 2305.909270] env[62824]: _type = "Task" [ 2305.909270] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2305.909696] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c8b85faf-564c-441b-8e61-3d49620ab662 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2305.915402] env[62824]: DEBUG oslo_vmware.api [None req-96907f0c-6087-4a41-a36e-34211547ef1e tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for the task: (returnval){ [ 2305.915402] env[62824]: value = "task-2146795" [ 2305.915402] env[62824]: _type = "Task" [ 2305.915402] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2305.926596] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146794, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2305.937313] env[62824]: DEBUG oslo_vmware.api [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2305.937313] env[62824]: value = "task-2146796" [ 2305.937313] env[62824]: _type = "Task" [ 2305.937313] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2305.937313] env[62824]: DEBUG oslo_vmware.api [None req-96907f0c-6087-4a41-a36e-34211547ef1e tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2146795, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2305.943094] env[62824]: DEBUG nova.compute.manager [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2305.949459] env[62824]: DEBUG oslo_vmware.api [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146796, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2306.118100] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2306.140887] env[62824]: DEBUG nova.network.neutron [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3] Successfully created port: 2f4a7326-1463-4942-b137-573646230e57 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2306.281227] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Creating linked-clone VM from snapshot {{(pid=62824) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 2306.281717] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-a4b83fa3-5393-49a9-8480-84c99f91e18a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2306.290287] env[62824]: DEBUG oslo_vmware.api [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Waiting for the task: (returnval){ [ 2306.290287] env[62824]: value = "task-2146797" [ 2306.290287] env[62824]: _type = "Task" [ 2306.290287] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2306.303302] env[62824]: DEBUG nova.scheduler.client.report [None req-d9f6b187-61c9-454d-8fbe-17f8e63decd3 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2306.309161] env[62824]: DEBUG oslo_vmware.api [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146797, 'name': CloneVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2306.330464] env[62824]: DEBUG oslo_concurrency.lockutils [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquiring lock "511ef5bb-1dca-4eae-a0f1-19d1d7a17371" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2306.330790] env[62824]: DEBUG oslo_concurrency.lockutils [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lock "511ef5bb-1dca-4eae-a0f1-19d1d7a17371" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2306.409012] env[62824]: DEBUG nova.network.neutron [req-ebaa183f-e97c-432c-b898-5755fae30120 req-e826b108-b942-4d27-ad69-3813372b12b1 service nova] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Updated VIF entry in instance network info cache for port d5c6ba87-e0c7-49c2-bbaf-b7c64f84cdf8. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2306.409418] env[62824]: DEBUG nova.network.neutron [req-ebaa183f-e97c-432c-b898-5755fae30120 req-e826b108-b942-4d27-ad69-3813372b12b1 service nova] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Updating instance_info_cache with network_info: [{"id": "d5c6ba87-e0c7-49c2-bbaf-b7c64f84cdf8", "address": "fa:16:3e:c4:9c:f4", "network": {"id": "aa1c09e5-8d72-43ad-a903-a0f5e711da80", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1643616572-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.166", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c987c5f844be42c99cc9d57a9ca84c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9bc2632-36f9-4912-8782-8bbb789f909d", "external-id": "nsx-vlan-transportzone-897", "segmentation_id": 897, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd5c6ba87-e0", "ovs_interfaceid": "d5c6ba87-e0c7-49c2-bbaf-b7c64f84cdf8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2306.423420] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146794, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2306.428520] env[62824]: DEBUG oslo_vmware.api [None req-96907f0c-6087-4a41-a36e-34211547ef1e tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2146795, 'name': PowerOffVM_Task, 'duration_secs': 0.203658} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2306.429579] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-96907f0c-6087-4a41-a36e-34211547ef1e tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2306.429845] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-96907f0c-6087-4a41-a36e-34211547ef1e tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2306.430259] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f4633f01-b26d-440b-8c61-86815a37730c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2306.448672] env[62824]: DEBUG oslo_vmware.api [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146796, 'name': PowerOnVM_Task, 'duration_secs': 0.465824} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2306.449009] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2306.449392] env[62824]: INFO nova.compute.manager [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Took 7.76 seconds to spawn the instance on the hypervisor. [ 2306.449672] env[62824]: DEBUG nova.compute.manager [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2306.453597] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9d16c8f-36d5-4e27-a1fb-4de35e362ae5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2306.799798] env[62824]: DEBUG oslo_vmware.api [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146797, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2306.810858] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d9f6b187-61c9-454d-8fbe-17f8e63decd3 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.876s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2306.813404] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.696s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2306.814930] env[62824]: INFO nova.compute.claims [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2306.833829] env[62824]: DEBUG nova.compute.manager [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2306.857870] env[62824]: INFO nova.scheduler.client.report [None req-d9f6b187-61c9-454d-8fbe-17f8e63decd3 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Deleted allocations for instance 19bcead9-b27c-471a-b217-cf7bfe101ed8 [ 2306.912573] env[62824]: DEBUG oslo_concurrency.lockutils [req-ebaa183f-e97c-432c-b898-5755fae30120 req-e826b108-b942-4d27-ad69-3813372b12b1 service nova] Releasing lock "refresh_cache-e1574266-e53b-4169-9db6-c66b84895edb" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2306.922378] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146794, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2306.958754] env[62824]: DEBUG nova.compute.manager [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2306.971266] env[62824]: INFO nova.compute.manager [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Took 17.46 seconds to build instance. [ 2306.991727] env[62824]: DEBUG nova.virt.hardware [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2306.992011] env[62824]: DEBUG nova.virt.hardware [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2306.992184] env[62824]: DEBUG nova.virt.hardware [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2306.992366] env[62824]: DEBUG nova.virt.hardware [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2306.992510] env[62824]: DEBUG nova.virt.hardware [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2306.992654] env[62824]: DEBUG nova.virt.hardware [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2306.992859] env[62824]: DEBUG nova.virt.hardware [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2306.993024] env[62824]: DEBUG nova.virt.hardware [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2306.993194] env[62824]: DEBUG nova.virt.hardware [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2306.993353] env[62824]: DEBUG nova.virt.hardware [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2306.993695] env[62824]: DEBUG nova.virt.hardware [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2306.994621] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8248acf3-d450-44fd-839d-a46f91b2de90 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2307.002343] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d97258c6-ee1e-4080-963f-0389e5bce86c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2307.300257] env[62824]: DEBUG oslo_vmware.api [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146797, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2307.359910] env[62824]: DEBUG oslo_concurrency.lockutils [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2307.369328] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d9f6b187-61c9-454d-8fbe-17f8e63decd3 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "19bcead9-b27c-471a-b217-cf7bfe101ed8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.098s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2307.422618] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146794, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2307.473857] env[62824]: DEBUG oslo_concurrency.lockutils [None req-98d12844-cc61-4c3f-89a6-71a59f472796 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "3d294ee9-92f6-4115-95db-9d4b66562b56" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.969s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2307.652658] env[62824]: DEBUG nova.compute.manager [req-359ea165-1761-47a0-a109-b6a6b9039b55 req-58fb749c-12b8-49b8-bbf5-f1a72496251f service nova] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Received event network-changed-f1800205-8a92-425e-9f64-1dabd5809221 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2307.653773] env[62824]: DEBUG nova.compute.manager [req-359ea165-1761-47a0-a109-b6a6b9039b55 req-58fb749c-12b8-49b8-bbf5-f1a72496251f service nova] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Refreshing instance network info cache due to event network-changed-f1800205-8a92-425e-9f64-1dabd5809221. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2307.653773] env[62824]: DEBUG oslo_concurrency.lockutils [req-359ea165-1761-47a0-a109-b6a6b9039b55 req-58fb749c-12b8-49b8-bbf5-f1a72496251f service nova] Acquiring lock "refresh_cache-3d294ee9-92f6-4115-95db-9d4b66562b56" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2307.653773] env[62824]: DEBUG oslo_concurrency.lockutils [req-359ea165-1761-47a0-a109-b6a6b9039b55 req-58fb749c-12b8-49b8-bbf5-f1a72496251f service nova] Acquired lock "refresh_cache-3d294ee9-92f6-4115-95db-9d4b66562b56" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2307.653773] env[62824]: DEBUG nova.network.neutron [req-359ea165-1761-47a0-a109-b6a6b9039b55 req-58fb749c-12b8-49b8-bbf5-f1a72496251f service nova] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Refreshing network info cache for port f1800205-8a92-425e-9f64-1dabd5809221 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2307.801875] env[62824]: DEBUG oslo_vmware.api [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146797, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2307.932302] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146794, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2308.045982] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e64d10b-3933-45b3-8588-faec5b693151 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2308.060040] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdbf3a8e-1a4e-4c27-b852-33525c7d1d4d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2308.089192] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-694a1d17-fd37-4914-a977-e27d9307cf8c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2308.097155] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05dc3d44-ed8f-49b4-bbea-8fff16e3b8ba {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2308.110921] env[62824]: DEBUG nova.compute.provider_tree [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2308.301965] env[62824]: DEBUG oslo_vmware.api [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146797, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2308.430049] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146794, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2308.613542] env[62824]: DEBUG nova.scheduler.client.report [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2308.745582] env[62824]: DEBUG nova.network.neutron [req-359ea165-1761-47a0-a109-b6a6b9039b55 req-58fb749c-12b8-49b8-bbf5-f1a72496251f service nova] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Updated VIF entry in instance network info cache for port f1800205-8a92-425e-9f64-1dabd5809221. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2308.745977] env[62824]: DEBUG nova.network.neutron [req-359ea165-1761-47a0-a109-b6a6b9039b55 req-58fb749c-12b8-49b8-bbf5-f1a72496251f service nova] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Updating instance_info_cache with network_info: [{"id": "f1800205-8a92-425e-9f64-1dabd5809221", "address": "fa:16:3e:17:9c:53", "network": {"id": "b4ed285c-26a7-461b-8a71-d18e07582ab8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-997150974-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.243", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "815ecf9340934733a83d4dcb26612fd4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69f65356-c85e-4b7f-ad28-7c7b5e8cf50c", "external-id": "nsx-vlan-transportzone-281", "segmentation_id": 281, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf1800205-8a", "ovs_interfaceid": "f1800205-8a92-425e-9f64-1dabd5809221", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2308.802543] env[62824]: DEBUG oslo_vmware.api [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146797, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2308.924877] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146794, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2309.118840] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.305s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2309.120239] env[62824]: DEBUG nova.compute.manager [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2309.122604] env[62824]: DEBUG oslo_concurrency.lockutils [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.763s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2309.124283] env[62824]: INFO nova.compute.claims [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2309.174951] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquiring lock "c480f680-61fd-496b-8cf7-fa50f580b10d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2309.175373] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "c480f680-61fd-496b-8cf7-fa50f580b10d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2309.252920] env[62824]: DEBUG oslo_concurrency.lockutils [req-359ea165-1761-47a0-a109-b6a6b9039b55 req-58fb749c-12b8-49b8-bbf5-f1a72496251f service nova] Releasing lock "refresh_cache-3d294ee9-92f6-4115-95db-9d4b66562b56" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2309.303278] env[62824]: DEBUG oslo_vmware.api [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146797, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2309.425209] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146794, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2309.630020] env[62824]: DEBUG nova.compute.utils [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2309.633819] env[62824]: DEBUG nova.compute.manager [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2309.634019] env[62824]: DEBUG nova.network.neutron [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2309.678701] env[62824]: DEBUG nova.compute.manager [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2309.691376] env[62824]: DEBUG nova.policy [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5a5891a75fee4caf92d4bba152254ad5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7fcdbfc29df64a1ba8d982bdcc667b64', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2309.807396] env[62824]: DEBUG oslo_vmware.api [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146797, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2309.926364] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146794, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2310.058653] env[62824]: DEBUG nova.network.neutron [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Successfully created port: aeeadfdd-d869-4555-bda2-391c435de1bb {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2310.140513] env[62824]: DEBUG nova.compute.manager [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2310.199243] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2310.311055] env[62824]: DEBUG oslo_vmware.api [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146797, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2310.362618] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfb4d471-a733-41d6-9099-51b1243e1ec7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2310.370715] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa56da60-7ebe-4e4b-b4cd-9304931cbe84 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2310.403155] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56ba5ad0-8367-4401-adf3-1633bd02b4e2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2310.410943] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6856e90-4d1b-49d6-bc09-015501c94b5d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2310.424724] env[62824]: DEBUG nova.compute.provider_tree [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2310.441571] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146794, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2310.808116] env[62824]: DEBUG oslo_vmware.api [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146797, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2310.934607] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146794, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2310.937033] env[62824]: DEBUG nova.scheduler.client.report [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2311.154247] env[62824]: DEBUG nova.compute.manager [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2311.181831] env[62824]: DEBUG nova.virt.hardware [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2311.182111] env[62824]: DEBUG nova.virt.hardware [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2311.182292] env[62824]: DEBUG nova.virt.hardware [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2311.182465] env[62824]: DEBUG nova.virt.hardware [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2311.182611] env[62824]: DEBUG nova.virt.hardware [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2311.182755] env[62824]: DEBUG nova.virt.hardware [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2311.183089] env[62824]: DEBUG nova.virt.hardware [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2311.183308] env[62824]: DEBUG nova.virt.hardware [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2311.183471] env[62824]: DEBUG nova.virt.hardware [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2311.183636] env[62824]: DEBUG nova.virt.hardware [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2311.183805] env[62824]: DEBUG nova.virt.hardware [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2311.184710] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-607f5658-163a-4a9c-ba6b-fc42fc7a9b57 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2311.193275] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c66d4807-3f75-4477-9709-1e1e1f341600 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2311.309737] env[62824]: DEBUG oslo_vmware.api [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146797, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2311.435384] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146794, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2311.441313] env[62824]: DEBUG oslo_concurrency.lockutils [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.319s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2311.441850] env[62824]: DEBUG nova.compute.manager [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2311.444563] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.246s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2311.445953] env[62824]: INFO nova.compute.claims [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2311.809677] env[62824]: DEBUG oslo_vmware.api [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146797, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2311.936383] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146794, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2311.950051] env[62824]: DEBUG nova.compute.utils [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2311.953313] env[62824]: DEBUG nova.compute.manager [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2311.953476] env[62824]: DEBUG nova.network.neutron [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2312.015484] env[62824]: DEBUG nova.policy [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b989eb89498b45dab6604116870d2fa1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c987c5f844be42c99cc9d57a9ca84c31', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2312.311063] env[62824]: DEBUG oslo_vmware.api [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146797, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2312.379185] env[62824]: DEBUG nova.network.neutron [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Successfully created port: 3c1f3341-20a4-4b77-8dae-12ef5901a16e {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2312.437018] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146794, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2312.453746] env[62824]: DEBUG nova.compute.manager [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2312.700890] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af6673aa-11ee-41c6-8440-fbed6f00fee2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2312.709130] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b516f5e0-21dc-49f2-9133-9607f3a1d05e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2312.740197] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9e5d596-66fa-4553-b6ac-32efff68772e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2312.751187] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-408f6df4-c2b6-4929-87cd-3cf9b0e506a9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2312.765115] env[62824]: DEBUG nova.compute.provider_tree [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2312.810300] env[62824]: DEBUG oslo_vmware.api [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146797, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2312.937929] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146794, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2313.268678] env[62824]: DEBUG nova.scheduler.client.report [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2313.311472] env[62824]: DEBUG oslo_vmware.api [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146797, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2313.438817] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146794, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2313.466358] env[62824]: DEBUG nova.compute.manager [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2313.490125] env[62824]: DEBUG nova.virt.hardware [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2313.490385] env[62824]: DEBUG nova.virt.hardware [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2313.490545] env[62824]: DEBUG nova.virt.hardware [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2313.490726] env[62824]: DEBUG nova.virt.hardware [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2313.490874] env[62824]: DEBUG nova.virt.hardware [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2313.491037] env[62824]: DEBUG nova.virt.hardware [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2313.491252] env[62824]: DEBUG nova.virt.hardware [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2313.491411] env[62824]: DEBUG nova.virt.hardware [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2313.491578] env[62824]: DEBUG nova.virt.hardware [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2313.491741] env[62824]: DEBUG nova.virt.hardware [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2313.491912] env[62824]: DEBUG nova.virt.hardware [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2313.492779] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6d32cf3-0d00-42df-9738-7cc052ea7302 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2313.501222] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10d87bc4-6716-448c-8baa-c985c0872ce7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2313.744083] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-96907f0c-6087-4a41-a36e-34211547ef1e tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2313.744409] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-96907f0c-6087-4a41-a36e-34211547ef1e tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Deleting contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2313.744738] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-96907f0c-6087-4a41-a36e-34211547ef1e tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Deleting the datastore file [datastore2] bbfcb2e3-9326-4548-b15b-e054cbfd192e {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2313.745088] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0a9a2e87-c940-45b7-b3b1-785a9ca62fb7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2313.754291] env[62824]: DEBUG oslo_vmware.api [None req-96907f0c-6087-4a41-a36e-34211547ef1e tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for the task: (returnval){ [ 2313.754291] env[62824]: value = "task-2146799" [ 2313.754291] env[62824]: _type = "Task" [ 2313.754291] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2313.765358] env[62824]: DEBUG oslo_vmware.api [None req-96907f0c-6087-4a41-a36e-34211547ef1e tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2146799, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2313.774464] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.330s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2313.775048] env[62824]: DEBUG nova.compute.manager [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2313.811625] env[62824]: DEBUG oslo_vmware.api [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146797, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2313.940027] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146794, 'name': CreateVM_Task, 'duration_secs': 7.950822} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2313.940027] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2313.940225] env[62824]: DEBUG oslo_concurrency.lockutils [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2313.940259] env[62824]: DEBUG oslo_concurrency.lockutils [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2313.940571] env[62824]: DEBUG oslo_concurrency.lockutils [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2313.940822] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-15654be7-348f-48be-848e-1aeede696b56 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2313.946469] env[62824]: DEBUG oslo_vmware.api [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2313.946469] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52a7dfa2-ff1f-5c9d-de4a-00b50eb28778" [ 2313.946469] env[62824]: _type = "Task" [ 2313.946469] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2313.955155] env[62824]: DEBUG oslo_vmware.api [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52a7dfa2-ff1f-5c9d-de4a-00b50eb28778, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2314.264877] env[62824]: DEBUG oslo_vmware.api [None req-96907f0c-6087-4a41-a36e-34211547ef1e tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2146799, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.149738} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2314.265221] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-96907f0c-6087-4a41-a36e-34211547ef1e tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2314.265601] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-96907f0c-6087-4a41-a36e-34211547ef1e tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Deleted contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2314.265691] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-96907f0c-6087-4a41-a36e-34211547ef1e tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2314.265860] env[62824]: INFO nova.compute.manager [None req-96907f0c-6087-4a41-a36e-34211547ef1e tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Took 8.40 seconds to destroy the instance on the hypervisor. [ 2314.266113] env[62824]: DEBUG oslo.service.loopingcall [None req-96907f0c-6087-4a41-a36e-34211547ef1e tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2314.266369] env[62824]: DEBUG nova.compute.manager [-] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2314.266437] env[62824]: DEBUG nova.network.neutron [-] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2314.280881] env[62824]: DEBUG nova.compute.utils [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2314.282459] env[62824]: DEBUG nova.compute.manager [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2314.282652] env[62824]: DEBUG nova.network.neutron [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2314.312304] env[62824]: DEBUG oslo_vmware.api [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146797, 'name': CloneVM_Task} progress is 95%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2314.347097] env[62824]: DEBUG nova.policy [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c379c824a178444e99bb121716fc7887', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0c11aef7fb6247cb9b5272a6a063cd12', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2314.457682] env[62824]: DEBUG oslo_vmware.api [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52a7dfa2-ff1f-5c9d-de4a-00b50eb28778, 'name': SearchDatastore_Task, 'duration_secs': 0.010167} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2314.460168] env[62824]: DEBUG oslo_concurrency.lockutils [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2314.460168] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2314.460168] env[62824]: DEBUG oslo_concurrency.lockutils [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2314.460168] env[62824]: DEBUG oslo_concurrency.lockutils [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2314.460168] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2314.460168] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-79fc2048-e2ea-4a7e-a5c3-bd6f6cac7fdd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2314.469467] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2314.469567] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2314.470427] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-685e47f9-0fa2-45b3-aae5-2f564a24c8e4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2314.477106] env[62824]: DEBUG oslo_vmware.api [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2314.477106] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5289417b-3703-3079-b5e6-ef7843cec63a" [ 2314.477106] env[62824]: _type = "Task" [ 2314.477106] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2314.485416] env[62824]: DEBUG oslo_vmware.api [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5289417b-3703-3079-b5e6-ef7843cec63a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2314.785997] env[62824]: DEBUG nova.compute.manager [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2314.792727] env[62824]: DEBUG nova.network.neutron [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Successfully created port: 9190a574-76bc-4236-9f09-69d08897d634 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2314.813034] env[62824]: DEBUG oslo_vmware.api [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146797, 'name': CloneVM_Task, 'duration_secs': 8.267189} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2314.813312] env[62824]: INFO nova.virt.vmwareapi.vmops [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Created linked-clone VM from snapshot [ 2314.814084] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-915440df-2203-4158-987c-ee8274140602 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2314.822813] env[62824]: DEBUG nova.virt.vmwareapi.images [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Uploading image 91de84ba-a43c-4db9-92a7-d2e20dada664 {{(pid=62824) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 2314.848888] env[62824]: DEBUG oslo_vmware.rw_handles [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 2314.848888] env[62824]: value = "vm-438752" [ 2314.848888] env[62824]: _type = "VirtualMachine" [ 2314.848888] env[62824]: }. {{(pid=62824) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 2314.849184] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-6f1d93f1-2531-4078-856b-9f45a12705f5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2314.857457] env[62824]: DEBUG oslo_vmware.rw_handles [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Lease: (returnval){ [ 2314.857457] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5210105c-b097-42e3-ab34-4e318468a90e" [ 2314.857457] env[62824]: _type = "HttpNfcLease" [ 2314.857457] env[62824]: } obtained for exporting VM: (result){ [ 2314.857457] env[62824]: value = "vm-438752" [ 2314.857457] env[62824]: _type = "VirtualMachine" [ 2314.857457] env[62824]: }. {{(pid=62824) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 2314.857684] env[62824]: DEBUG oslo_vmware.api [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Waiting for the lease: (returnval){ [ 2314.857684] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5210105c-b097-42e3-ab34-4e318468a90e" [ 2314.857684] env[62824]: _type = "HttpNfcLease" [ 2314.857684] env[62824]: } to be ready. {{(pid=62824) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 2314.866049] env[62824]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 2314.866049] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5210105c-b097-42e3-ab34-4e318468a90e" [ 2314.866049] env[62824]: _type = "HttpNfcLease" [ 2314.866049] env[62824]: } is initializing. {{(pid=62824) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 2314.896174] env[62824]: DEBUG nova.compute.manager [req-77bb47a5-f09d-4477-bd86-0c4c9f6ad035 req-2cada8f0-8515-43a7-8b03-08e140b2d257 service nova] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Received event network-vif-plugged-aeeadfdd-d869-4555-bda2-391c435de1bb {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2314.896174] env[62824]: DEBUG oslo_concurrency.lockutils [req-77bb47a5-f09d-4477-bd86-0c4c9f6ad035 req-2cada8f0-8515-43a7-8b03-08e140b2d257 service nova] Acquiring lock "1db9ff97-e59c-43f6-984e-6405d8e993a4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2314.896402] env[62824]: DEBUG oslo_concurrency.lockutils [req-77bb47a5-f09d-4477-bd86-0c4c9f6ad035 req-2cada8f0-8515-43a7-8b03-08e140b2d257 service nova] Lock "1db9ff97-e59c-43f6-984e-6405d8e993a4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2314.896502] env[62824]: DEBUG oslo_concurrency.lockutils [req-77bb47a5-f09d-4477-bd86-0c4c9f6ad035 req-2cada8f0-8515-43a7-8b03-08e140b2d257 service nova] Lock "1db9ff97-e59c-43f6-984e-6405d8e993a4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2314.896673] env[62824]: DEBUG nova.compute.manager [req-77bb47a5-f09d-4477-bd86-0c4c9f6ad035 req-2cada8f0-8515-43a7-8b03-08e140b2d257 service nova] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] No waiting events found dispatching network-vif-plugged-aeeadfdd-d869-4555-bda2-391c435de1bb {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2314.896838] env[62824]: WARNING nova.compute.manager [req-77bb47a5-f09d-4477-bd86-0c4c9f6ad035 req-2cada8f0-8515-43a7-8b03-08e140b2d257 service nova] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Received unexpected event network-vif-plugged-aeeadfdd-d869-4555-bda2-391c435de1bb for instance with vm_state building and task_state spawning. [ 2314.993431] env[62824]: DEBUG oslo_vmware.api [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5289417b-3703-3079-b5e6-ef7843cec63a, 'name': SearchDatastore_Task, 'duration_secs': 0.010638} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2314.996214] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e99f9cdc-0871-4096-9a12-1a01f8a11b85 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2315.002555] env[62824]: DEBUG oslo_vmware.api [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2315.002555] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]525296ee-b2a5-eb24-a167-f4c313bd7000" [ 2315.002555] env[62824]: _type = "Task" [ 2315.002555] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2315.014429] env[62824]: DEBUG oslo_vmware.api [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]525296ee-b2a5-eb24-a167-f4c313bd7000, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2315.073116] env[62824]: DEBUG nova.compute.manager [req-2aa680af-51da-42e0-aec7-821022894964 req-0e7a9eb3-2d91-47de-a745-dd4432eecd48 service nova] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Received event network-vif-deleted-d85a0acc-fed6-4797-9f81-2aafa5bb6967 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2315.073301] env[62824]: INFO nova.compute.manager [req-2aa680af-51da-42e0-aec7-821022894964 req-0e7a9eb3-2d91-47de-a745-dd4432eecd48 service nova] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Neutron deleted interface d85a0acc-fed6-4797-9f81-2aafa5bb6967; detaching it from the instance and deleting it from the info cache [ 2315.073651] env[62824]: DEBUG nova.network.neutron [req-2aa680af-51da-42e0-aec7-821022894964 req-0e7a9eb3-2d91-47de-a745-dd4432eecd48 service nova] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2315.180721] env[62824]: DEBUG nova.network.neutron [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Successfully updated port: aeeadfdd-d869-4555-bda2-391c435de1bb {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2315.256610] env[62824]: DEBUG nova.network.neutron [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Successfully updated port: 3c1f3341-20a4-4b77-8dae-12ef5901a16e {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2315.335983] env[62824]: DEBUG nova.network.neutron [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3] Successfully updated port: 2f4a7326-1463-4942-b137-573646230e57 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2315.369823] env[62824]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 2315.369823] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5210105c-b097-42e3-ab34-4e318468a90e" [ 2315.369823] env[62824]: _type = "HttpNfcLease" [ 2315.369823] env[62824]: } is ready. {{(pid=62824) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 2315.370841] env[62824]: DEBUG oslo_vmware.rw_handles [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 2315.370841] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5210105c-b097-42e3-ab34-4e318468a90e" [ 2315.370841] env[62824]: _type = "HttpNfcLease" [ 2315.370841] env[62824]: }. {{(pid=62824) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 2315.371973] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcc5f8fc-194e-4173-8a52-efe7ba4c3b50 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2315.384188] env[62824]: DEBUG oslo_vmware.rw_handles [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52724aed-6614-2bc5-de89-88a16d737b6b/disk-0.vmdk from lease info. {{(pid=62824) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 2315.384510] env[62824]: DEBUG oslo_vmware.rw_handles [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52724aed-6614-2bc5-de89-88a16d737b6b/disk-0.vmdk for reading. {{(pid=62824) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 2315.486442] env[62824]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-dc124565-0f91-4f01-9e74-2d1032d8c01d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2315.518187] env[62824]: DEBUG oslo_vmware.api [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]525296ee-b2a5-eb24-a167-f4c313bd7000, 'name': SearchDatastore_Task, 'duration_secs': 0.011249} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2315.521718] env[62824]: DEBUG oslo_concurrency.lockutils [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2315.522035] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 1805a6e7-48f2-47b4-b097-0644ba8217fa/1805a6e7-48f2-47b4-b097-0644ba8217fa.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2315.522721] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6985b2c9-a117-4370-9874-741d220a543a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2315.525045] env[62824]: DEBUG nova.network.neutron [-] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2315.534044] env[62824]: DEBUG oslo_vmware.api [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2315.534044] env[62824]: value = "task-2146801" [ 2315.534044] env[62824]: _type = "Task" [ 2315.534044] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2315.543897] env[62824]: DEBUG oslo_vmware.api [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146801, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2315.577017] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e928d198-1875-4538-9203-88dbe945f1f3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2315.587176] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2632dc77-b8a0-4931-b1a7-9ace3a8e62cf {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2315.624492] env[62824]: DEBUG nova.compute.manager [req-2aa680af-51da-42e0-aec7-821022894964 req-0e7a9eb3-2d91-47de-a745-dd4432eecd48 service nova] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Detach interface failed, port_id=d85a0acc-fed6-4797-9f81-2aafa5bb6967, reason: Instance bbfcb2e3-9326-4548-b15b-e054cbfd192e could not be found. {{(pid=62824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11482}} [ 2315.687177] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquiring lock "refresh_cache-1db9ff97-e59c-43f6-984e-6405d8e993a4" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2315.687346] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquired lock "refresh_cache-1db9ff97-e59c-43f6-984e-6405d8e993a4" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2315.687623] env[62824]: DEBUG nova.network.neutron [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2315.759613] env[62824]: DEBUG oslo_concurrency.lockutils [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquiring lock "refresh_cache-511ef5bb-1dca-4eae-a0f1-19d1d7a17371" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2315.759782] env[62824]: DEBUG oslo_concurrency.lockutils [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquired lock "refresh_cache-511ef5bb-1dca-4eae-a0f1-19d1d7a17371" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2315.759917] env[62824]: DEBUG nova.network.neutron [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2315.795817] env[62824]: DEBUG nova.compute.manager [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2315.822468] env[62824]: DEBUG nova.virt.hardware [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2315.822748] env[62824]: DEBUG nova.virt.hardware [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2315.822906] env[62824]: DEBUG nova.virt.hardware [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2315.823103] env[62824]: DEBUG nova.virt.hardware [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2315.823304] env[62824]: DEBUG nova.virt.hardware [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2315.823484] env[62824]: DEBUG nova.virt.hardware [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2315.823764] env[62824]: DEBUG nova.virt.hardware [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2315.824098] env[62824]: DEBUG nova.virt.hardware [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2315.824412] env[62824]: DEBUG nova.virt.hardware [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2315.824589] env[62824]: DEBUG nova.virt.hardware [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2315.824865] env[62824]: DEBUG nova.virt.hardware [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2315.825938] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e07452e-b133-43d4-b748-439c919323f7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2315.838543] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbe9766f-c3b1-4f06-b262-09f5650f84c8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2315.843249] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquiring lock "refresh_cache-68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2315.843408] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquired lock "refresh_cache-68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2315.843601] env[62824]: DEBUG nova.network.neutron [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2316.027717] env[62824]: INFO nova.compute.manager [-] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Took 1.76 seconds to deallocate network for instance. [ 2316.046328] env[62824]: DEBUG oslo_vmware.api [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146801, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.497221} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2316.046787] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 1805a6e7-48f2-47b4-b097-0644ba8217fa/1805a6e7-48f2-47b4-b097-0644ba8217fa.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2316.047418] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2316.047567] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bcaddc90-a0cd-4480-8492-d83f991c130c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2316.056844] env[62824]: DEBUG oslo_vmware.api [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2316.056844] env[62824]: value = "task-2146802" [ 2316.056844] env[62824]: _type = "Task" [ 2316.056844] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2316.069466] env[62824]: DEBUG oslo_vmware.api [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146802, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2316.220794] env[62824]: DEBUG nova.network.neutron [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2316.318383] env[62824]: DEBUG nova.network.neutron [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2316.406152] env[62824]: DEBUG nova.network.neutron [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2316.429988] env[62824]: DEBUG nova.network.neutron [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Updating instance_info_cache with network_info: [{"id": "aeeadfdd-d869-4555-bda2-391c435de1bb", "address": "fa:16:3e:43:2e:10", "network": {"id": "bba9b98c-c7eb-4890-8de5-eafacfdebd7d", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1928394728-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7fcdbfc29df64a1ba8d982bdcc667b64", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b49c5024-2ced-42ca-90cc-6066766d43e6", "external-id": "nsx-vlan-transportzone-239", "segmentation_id": 239, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaeeadfdd-d8", "ovs_interfaceid": "aeeadfdd-d869-4555-bda2-391c435de1bb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2316.536151] env[62824]: DEBUG oslo_concurrency.lockutils [None req-96907f0c-6087-4a41-a36e-34211547ef1e tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2316.536432] env[62824]: DEBUG oslo_concurrency.lockutils [None req-96907f0c-6087-4a41-a36e-34211547ef1e tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2316.537377] env[62824]: DEBUG nova.objects.instance [None req-96907f0c-6087-4a41-a36e-34211547ef1e tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Lazy-loading 'resources' on Instance uuid bbfcb2e3-9326-4548-b15b-e054cbfd192e {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2316.569884] env[62824]: DEBUG oslo_vmware.api [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146802, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079035} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2316.570297] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2316.571330] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12a21c57-9390-456f-bc0a-10b4663b4062 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2316.598656] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Reconfiguring VM instance instance-00000064 to attach disk [datastore1] 1805a6e7-48f2-47b4-b097-0644ba8217fa/1805a6e7-48f2-47b4-b097-0644ba8217fa.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2316.599137] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b6c57d16-bc37-412c-9c7b-48195e2034e9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2316.618240] env[62824]: DEBUG nova.network.neutron [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Successfully updated port: 9190a574-76bc-4236-9f09-69d08897d634 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2316.631582] env[62824]: DEBUG oslo_vmware.api [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2316.631582] env[62824]: value = "task-2146803" [ 2316.631582] env[62824]: _type = "Task" [ 2316.631582] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2316.643815] env[62824]: DEBUG oslo_vmware.api [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146803, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2316.672697] env[62824]: DEBUG nova.network.neutron [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Updating instance_info_cache with network_info: [{"id": "3c1f3341-20a4-4b77-8dae-12ef5901a16e", "address": "fa:16:3e:1c:3a:8c", "network": {"id": "aa1c09e5-8d72-43ad-a903-a0f5e711da80", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1643616572-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c987c5f844be42c99cc9d57a9ca84c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9bc2632-36f9-4912-8782-8bbb789f909d", "external-id": "nsx-vlan-transportzone-897", "segmentation_id": 897, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3c1f3341-20", "ovs_interfaceid": "3c1f3341-20a4-4b77-8dae-12ef5901a16e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2316.704618] env[62824]: DEBUG nova.network.neutron [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3] Updating instance_info_cache with network_info: [{"id": "2f4a7326-1463-4942-b137-573646230e57", "address": "fa:16:3e:a3:75:9b", "network": {"id": "0f512108-670f-4a85-9381-108e18b3b5ed", "bridge": "br-int", "label": "tempest-ServersTestJSON-766397782-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dfff08982dad4790bf4d555e2b4db5e4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f3a2eb5-353f-45c5-a73b-869626f4bb13", "external-id": "nsx-vlan-transportzone-411", "segmentation_id": 411, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2f4a7326-14", "ovs_interfaceid": "2f4a7326-1463-4942-b137-573646230e57", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2316.927861] env[62824]: DEBUG nova.compute.manager [req-2ada9a71-6ae0-4110-9e35-b3c3e4b88a39 req-eab6b440-b266-4c55-9b4e-98f6a93b145e service nova] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Received event network-vif-plugged-3c1f3341-20a4-4b77-8dae-12ef5901a16e {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2316.928320] env[62824]: DEBUG oslo_concurrency.lockutils [req-2ada9a71-6ae0-4110-9e35-b3c3e4b88a39 req-eab6b440-b266-4c55-9b4e-98f6a93b145e service nova] Acquiring lock "511ef5bb-1dca-4eae-a0f1-19d1d7a17371-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2316.928542] env[62824]: DEBUG oslo_concurrency.lockutils [req-2ada9a71-6ae0-4110-9e35-b3c3e4b88a39 req-eab6b440-b266-4c55-9b4e-98f6a93b145e service nova] Lock "511ef5bb-1dca-4eae-a0f1-19d1d7a17371-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2316.928714] env[62824]: DEBUG oslo_concurrency.lockutils [req-2ada9a71-6ae0-4110-9e35-b3c3e4b88a39 req-eab6b440-b266-4c55-9b4e-98f6a93b145e service nova] Lock "511ef5bb-1dca-4eae-a0f1-19d1d7a17371-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2316.928905] env[62824]: DEBUG nova.compute.manager [req-2ada9a71-6ae0-4110-9e35-b3c3e4b88a39 req-eab6b440-b266-4c55-9b4e-98f6a93b145e service nova] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] No waiting events found dispatching network-vif-plugged-3c1f3341-20a4-4b77-8dae-12ef5901a16e {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2316.929118] env[62824]: WARNING nova.compute.manager [req-2ada9a71-6ae0-4110-9e35-b3c3e4b88a39 req-eab6b440-b266-4c55-9b4e-98f6a93b145e service nova] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Received unexpected event network-vif-plugged-3c1f3341-20a4-4b77-8dae-12ef5901a16e for instance with vm_state building and task_state spawning. [ 2316.929435] env[62824]: DEBUG nova.compute.manager [req-2ada9a71-6ae0-4110-9e35-b3c3e4b88a39 req-eab6b440-b266-4c55-9b4e-98f6a93b145e service nova] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Received event network-changed-aeeadfdd-d869-4555-bda2-391c435de1bb {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2316.929673] env[62824]: DEBUG nova.compute.manager [req-2ada9a71-6ae0-4110-9e35-b3c3e4b88a39 req-eab6b440-b266-4c55-9b4e-98f6a93b145e service nova] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Refreshing instance network info cache due to event network-changed-aeeadfdd-d869-4555-bda2-391c435de1bb. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2316.929900] env[62824]: DEBUG oslo_concurrency.lockutils [req-2ada9a71-6ae0-4110-9e35-b3c3e4b88a39 req-eab6b440-b266-4c55-9b4e-98f6a93b145e service nova] Acquiring lock "refresh_cache-1db9ff97-e59c-43f6-984e-6405d8e993a4" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2316.932782] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Releasing lock "refresh_cache-1db9ff97-e59c-43f6-984e-6405d8e993a4" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2316.933093] env[62824]: DEBUG nova.compute.manager [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Instance network_info: |[{"id": "aeeadfdd-d869-4555-bda2-391c435de1bb", "address": "fa:16:3e:43:2e:10", "network": {"id": "bba9b98c-c7eb-4890-8de5-eafacfdebd7d", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1928394728-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7fcdbfc29df64a1ba8d982bdcc667b64", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b49c5024-2ced-42ca-90cc-6066766d43e6", "external-id": "nsx-vlan-transportzone-239", "segmentation_id": 239, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaeeadfdd-d8", "ovs_interfaceid": "aeeadfdd-d869-4555-bda2-391c435de1bb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2316.933635] env[62824]: DEBUG oslo_concurrency.lockutils [req-2ada9a71-6ae0-4110-9e35-b3c3e4b88a39 req-eab6b440-b266-4c55-9b4e-98f6a93b145e service nova] Acquired lock "refresh_cache-1db9ff97-e59c-43f6-984e-6405d8e993a4" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2316.933984] env[62824]: DEBUG nova.network.neutron [req-2ada9a71-6ae0-4110-9e35-b3c3e4b88a39 req-eab6b440-b266-4c55-9b4e-98f6a93b145e service nova] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Refreshing network info cache for port aeeadfdd-d869-4555-bda2-391c435de1bb {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2316.935460] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:43:2e:10', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b49c5024-2ced-42ca-90cc-6066766d43e6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'aeeadfdd-d869-4555-bda2-391c435de1bb', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2316.944752] env[62824]: DEBUG oslo.service.loopingcall [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2316.946143] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2316.946373] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ffc11c32-1bd8-47ce-b96b-c1e55559a6d7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2316.968270] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2316.968270] env[62824]: value = "task-2146804" [ 2316.968270] env[62824]: _type = "Task" [ 2316.968270] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2316.978298] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146804, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2317.105314] env[62824]: DEBUG nova.compute.manager [req-ff980577-74cd-4d01-a99b-de18ad8801ff req-c3aa2b95-4fc3-498b-b7b6-7bc2bbc669fb service nova] [instance: 68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3] Received event network-vif-plugged-2f4a7326-1463-4942-b137-573646230e57 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2317.105720] env[62824]: DEBUG oslo_concurrency.lockutils [req-ff980577-74cd-4d01-a99b-de18ad8801ff req-c3aa2b95-4fc3-498b-b7b6-7bc2bbc669fb service nova] Acquiring lock "68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2317.106181] env[62824]: DEBUG oslo_concurrency.lockutils [req-ff980577-74cd-4d01-a99b-de18ad8801ff req-c3aa2b95-4fc3-498b-b7b6-7bc2bbc669fb service nova] Lock "68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2317.106434] env[62824]: DEBUG oslo_concurrency.lockutils [req-ff980577-74cd-4d01-a99b-de18ad8801ff req-c3aa2b95-4fc3-498b-b7b6-7bc2bbc669fb service nova] Lock "68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2317.106712] env[62824]: DEBUG nova.compute.manager [req-ff980577-74cd-4d01-a99b-de18ad8801ff req-c3aa2b95-4fc3-498b-b7b6-7bc2bbc669fb service nova] [instance: 68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3] No waiting events found dispatching network-vif-plugged-2f4a7326-1463-4942-b137-573646230e57 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2317.107175] env[62824]: WARNING nova.compute.manager [req-ff980577-74cd-4d01-a99b-de18ad8801ff req-c3aa2b95-4fc3-498b-b7b6-7bc2bbc669fb service nova] [instance: 68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3] Received unexpected event network-vif-plugged-2f4a7326-1463-4942-b137-573646230e57 for instance with vm_state building and task_state spawning. [ 2317.107370] env[62824]: DEBUG nova.compute.manager [req-ff980577-74cd-4d01-a99b-de18ad8801ff req-c3aa2b95-4fc3-498b-b7b6-7bc2bbc669fb service nova] [instance: 68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3] Received event network-changed-2f4a7326-1463-4942-b137-573646230e57 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2317.107637] env[62824]: DEBUG nova.compute.manager [req-ff980577-74cd-4d01-a99b-de18ad8801ff req-c3aa2b95-4fc3-498b-b7b6-7bc2bbc669fb service nova] [instance: 68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3] Refreshing instance network info cache due to event network-changed-2f4a7326-1463-4942-b137-573646230e57. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2317.107865] env[62824]: DEBUG oslo_concurrency.lockutils [req-ff980577-74cd-4d01-a99b-de18ad8801ff req-c3aa2b95-4fc3-498b-b7b6-7bc2bbc669fb service nova] Acquiring lock "refresh_cache-68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2317.123636] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquiring lock "refresh_cache-c480f680-61fd-496b-8cf7-fa50f580b10d" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2317.123869] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquired lock "refresh_cache-c480f680-61fd-496b-8cf7-fa50f580b10d" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2317.124094] env[62824]: DEBUG nova.network.neutron [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2317.146731] env[62824]: DEBUG oslo_vmware.api [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146803, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2317.176405] env[62824]: DEBUG oslo_concurrency.lockutils [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Releasing lock "refresh_cache-511ef5bb-1dca-4eae-a0f1-19d1d7a17371" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2317.176732] env[62824]: DEBUG nova.compute.manager [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Instance network_info: |[{"id": "3c1f3341-20a4-4b77-8dae-12ef5901a16e", "address": "fa:16:3e:1c:3a:8c", "network": {"id": "aa1c09e5-8d72-43ad-a903-a0f5e711da80", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1643616572-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c987c5f844be42c99cc9d57a9ca84c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9bc2632-36f9-4912-8782-8bbb789f909d", "external-id": "nsx-vlan-transportzone-897", "segmentation_id": 897, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3c1f3341-20", "ovs_interfaceid": "3c1f3341-20a4-4b77-8dae-12ef5901a16e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2317.177690] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1c:3a:8c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c9bc2632-36f9-4912-8782-8bbb789f909d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3c1f3341-20a4-4b77-8dae-12ef5901a16e', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2317.185669] env[62824]: DEBUG oslo.service.loopingcall [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2317.186669] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2317.187098] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ed1f6c6d-91f7-4e81-a68b-052b38bf0c8f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2317.207635] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Releasing lock "refresh_cache-68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2317.208083] env[62824]: DEBUG nova.compute.manager [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3] Instance network_info: |[{"id": "2f4a7326-1463-4942-b137-573646230e57", "address": "fa:16:3e:a3:75:9b", "network": {"id": "0f512108-670f-4a85-9381-108e18b3b5ed", "bridge": "br-int", "label": "tempest-ServersTestJSON-766397782-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dfff08982dad4790bf4d555e2b4db5e4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f3a2eb5-353f-45c5-a73b-869626f4bb13", "external-id": "nsx-vlan-transportzone-411", "segmentation_id": 411, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2f4a7326-14", "ovs_interfaceid": "2f4a7326-1463-4942-b137-573646230e57", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2317.208375] env[62824]: DEBUG oslo_concurrency.lockutils [req-ff980577-74cd-4d01-a99b-de18ad8801ff req-c3aa2b95-4fc3-498b-b7b6-7bc2bbc669fb service nova] Acquired lock "refresh_cache-68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2317.208563] env[62824]: DEBUG nova.network.neutron [req-ff980577-74cd-4d01-a99b-de18ad8801ff req-c3aa2b95-4fc3-498b-b7b6-7bc2bbc669fb service nova] [instance: 68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3] Refreshing network info cache for port 2f4a7326-1463-4942-b137-573646230e57 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2317.209995] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a3:75:9b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9f3a2eb5-353f-45c5-a73b-869626f4bb13', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2f4a7326-1463-4942-b137-573646230e57', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2317.218392] env[62824]: DEBUG oslo.service.loopingcall [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2317.223526] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2317.224360] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2317.224360] env[62824]: value = "task-2146805" [ 2317.224360] env[62824]: _type = "Task" [ 2317.224360] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2317.224360] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9345663d-1cde-4622-8eeb-bb519a4ddaf4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2317.259112] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146805, 'name': CreateVM_Task} progress is 15%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2317.261217] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2317.261217] env[62824]: value = "task-2146806" [ 2317.261217] env[62824]: _type = "Task" [ 2317.261217] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2317.275956] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146806, 'name': CreateVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2317.330765] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb8dd3e4-7504-44d4-b751-877a1f8c8d5c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2317.339877] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1530b2c4-6434-45f9-a769-c4a21e25d91a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2317.379440] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d47141b-1c20-4e5e-827c-a5581336dd9a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2317.389407] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e54dba5-1d1c-4dc6-a60f-467f6722880b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2317.405480] env[62824]: DEBUG nova.compute.provider_tree [None req-96907f0c-6087-4a41-a36e-34211547ef1e tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2317.484303] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146804, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2317.644105] env[62824]: DEBUG oslo_vmware.api [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146803, 'name': ReconfigVM_Task, 'duration_secs': 0.700851} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2317.644503] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Reconfigured VM instance instance-00000064 to attach disk [datastore1] 1805a6e7-48f2-47b4-b097-0644ba8217fa/1805a6e7-48f2-47b4-b097-0644ba8217fa.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2317.645249] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-30e69176-74ab-45d5-b394-22edce93ce3e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2317.654753] env[62824]: DEBUG oslo_vmware.api [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2317.654753] env[62824]: value = "task-2146807" [ 2317.654753] env[62824]: _type = "Task" [ 2317.654753] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2317.666741] env[62824]: DEBUG oslo_vmware.api [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146807, 'name': Rename_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2317.679507] env[62824]: DEBUG nova.network.neutron [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2317.759663] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146805, 'name': CreateVM_Task, 'duration_secs': 0.532547} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2317.762424] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2317.763220] env[62824]: DEBUG oslo_concurrency.lockutils [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2317.763402] env[62824]: DEBUG oslo_concurrency.lockutils [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2317.763981] env[62824]: DEBUG oslo_concurrency.lockutils [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2317.767264] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9fae7414-40dd-4808-bb63-2da7ad4a64b9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2317.772867] env[62824]: DEBUG oslo_vmware.api [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Waiting for the task: (returnval){ [ 2317.772867] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]524bb9f3-2d8f-b4e4-7fe4-2df1b06b50e2" [ 2317.772867] env[62824]: _type = "Task" [ 2317.772867] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2317.777500] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146806, 'name': CreateVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2317.787430] env[62824]: DEBUG oslo_vmware.api [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]524bb9f3-2d8f-b4e4-7fe4-2df1b06b50e2, 'name': SearchDatastore_Task, 'duration_secs': 0.01165} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2317.787825] env[62824]: DEBUG oslo_concurrency.lockutils [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2317.788103] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2317.788366] env[62824]: DEBUG oslo_concurrency.lockutils [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2317.788619] env[62824]: DEBUG oslo_concurrency.lockutils [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2317.789138] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2317.789520] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b6374817-590e-4135-8163-0753c1a75653 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2317.800083] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2317.800337] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2317.801266] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3a3b6b5a-765a-446f-8019-d47548d26414 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2317.811826] env[62824]: DEBUG oslo_vmware.api [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Waiting for the task: (returnval){ [ 2317.811826] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52b5b8e0-b268-301a-fd66-f6eaaf3b779a" [ 2317.811826] env[62824]: _type = "Task" [ 2317.811826] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2317.821693] env[62824]: DEBUG oslo_vmware.api [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52b5b8e0-b268-301a-fd66-f6eaaf3b779a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2317.908931] env[62824]: DEBUG nova.scheduler.client.report [None req-96907f0c-6087-4a41-a36e-34211547ef1e tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2317.913726] env[62824]: DEBUG nova.network.neutron [req-2ada9a71-6ae0-4110-9e35-b3c3e4b88a39 req-eab6b440-b266-4c55-9b4e-98f6a93b145e service nova] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Updated VIF entry in instance network info cache for port aeeadfdd-d869-4555-bda2-391c435de1bb. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2317.913726] env[62824]: DEBUG nova.network.neutron [req-2ada9a71-6ae0-4110-9e35-b3c3e4b88a39 req-eab6b440-b266-4c55-9b4e-98f6a93b145e service nova] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Updating instance_info_cache with network_info: [{"id": "aeeadfdd-d869-4555-bda2-391c435de1bb", "address": "fa:16:3e:43:2e:10", "network": {"id": "bba9b98c-c7eb-4890-8de5-eafacfdebd7d", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1928394728-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7fcdbfc29df64a1ba8d982bdcc667b64", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b49c5024-2ced-42ca-90cc-6066766d43e6", "external-id": "nsx-vlan-transportzone-239", "segmentation_id": 239, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaeeadfdd-d8", "ovs_interfaceid": "aeeadfdd-d869-4555-bda2-391c435de1bb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2317.981462] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146804, 'name': CreateVM_Task, 'duration_secs': 0.623393} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2317.981889] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2317.982756] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2317.983167] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2317.983493] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2317.983764] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-80ed9b29-f0e2-4102-8b70-33a725433c6d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2317.990403] env[62824]: DEBUG oslo_vmware.api [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2317.990403] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52651a7d-909d-ba54-f611-4f98aabe1bd5" [ 2317.990403] env[62824]: _type = "Task" [ 2317.990403] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2318.003211] env[62824]: DEBUG oslo_vmware.api [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52651a7d-909d-ba54-f611-4f98aabe1bd5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2318.003211] env[62824]: DEBUG nova.network.neutron [req-ff980577-74cd-4d01-a99b-de18ad8801ff req-c3aa2b95-4fc3-498b-b7b6-7bc2bbc669fb service nova] [instance: 68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3] Updated VIF entry in instance network info cache for port 2f4a7326-1463-4942-b137-573646230e57. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2318.003600] env[62824]: DEBUG nova.network.neutron [req-ff980577-74cd-4d01-a99b-de18ad8801ff req-c3aa2b95-4fc3-498b-b7b6-7bc2bbc669fb service nova] [instance: 68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3] Updating instance_info_cache with network_info: [{"id": "2f4a7326-1463-4942-b137-573646230e57", "address": "fa:16:3e:a3:75:9b", "network": {"id": "0f512108-670f-4a85-9381-108e18b3b5ed", "bridge": "br-int", "label": "tempest-ServersTestJSON-766397782-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dfff08982dad4790bf4d555e2b4db5e4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f3a2eb5-353f-45c5-a73b-869626f4bb13", "external-id": "nsx-vlan-transportzone-411", "segmentation_id": 411, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2f4a7326-14", "ovs_interfaceid": "2f4a7326-1463-4942-b137-573646230e57", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2318.007713] env[62824]: DEBUG nova.network.neutron [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Updating instance_info_cache with network_info: [{"id": "9190a574-76bc-4236-9f09-69d08897d634", "address": "fa:16:3e:a9:57:8f", "network": {"id": "b995934c-4934-4469-884c-73cb4de1796f", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1892307918-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c11aef7fb6247cb9b5272a6a063cd12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15165046-2de9-4ada-9e99-0126e20854a9", "external-id": "nsx-vlan-transportzone-974", "segmentation_id": 974, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9190a574-76", "ovs_interfaceid": "9190a574-76bc-4236-9f09-69d08897d634", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2318.166568] env[62824]: DEBUG oslo_vmware.api [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146807, 'name': Rename_Task, 'duration_secs': 0.201288} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2318.166933] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2318.167206] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bf2b03b5-adec-44a4-9f2b-aa19b40e0973 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2318.175881] env[62824]: DEBUG oslo_vmware.api [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2318.175881] env[62824]: value = "task-2146808" [ 2318.175881] env[62824]: _type = "Task" [ 2318.175881] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2318.185026] env[62824]: DEBUG oslo_vmware.api [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146808, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2318.273757] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146806, 'name': CreateVM_Task, 'duration_secs': 0.559026} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2318.273757] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2318.274017] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2318.324415] env[62824]: DEBUG oslo_vmware.api [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52b5b8e0-b268-301a-fd66-f6eaaf3b779a, 'name': SearchDatastore_Task, 'duration_secs': 0.012798} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2318.325208] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7f32a564-dd62-4cf3-8249-64cb0ef751b5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2318.332250] env[62824]: DEBUG oslo_vmware.api [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Waiting for the task: (returnval){ [ 2318.332250] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52cad241-f451-5da6-b87f-6132a4946691" [ 2318.332250] env[62824]: _type = "Task" [ 2318.332250] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2318.340821] env[62824]: DEBUG oslo_vmware.api [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52cad241-f451-5da6-b87f-6132a4946691, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2318.416207] env[62824]: DEBUG oslo_concurrency.lockutils [None req-96907f0c-6087-4a41-a36e-34211547ef1e tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.880s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2318.419461] env[62824]: DEBUG oslo_concurrency.lockutils [req-2ada9a71-6ae0-4110-9e35-b3c3e4b88a39 req-eab6b440-b266-4c55-9b4e-98f6a93b145e service nova] Releasing lock "refresh_cache-1db9ff97-e59c-43f6-984e-6405d8e993a4" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2318.419819] env[62824]: DEBUG nova.compute.manager [req-2ada9a71-6ae0-4110-9e35-b3c3e4b88a39 req-eab6b440-b266-4c55-9b4e-98f6a93b145e service nova] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Received event network-changed-3c1f3341-20a4-4b77-8dae-12ef5901a16e {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2318.419977] env[62824]: DEBUG nova.compute.manager [req-2ada9a71-6ae0-4110-9e35-b3c3e4b88a39 req-eab6b440-b266-4c55-9b4e-98f6a93b145e service nova] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Refreshing instance network info cache due to event network-changed-3c1f3341-20a4-4b77-8dae-12ef5901a16e. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2318.420286] env[62824]: DEBUG oslo_concurrency.lockutils [req-2ada9a71-6ae0-4110-9e35-b3c3e4b88a39 req-eab6b440-b266-4c55-9b4e-98f6a93b145e service nova] Acquiring lock "refresh_cache-511ef5bb-1dca-4eae-a0f1-19d1d7a17371" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2318.420447] env[62824]: DEBUG oslo_concurrency.lockutils [req-2ada9a71-6ae0-4110-9e35-b3c3e4b88a39 req-eab6b440-b266-4c55-9b4e-98f6a93b145e service nova] Acquired lock "refresh_cache-511ef5bb-1dca-4eae-a0f1-19d1d7a17371" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2318.420617] env[62824]: DEBUG nova.network.neutron [req-2ada9a71-6ae0-4110-9e35-b3c3e4b88a39 req-eab6b440-b266-4c55-9b4e-98f6a93b145e service nova] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Refreshing network info cache for port 3c1f3341-20a4-4b77-8dae-12ef5901a16e {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2318.438666] env[62824]: INFO nova.scheduler.client.report [None req-96907f0c-6087-4a41-a36e-34211547ef1e tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Deleted allocations for instance bbfcb2e3-9326-4548-b15b-e054cbfd192e [ 2318.503538] env[62824]: DEBUG oslo_vmware.api [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52651a7d-909d-ba54-f611-4f98aabe1bd5, 'name': SearchDatastore_Task, 'duration_secs': 0.010627} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2318.503952] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2318.503952] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2318.504309] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2318.504619] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2318.504974] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2318.505591] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b28eaf9d-5b37-4e1f-b7c4-c0134f75e11d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2318.507654] env[62824]: DEBUG oslo_concurrency.lockutils [req-ff980577-74cd-4d01-a99b-de18ad8801ff req-c3aa2b95-4fc3-498b-b7b6-7bc2bbc669fb service nova] Releasing lock "refresh_cache-68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2318.507902] env[62824]: DEBUG nova.compute.manager [req-ff980577-74cd-4d01-a99b-de18ad8801ff req-c3aa2b95-4fc3-498b-b7b6-7bc2bbc669fb service nova] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Received event network-vif-plugged-9190a574-76bc-4236-9f09-69d08897d634 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2318.508666] env[62824]: DEBUG oslo_concurrency.lockutils [req-ff980577-74cd-4d01-a99b-de18ad8801ff req-c3aa2b95-4fc3-498b-b7b6-7bc2bbc669fb service nova] Acquiring lock "c480f680-61fd-496b-8cf7-fa50f580b10d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2318.508666] env[62824]: DEBUG oslo_concurrency.lockutils [req-ff980577-74cd-4d01-a99b-de18ad8801ff req-c3aa2b95-4fc3-498b-b7b6-7bc2bbc669fb service nova] Lock "c480f680-61fd-496b-8cf7-fa50f580b10d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2318.508666] env[62824]: DEBUG oslo_concurrency.lockutils [req-ff980577-74cd-4d01-a99b-de18ad8801ff req-c3aa2b95-4fc3-498b-b7b6-7bc2bbc669fb service nova] Lock "c480f680-61fd-496b-8cf7-fa50f580b10d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2318.508666] env[62824]: DEBUG nova.compute.manager [req-ff980577-74cd-4d01-a99b-de18ad8801ff req-c3aa2b95-4fc3-498b-b7b6-7bc2bbc669fb service nova] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] No waiting events found dispatching network-vif-plugged-9190a574-76bc-4236-9f09-69d08897d634 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2318.508860] env[62824]: WARNING nova.compute.manager [req-ff980577-74cd-4d01-a99b-de18ad8801ff req-c3aa2b95-4fc3-498b-b7b6-7bc2bbc669fb service nova] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Received unexpected event network-vif-plugged-9190a574-76bc-4236-9f09-69d08897d634 for instance with vm_state building and task_state spawning. [ 2318.509187] env[62824]: DEBUG nova.compute.manager [req-ff980577-74cd-4d01-a99b-de18ad8801ff req-c3aa2b95-4fc3-498b-b7b6-7bc2bbc669fb service nova] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Received event network-changed-9190a574-76bc-4236-9f09-69d08897d634 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2318.509187] env[62824]: DEBUG nova.compute.manager [req-ff980577-74cd-4d01-a99b-de18ad8801ff req-c3aa2b95-4fc3-498b-b7b6-7bc2bbc669fb service nova] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Refreshing instance network info cache due to event network-changed-9190a574-76bc-4236-9f09-69d08897d634. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2318.509354] env[62824]: DEBUG oslo_concurrency.lockutils [req-ff980577-74cd-4d01-a99b-de18ad8801ff req-c3aa2b95-4fc3-498b-b7b6-7bc2bbc669fb service nova] Acquiring lock "refresh_cache-c480f680-61fd-496b-8cf7-fa50f580b10d" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2318.509996] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Releasing lock "refresh_cache-c480f680-61fd-496b-8cf7-fa50f580b10d" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2318.510567] env[62824]: DEBUG nova.compute.manager [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Instance network_info: |[{"id": "9190a574-76bc-4236-9f09-69d08897d634", "address": "fa:16:3e:a9:57:8f", "network": {"id": "b995934c-4934-4469-884c-73cb4de1796f", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1892307918-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c11aef7fb6247cb9b5272a6a063cd12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15165046-2de9-4ada-9e99-0126e20854a9", "external-id": "nsx-vlan-transportzone-974", "segmentation_id": 974, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9190a574-76", "ovs_interfaceid": "9190a574-76bc-4236-9f09-69d08897d634", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2318.511193] env[62824]: DEBUG oslo_concurrency.lockutils [req-ff980577-74cd-4d01-a99b-de18ad8801ff req-c3aa2b95-4fc3-498b-b7b6-7bc2bbc669fb service nova] Acquired lock "refresh_cache-c480f680-61fd-496b-8cf7-fa50f580b10d" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2318.511388] env[62824]: DEBUG nova.network.neutron [req-ff980577-74cd-4d01-a99b-de18ad8801ff req-c3aa2b95-4fc3-498b-b7b6-7bc2bbc669fb service nova] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Refreshing network info cache for port 9190a574-76bc-4236-9f09-69d08897d634 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2318.513263] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a9:57:8f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '15165046-2de9-4ada-9e99-0126e20854a9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9190a574-76bc-4236-9f09-69d08897d634', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2318.522093] env[62824]: DEBUG oslo.service.loopingcall [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2318.526860] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2318.527524] env[62824]: DEBUG oslo_vmware.api [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2318.527524] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52fad573-e2b7-cfad-53aa-a31b153bb218" [ 2318.527524] env[62824]: _type = "Task" [ 2318.527524] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2318.527750] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b1c445c2-40e2-4f16-b800-6e1420cbf4d8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2318.555649] env[62824]: DEBUG oslo_vmware.api [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52fad573-e2b7-cfad-53aa-a31b153bb218, 'name': SearchDatastore_Task, 'duration_secs': 0.010586} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2318.557602] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2318.557924] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2318.558243] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2318.558907] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2318.558907] env[62824]: value = "task-2146809" [ 2318.558907] env[62824]: _type = "Task" [ 2318.558907] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2318.569289] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146809, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2318.688329] env[62824]: DEBUG oslo_vmware.api [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146808, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2318.770485] env[62824]: DEBUG nova.network.neutron [req-ff980577-74cd-4d01-a99b-de18ad8801ff req-c3aa2b95-4fc3-498b-b7b6-7bc2bbc669fb service nova] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Updated VIF entry in instance network info cache for port 9190a574-76bc-4236-9f09-69d08897d634. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2318.770724] env[62824]: DEBUG nova.network.neutron [req-ff980577-74cd-4d01-a99b-de18ad8801ff req-c3aa2b95-4fc3-498b-b7b6-7bc2bbc669fb service nova] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Updating instance_info_cache with network_info: [{"id": "9190a574-76bc-4236-9f09-69d08897d634", "address": "fa:16:3e:a9:57:8f", "network": {"id": "b995934c-4934-4469-884c-73cb4de1796f", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1892307918-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c11aef7fb6247cb9b5272a6a063cd12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15165046-2de9-4ada-9e99-0126e20854a9", "external-id": "nsx-vlan-transportzone-974", "segmentation_id": 974, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9190a574-76", "ovs_interfaceid": "9190a574-76bc-4236-9f09-69d08897d634", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2318.773237] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2318.773831] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2318.773831] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2318.773951] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2318.774255] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager.update_available_resource {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2318.845031] env[62824]: DEBUG oslo_vmware.api [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52cad241-f451-5da6-b87f-6132a4946691, 'name': SearchDatastore_Task, 'duration_secs': 0.012512} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2318.845031] env[62824]: DEBUG oslo_concurrency.lockutils [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2318.845031] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 511ef5bb-1dca-4eae-a0f1-19d1d7a17371/511ef5bb-1dca-4eae-a0f1-19d1d7a17371.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2318.845316] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2318.845316] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2318.846056] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4074bb6d-b521-4af8-baad-c0df9e681e55 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2318.847692] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d3a40be4-6700-49f6-a99a-aadbaa6eccca {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2318.857324] env[62824]: DEBUG oslo_vmware.api [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Waiting for the task: (returnval){ [ 2318.857324] env[62824]: value = "task-2146810" [ 2318.857324] env[62824]: _type = "Task" [ 2318.857324] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2318.858868] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2318.859169] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2318.862969] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ddf1ca51-784b-4673-8cf3-df14c4781120 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2318.872222] env[62824]: DEBUG oslo_vmware.api [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146810, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2318.872882] env[62824]: DEBUG oslo_vmware.api [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2318.872882] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52f063cc-057e-6d3f-5624-f02570d91dda" [ 2318.872882] env[62824]: _type = "Task" [ 2318.872882] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2318.883920] env[62824]: DEBUG oslo_vmware.api [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52f063cc-057e-6d3f-5624-f02570d91dda, 'name': SearchDatastore_Task, 'duration_secs': 0.011479} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2318.884906] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b08ed36e-4ac9-4f38-bcce-d728a4de7899 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2318.891705] env[62824]: DEBUG oslo_vmware.api [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2318.891705] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5240b570-6ef4-7870-3b67-3279c098a459" [ 2318.891705] env[62824]: _type = "Task" [ 2318.891705] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2318.901375] env[62824]: DEBUG oslo_vmware.api [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5240b570-6ef4-7870-3b67-3279c098a459, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2318.948385] env[62824]: DEBUG oslo_concurrency.lockutils [None req-96907f0c-6087-4a41-a36e-34211547ef1e tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Lock "bbfcb2e3-9326-4548-b15b-e054cbfd192e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.599s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2319.076043] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146809, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2319.193749] env[62824]: DEBUG oslo_vmware.api [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146808, 'name': PowerOnVM_Task, 'duration_secs': 0.625406} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2319.194090] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2319.194331] env[62824]: DEBUG nova.compute.manager [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2319.195613] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-208738d8-4722-4922-99f8-3b8591b682f3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2319.274535] env[62824]: DEBUG oslo_concurrency.lockutils [req-ff980577-74cd-4d01-a99b-de18ad8801ff req-c3aa2b95-4fc3-498b-b7b6-7bc2bbc669fb service nova] Releasing lock "refresh_cache-c480f680-61fd-496b-8cf7-fa50f580b10d" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2319.279045] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2319.279324] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2319.279518] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2319.279680] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62824) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 2319.281206] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78d3cd25-86cc-4c12-b2cc-a2be9d3da06c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2319.294852] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fea3f47e-51b5-4870-98fa-5934e9c6aa5e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2319.320400] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0d60d2c-6df9-4919-a0d1-62d7622594b4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2319.332378] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd9304ba-dc3d-47cf-bc38-e822bd5bc41b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2319.368634] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179295MB free_disk=174GB free_vcpus=48 pci_devices=None {{(pid=62824) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 2319.368837] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2319.369017] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2319.382024] env[62824]: DEBUG oslo_vmware.api [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146810, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.496603} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2319.382024] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 511ef5bb-1dca-4eae-a0f1-19d1d7a17371/511ef5bb-1dca-4eae-a0f1-19d1d7a17371.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2319.382024] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2319.382024] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9c86f19d-d3c1-42d5-bf14-e965b4849f3c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2319.388736] env[62824]: DEBUG oslo_vmware.api [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Waiting for the task: (returnval){ [ 2319.388736] env[62824]: value = "task-2146812" [ 2319.388736] env[62824]: _type = "Task" [ 2319.388736] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2319.401290] env[62824]: DEBUG oslo_vmware.api [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146812, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2319.405041] env[62824]: DEBUG oslo_vmware.api [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5240b570-6ef4-7870-3b67-3279c098a459, 'name': SearchDatastore_Task, 'duration_secs': 0.010328} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2319.405368] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2319.406187] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 1db9ff97-e59c-43f6-984e-6405d8e993a4/1db9ff97-e59c-43f6-984e-6405d8e993a4.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2319.406187] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2319.406187] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2319.406585] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d4934292-bce0-4990-95cc-7ba9c3132681 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2319.408939] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-11255252-d29f-4f76-95d2-d5e5c9a1a8e8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2319.415136] env[62824]: DEBUG nova.network.neutron [req-2ada9a71-6ae0-4110-9e35-b3c3e4b88a39 req-eab6b440-b266-4c55-9b4e-98f6a93b145e service nova] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Updated VIF entry in instance network info cache for port 3c1f3341-20a4-4b77-8dae-12ef5901a16e. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2319.415136] env[62824]: DEBUG nova.network.neutron [req-2ada9a71-6ae0-4110-9e35-b3c3e4b88a39 req-eab6b440-b266-4c55-9b4e-98f6a93b145e service nova] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Updating instance_info_cache with network_info: [{"id": "3c1f3341-20a4-4b77-8dae-12ef5901a16e", "address": "fa:16:3e:1c:3a:8c", "network": {"id": "aa1c09e5-8d72-43ad-a903-a0f5e711da80", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1643616572-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c987c5f844be42c99cc9d57a9ca84c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9bc2632-36f9-4912-8782-8bbb789f909d", "external-id": "nsx-vlan-transportzone-897", "segmentation_id": 897, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3c1f3341-20", "ovs_interfaceid": "3c1f3341-20a4-4b77-8dae-12ef5901a16e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2319.417690] env[62824]: DEBUG oslo_vmware.api [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2319.417690] env[62824]: value = "task-2146813" [ 2319.417690] env[62824]: _type = "Task" [ 2319.417690] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2319.419142] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2319.419142] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2319.422727] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c14e54cd-6fae-46d0-9e13-09ac8ae6c908 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2319.429497] env[62824]: DEBUG oslo_vmware.api [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2319.429497] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]522ca1eb-cc09-5b6b-d742-fa293c859790" [ 2319.429497] env[62824]: _type = "Task" [ 2319.429497] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2319.433669] env[62824]: DEBUG oslo_vmware.api [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2146813, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2319.443774] env[62824]: DEBUG oslo_vmware.api [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]522ca1eb-cc09-5b6b-d742-fa293c859790, 'name': SearchDatastore_Task, 'duration_secs': 0.011236} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2319.445033] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-34493149-98a7-4a8f-b8a7-96ff34719bae {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2319.450636] env[62824]: DEBUG oslo_vmware.api [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2319.450636] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]527d439c-da01-5c6c-c048-32410d08c0ea" [ 2319.450636] env[62824]: _type = "Task" [ 2319.450636] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2319.460345] env[62824]: DEBUG oslo_vmware.api [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]527d439c-da01-5c6c-c048-32410d08c0ea, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2319.571174] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146809, 'name': CreateVM_Task, 'duration_secs': 0.808644} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2319.571450] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2319.572166] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2319.572364] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2319.572732] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2319.573016] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-45b4ccd8-0fd9-4c7c-b22c-a572703affa6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2319.578909] env[62824]: DEBUG oslo_vmware.api [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 2319.578909] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]520fbe8e-d613-8028-2f14-e2fba106e1b2" [ 2319.578909] env[62824]: _type = "Task" [ 2319.578909] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2319.595546] env[62824]: DEBUG oslo_vmware.api [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]520fbe8e-d613-8028-2f14-e2fba106e1b2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2319.719219] env[62824]: DEBUG oslo_concurrency.lockutils [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2319.899659] env[62824]: DEBUG oslo_vmware.api [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146812, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071933} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2319.899990] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2319.900846] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4ecf080-8cad-4c4a-a062-2edbf23c98ce {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2319.923851] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Reconfiguring VM instance instance-00000069 to attach disk [datastore1] 511ef5bb-1dca-4eae-a0f1-19d1d7a17371/511ef5bb-1dca-4eae-a0f1-19d1d7a17371.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2319.924434] env[62824]: DEBUG oslo_concurrency.lockutils [req-2ada9a71-6ae0-4110-9e35-b3c3e4b88a39 req-eab6b440-b266-4c55-9b4e-98f6a93b145e service nova] Releasing lock "refresh_cache-511ef5bb-1dca-4eae-a0f1-19d1d7a17371" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2319.924873] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-08658072-329c-4cd4-9d6d-c04478785ee9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2319.948714] env[62824]: DEBUG oslo_vmware.api [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2146813, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.455134} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2319.949996] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 1db9ff97-e59c-43f6-984e-6405d8e993a4/1db9ff97-e59c-43f6-984e-6405d8e993a4.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2319.950256] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2319.950567] env[62824]: DEBUG oslo_vmware.api [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Waiting for the task: (returnval){ [ 2319.950567] env[62824]: value = "task-2146814" [ 2319.950567] env[62824]: _type = "Task" [ 2319.950567] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2319.950752] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-89678bd8-e090-4c9f-aa0f-b7a746a51979 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2319.964422] env[62824]: DEBUG oslo_vmware.api [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146814, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2319.968338] env[62824]: DEBUG oslo_vmware.api [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]527d439c-da01-5c6c-c048-32410d08c0ea, 'name': SearchDatastore_Task, 'duration_secs': 0.009914} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2319.968626] env[62824]: DEBUG oslo_vmware.api [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2319.968626] env[62824]: value = "task-2146815" [ 2319.968626] env[62824]: _type = "Task" [ 2319.968626] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2319.968846] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2319.969131] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3/68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2319.969487] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a1541721-0d4d-4d76-a467-c1c3517b2b39 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2319.982111] env[62824]: DEBUG oslo_vmware.api [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2146815, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2319.983764] env[62824]: DEBUG oslo_vmware.api [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2319.983764] env[62824]: value = "task-2146816" [ 2319.983764] env[62824]: _type = "Task" [ 2319.983764] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2319.993884] env[62824]: DEBUG oslo_vmware.api [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146816, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2320.090550] env[62824]: DEBUG oslo_vmware.api [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]520fbe8e-d613-8028-2f14-e2fba106e1b2, 'name': SearchDatastore_Task, 'duration_secs': 0.05553} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2320.090876] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2320.091265] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2320.091417] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2320.091583] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2320.091763] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2320.092052] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e61e5881-0912-4aca-8ee2-d2880a7f08f1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2320.101955] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2320.102158] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2320.102944] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dcf2fa69-9ef8-480e-8d0c-69ddecd32820 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2320.108626] env[62824]: DEBUG oslo_vmware.api [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 2320.108626] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]523e96c5-437d-70b2-274c-0d97da2ee3de" [ 2320.108626] env[62824]: _type = "Task" [ 2320.108626] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2320.116823] env[62824]: DEBUG oslo_vmware.api [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]523e96c5-437d-70b2-274c-0d97da2ee3de, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2320.417416] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance b3003c4b-ae5a-48df-8c12-a915a76253f4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2320.417578] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 6d80ec06-8559-4964-8577-a2512aa366ed actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2320.417705] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance a6cd2032-de60-4f78-bf1e-79801d049df0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2320.417855] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance ca42fef0-1e90-4ab6-9d60-8ef7e4997884 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2320.418045] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 5b3df31f-15fe-473f-992c-ddb272661c53 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2320.418106] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2320.418255] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 71e9733b-c84b-4501-8faf-a487f089b498 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2320.418392] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 1805a6e7-48f2-47b4-b097-0644ba8217fa actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2320.418707] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance e1574266-e53b-4169-9db6-c66b84895edb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2320.418879] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 3d294ee9-92f6-4115-95db-9d4b66562b56 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2320.419054] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2320.419200] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 1db9ff97-e59c-43f6-984e-6405d8e993a4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2320.419375] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 511ef5bb-1dca-4eae-a0f1-19d1d7a17371 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2320.419550] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance c480f680-61fd-496b-8cf7-fa50f580b10d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2320.419825] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Total usable vcpus: 48, total allocated vcpus: 14 {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 2320.419972] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3200MB phys_disk=200GB used_disk=14GB total_vcpus=48 used_vcpus=14 pci_stats=[] {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 2320.467343] env[62824]: DEBUG oslo_vmware.api [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146814, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2320.479172] env[62824]: DEBUG oslo_vmware.api [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2146815, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071498} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2320.479463] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2320.480836] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db6c6f49-4e68-4f55-bdf5-131ccdd5d247 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2320.513453] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Reconfiguring VM instance instance-00000068 to attach disk [datastore1] 1db9ff97-e59c-43f6-984e-6405d8e993a4/1db9ff97-e59c-43f6-984e-6405d8e993a4.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2320.517182] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6e749769-26f5-4ce2-9c2c-7ed83535e3b4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2320.537506] env[62824]: DEBUG oslo_vmware.api [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146816, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.480084} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2320.539262] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3/68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2320.539517] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2320.539831] env[62824]: DEBUG oslo_vmware.api [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2320.539831] env[62824]: value = "task-2146817" [ 2320.539831] env[62824]: _type = "Task" [ 2320.539831] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2320.540063] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6acc4cf0-0cb1-46e9-a3ec-2963abfffa63 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2320.555269] env[62824]: DEBUG oslo_vmware.api [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2146817, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2320.556109] env[62824]: DEBUG oslo_vmware.api [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2320.556109] env[62824]: value = "task-2146818" [ 2320.556109] env[62824]: _type = "Task" [ 2320.556109] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2320.569333] env[62824]: DEBUG oslo_vmware.api [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146818, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2320.620748] env[62824]: DEBUG oslo_vmware.api [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]523e96c5-437d-70b2-274c-0d97da2ee3de, 'name': SearchDatastore_Task, 'duration_secs': 0.011335} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2320.626048] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-42869709-f470-4180-85c3-728662dd47be {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2320.633515] env[62824]: DEBUG oslo_vmware.api [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 2320.633515] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5250e57c-2455-6462-f06a-8bfbcf35be40" [ 2320.633515] env[62824]: _type = "Task" [ 2320.633515] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2320.648703] env[62824]: DEBUG oslo_vmware.api [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5250e57c-2455-6462-f06a-8bfbcf35be40, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2320.684157] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9ba85f9-dcf6-4d4d-93ed-eabc0882d2f7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2320.693648] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-673c280b-6706-4e0a-8c41-bb44d1a0a72e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2320.731380] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70a22d3b-6e05-4ee7-b0c1-2acf9f7c16e5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2320.740853] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fb0318b-ba19-4c2c-93cb-7f0f68610570 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2320.757937] env[62824]: DEBUG nova.compute.provider_tree [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2320.964852] env[62824]: DEBUG oslo_vmware.api [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146814, 'name': ReconfigVM_Task, 'duration_secs': 0.562062} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2320.965163] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Reconfigured VM instance instance-00000069 to attach disk [datastore1] 511ef5bb-1dca-4eae-a0f1-19d1d7a17371/511ef5bb-1dca-4eae-a0f1-19d1d7a17371.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2320.965809] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-78e85b67-3352-4662-9d33-4b86a0f0ab7f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2320.973333] env[62824]: DEBUG oslo_vmware.api [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Waiting for the task: (returnval){ [ 2320.973333] env[62824]: value = "task-2146819" [ 2320.973333] env[62824]: _type = "Task" [ 2320.973333] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2320.981529] env[62824]: DEBUG oslo_vmware.api [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146819, 'name': Rename_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2321.052915] env[62824]: DEBUG oslo_vmware.api [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2146817, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2321.069159] env[62824]: DEBUG oslo_vmware.api [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146818, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.142067} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2321.069473] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2321.070465] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-972e2256-3bd5-4627-8904-104e1fedbe8b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2321.095668] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3] Reconfiguring VM instance instance-00000067 to attach disk [datastore1] 68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3/68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2321.096224] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9eadcb07-c2f2-4daf-ab0e-537cd63453af tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquiring lock "1805a6e7-48f2-47b4-b097-0644ba8217fa" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2321.096391] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9eadcb07-c2f2-4daf-ab0e-537cd63453af tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "1805a6e7-48f2-47b4-b097-0644ba8217fa" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2321.096712] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9eadcb07-c2f2-4daf-ab0e-537cd63453af tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquiring lock "1805a6e7-48f2-47b4-b097-0644ba8217fa-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2321.096972] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9eadcb07-c2f2-4daf-ab0e-537cd63453af tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "1805a6e7-48f2-47b4-b097-0644ba8217fa-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2321.097116] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9eadcb07-c2f2-4daf-ab0e-537cd63453af tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "1805a6e7-48f2-47b4-b097-0644ba8217fa-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2321.098772] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-31e0c7a3-7389-4613-996c-f74e3acb2ed3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2321.113796] env[62824]: INFO nova.compute.manager [None req-9eadcb07-c2f2-4daf-ab0e-537cd63453af tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Terminating instance [ 2321.123485] env[62824]: DEBUG oslo_vmware.api [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2321.123485] env[62824]: value = "task-2146820" [ 2321.123485] env[62824]: _type = "Task" [ 2321.123485] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2321.133020] env[62824]: DEBUG oslo_vmware.api [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146820, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2321.144281] env[62824]: DEBUG oslo_vmware.api [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5250e57c-2455-6462-f06a-8bfbcf35be40, 'name': SearchDatastore_Task, 'duration_secs': 0.018864} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2321.144864] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2321.145365] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] c480f680-61fd-496b-8cf7-fa50f580b10d/c480f680-61fd-496b-8cf7-fa50f580b10d.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2321.145485] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-816da9f7-2f53-41eb-9adf-b9176dd11add {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2321.154916] env[62824]: DEBUG oslo_vmware.api [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 2321.154916] env[62824]: value = "task-2146821" [ 2321.154916] env[62824]: _type = "Task" [ 2321.154916] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2321.165492] env[62824]: DEBUG oslo_vmware.api [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146821, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2321.260989] env[62824]: DEBUG nova.scheduler.client.report [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2321.485331] env[62824]: DEBUG oslo_vmware.api [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146819, 'name': Rename_Task, 'duration_secs': 0.177066} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2321.485659] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2321.486060] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-64c4043d-9ca3-4e1f-90d4-b6ba688a2f96 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2321.495992] env[62824]: DEBUG oslo_vmware.api [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Waiting for the task: (returnval){ [ 2321.495992] env[62824]: value = "task-2146822" [ 2321.495992] env[62824]: _type = "Task" [ 2321.495992] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2321.507918] env[62824]: DEBUG oslo_vmware.api [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146822, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2321.555796] env[62824]: DEBUG oslo_vmware.api [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2146817, 'name': ReconfigVM_Task, 'duration_secs': 0.837987} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2321.555996] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Reconfigured VM instance instance-00000068 to attach disk [datastore1] 1db9ff97-e59c-43f6-984e-6405d8e993a4/1db9ff97-e59c-43f6-984e-6405d8e993a4.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2321.557832] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6333aa1e-f904-42be-9a6e-20aeb6ab0371 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2321.565236] env[62824]: DEBUG oslo_vmware.api [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2321.565236] env[62824]: value = "task-2146823" [ 2321.565236] env[62824]: _type = "Task" [ 2321.565236] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2321.575795] env[62824]: DEBUG oslo_vmware.api [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2146823, 'name': Rename_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2321.622704] env[62824]: DEBUG nova.compute.manager [None req-9eadcb07-c2f2-4daf-ab0e-537cd63453af tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2321.623242] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-9eadcb07-c2f2-4daf-ab0e-537cd63453af tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2321.623897] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df5d35de-6f83-4c6d-a663-ce4cc851f040 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2321.636533] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-9eadcb07-c2f2-4daf-ab0e-537cd63453af tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2321.640152] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-362b4140-ed11-476c-ba04-f0392f6f9bb7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2321.642172] env[62824]: DEBUG oslo_vmware.api [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146820, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2321.651374] env[62824]: DEBUG oslo_vmware.api [None req-9eadcb07-c2f2-4daf-ab0e-537cd63453af tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2321.651374] env[62824]: value = "task-2146824" [ 2321.651374] env[62824]: _type = "Task" [ 2321.651374] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2321.664353] env[62824]: DEBUG oslo_vmware.api [None req-9eadcb07-c2f2-4daf-ab0e-537cd63453af tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146824, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2321.675615] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquiring lock "bf6e681d-b773-4787-886a-c90586bc81bf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2321.675903] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "bf6e681d-b773-4787-886a-c90586bc81bf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2321.677183] env[62824]: DEBUG oslo_vmware.api [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146821, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.511863} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2321.678957] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] c480f680-61fd-496b-8cf7-fa50f580b10d/c480f680-61fd-496b-8cf7-fa50f580b10d.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2321.678957] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2321.678957] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-00fa18a5-a9fd-4807-a2ed-077fa3e31296 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2321.691077] env[62824]: DEBUG oslo_vmware.api [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 2321.691077] env[62824]: value = "task-2146825" [ 2321.691077] env[62824]: _type = "Task" [ 2321.691077] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2321.702783] env[62824]: DEBUG oslo_vmware.api [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146825, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2321.766875] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62824) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 2321.767143] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.398s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2321.767447] env[62824]: DEBUG oslo_concurrency.lockutils [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 2.048s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2321.767671] env[62824]: DEBUG nova.objects.instance [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62824) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 2321.770604] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2321.770814] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Cleaning up deleted instances with incomplete migration {{(pid=62824) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11783}} [ 2322.008079] env[62824]: DEBUG oslo_vmware.api [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146822, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2322.079515] env[62824]: DEBUG oslo_vmware.api [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2146823, 'name': Rename_Task, 'duration_secs': 0.229404} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2322.079959] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2322.080847] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-82a8349b-e402-4641-b956-48ca0a87b6fb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2322.095580] env[62824]: DEBUG oslo_vmware.api [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2322.095580] env[62824]: value = "task-2146826" [ 2322.095580] env[62824]: _type = "Task" [ 2322.095580] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2322.107129] env[62824]: DEBUG oslo_vmware.api [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2146826, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2322.141656] env[62824]: DEBUG oslo_vmware.api [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146820, 'name': ReconfigVM_Task, 'duration_secs': 0.606497} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2322.144354] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3] Reconfigured VM instance instance-00000067 to attach disk [datastore1] 68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3/68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2322.144354] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fc4723c3-519f-41c5-b5da-79146e064719 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2322.152457] env[62824]: DEBUG oslo_vmware.api [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2322.152457] env[62824]: value = "task-2146827" [ 2322.152457] env[62824]: _type = "Task" [ 2322.152457] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2322.166842] env[62824]: DEBUG oslo_vmware.api [None req-9eadcb07-c2f2-4daf-ab0e-537cd63453af tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146824, 'name': PowerOffVM_Task, 'duration_secs': 0.209516} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2322.171075] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-9eadcb07-c2f2-4daf-ab0e-537cd63453af tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2322.171315] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-9eadcb07-c2f2-4daf-ab0e-537cd63453af tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2322.171712] env[62824]: DEBUG oslo_vmware.api [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146827, 'name': Rename_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2322.172022] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4f3ac992-d535-46c0-8ca1-e4de50a69231 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2322.179960] env[62824]: DEBUG nova.compute.manager [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2322.205941] env[62824]: DEBUG oslo_vmware.api [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146825, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.084391} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2322.206061] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2322.207539] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c6d56b3-bbd6-4cac-b71c-24ede1da36ea {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2322.233284] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] c480f680-61fd-496b-8cf7-fa50f580b10d/c480f680-61fd-496b-8cf7-fa50f580b10d.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2322.234409] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7d2002aa-67f8-45d5-9e0c-32912c58a458 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2322.258540] env[62824]: DEBUG oslo_vmware.api [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 2322.258540] env[62824]: value = "task-2146829" [ 2322.258540] env[62824]: _type = "Task" [ 2322.258540] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2322.269344] env[62824]: DEBUG oslo_vmware.api [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146829, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2322.283174] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-9eadcb07-c2f2-4daf-ab0e-537cd63453af tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2322.283462] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-9eadcb07-c2f2-4daf-ab0e-537cd63453af tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Deleting contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2322.283652] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-9eadcb07-c2f2-4daf-ab0e-537cd63453af tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Deleting the datastore file [datastore1] 1805a6e7-48f2-47b4-b097-0644ba8217fa {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2322.284219] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1224a714-80d2-4ac3-a938-04009d73691a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2322.293039] env[62824]: DEBUG oslo_vmware.api [None req-9eadcb07-c2f2-4daf-ab0e-537cd63453af tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2322.293039] env[62824]: value = "task-2146830" [ 2322.293039] env[62824]: _type = "Task" [ 2322.293039] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2322.304494] env[62824]: DEBUG oslo_vmware.api [None req-9eadcb07-c2f2-4daf-ab0e-537cd63453af tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146830, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2322.507845] env[62824]: DEBUG oslo_vmware.api [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146822, 'name': PowerOnVM_Task, 'duration_secs': 0.619121} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2322.508088] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2322.508313] env[62824]: INFO nova.compute.manager [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Took 9.04 seconds to spawn the instance on the hypervisor. [ 2322.508499] env[62824]: DEBUG nova.compute.manager [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2322.509365] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00c18e35-1d1f-4cd4-9079-76b3e6671f01 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2322.605927] env[62824]: DEBUG oslo_vmware.api [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2146826, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2322.668752] env[62824]: DEBUG oslo_vmware.api [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146827, 'name': Rename_Task, 'duration_secs': 0.223684} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2322.669127] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2322.669320] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a79b7a1b-d845-406c-8ef7-487031a430f2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2322.677721] env[62824]: DEBUG oslo_vmware.api [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2322.677721] env[62824]: value = "task-2146831" [ 2322.677721] env[62824]: _type = "Task" [ 2322.677721] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2322.690033] env[62824]: DEBUG oslo_vmware.api [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146831, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2322.711262] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2322.772973] env[62824]: DEBUG oslo_vmware.api [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146829, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2322.785645] env[62824]: DEBUG oslo_concurrency.lockutils [None req-745be223-43a1-4789-bf4b-925d5b7284f2 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.018s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2322.787353] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.076s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2322.788986] env[62824]: INFO nova.compute.claims [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2322.806895] env[62824]: DEBUG oslo_vmware.api [None req-9eadcb07-c2f2-4daf-ab0e-537cd63453af tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146830, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.349953} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2322.807326] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-9eadcb07-c2f2-4daf-ab0e-537cd63453af tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2322.807642] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-9eadcb07-c2f2-4daf-ab0e-537cd63453af tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Deleted contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2322.807813] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-9eadcb07-c2f2-4daf-ab0e-537cd63453af tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2322.808093] env[62824]: INFO nova.compute.manager [None req-9eadcb07-c2f2-4daf-ab0e-537cd63453af tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Took 1.19 seconds to destroy the instance on the hypervisor. [ 2322.809052] env[62824]: DEBUG oslo.service.loopingcall [None req-9eadcb07-c2f2-4daf-ab0e-537cd63453af tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2322.809052] env[62824]: DEBUG nova.compute.manager [-] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2322.809052] env[62824]: DEBUG nova.network.neutron [-] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2323.027223] env[62824]: INFO nova.compute.manager [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Took 15.69 seconds to build instance. [ 2323.107764] env[62824]: DEBUG oslo_vmware.api [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2146826, 'name': PowerOnVM_Task, 'duration_secs': 0.703077} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2323.111222] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2323.111222] env[62824]: INFO nova.compute.manager [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Took 11.95 seconds to spawn the instance on the hypervisor. [ 2323.111222] env[62824]: DEBUG nova.compute.manager [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2323.111222] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8041f150-58e4-4e32-8ec0-6b3769c7a574 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2323.188889] env[62824]: DEBUG oslo_vmware.api [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146831, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2323.245905] env[62824]: DEBUG nova.compute.manager [req-71d1fc48-029c-4aa5-abae-15d6d79c22bb req-b53932ee-8566-4032-81ef-cc67dd5518c2 service nova] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Received event network-vif-deleted-37654518-0a1a-44e2-8297-2169c44d0aad {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2323.246163] env[62824]: INFO nova.compute.manager [req-71d1fc48-029c-4aa5-abae-15d6d79c22bb req-b53932ee-8566-4032-81ef-cc67dd5518c2 service nova] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Neutron deleted interface 37654518-0a1a-44e2-8297-2169c44d0aad; detaching it from the instance and deleting it from the info cache [ 2323.246350] env[62824]: DEBUG nova.network.neutron [req-71d1fc48-029c-4aa5-abae-15d6d79c22bb req-b53932ee-8566-4032-81ef-cc67dd5518c2 service nova] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2323.270112] env[62824]: DEBUG oslo_vmware.api [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146829, 'name': ReconfigVM_Task, 'duration_secs': 0.611681} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2323.270439] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Reconfigured VM instance instance-0000006a to attach disk [datastore1] c480f680-61fd-496b-8cf7-fa50f580b10d/c480f680-61fd-496b-8cf7-fa50f580b10d.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2323.271058] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b281a96f-f5be-466d-bcd4-7cfcbfae4945 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2323.277930] env[62824]: DEBUG oslo_vmware.api [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 2323.277930] env[62824]: value = "task-2146832" [ 2323.277930] env[62824]: _type = "Task" [ 2323.277930] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2323.281639] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2323.281639] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Starting heal instance info cache {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10445}} [ 2323.281788] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Rebuilding the list of instances to heal {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10449}} [ 2323.289969] env[62824]: DEBUG oslo_vmware.api [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146832, 'name': Rename_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2323.529707] env[62824]: DEBUG oslo_concurrency.lockutils [None req-74d60396-7b72-4970-9571-c985d0efb492 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lock "511ef5bb-1dca-4eae-a0f1-19d1d7a17371" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.199s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2323.612818] env[62824]: DEBUG nova.compute.manager [req-418f046e-0701-44e1-adce-923cd7efecff req-4e5b8d60-f25a-41c7-9d4f-e1079cc9dbc5 service nova] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Received event network-changed-d5c6ba87-e0c7-49c2-bbaf-b7c64f84cdf8 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2323.613371] env[62824]: DEBUG nova.compute.manager [req-418f046e-0701-44e1-adce-923cd7efecff req-4e5b8d60-f25a-41c7-9d4f-e1079cc9dbc5 service nova] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Refreshing instance network info cache due to event network-changed-d5c6ba87-e0c7-49c2-bbaf-b7c64f84cdf8. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2323.613610] env[62824]: DEBUG oslo_concurrency.lockutils [req-418f046e-0701-44e1-adce-923cd7efecff req-4e5b8d60-f25a-41c7-9d4f-e1079cc9dbc5 service nova] Acquiring lock "refresh_cache-e1574266-e53b-4169-9db6-c66b84895edb" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2323.613830] env[62824]: DEBUG oslo_concurrency.lockutils [req-418f046e-0701-44e1-adce-923cd7efecff req-4e5b8d60-f25a-41c7-9d4f-e1079cc9dbc5 service nova] Acquired lock "refresh_cache-e1574266-e53b-4169-9db6-c66b84895edb" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2323.614008] env[62824]: DEBUG nova.network.neutron [req-418f046e-0701-44e1-adce-923cd7efecff req-4e5b8d60-f25a-41c7-9d4f-e1079cc9dbc5 service nova] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Refreshing network info cache for port d5c6ba87-e0c7-49c2-bbaf-b7c64f84cdf8 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2323.635631] env[62824]: INFO nova.compute.manager [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Took 17.54 seconds to build instance. [ 2323.689017] env[62824]: DEBUG oslo_vmware.api [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146831, 'name': PowerOnVM_Task, 'duration_secs': 0.983553} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2323.689397] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2323.689397] env[62824]: INFO nova.compute.manager [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3] Took 16.73 seconds to spawn the instance on the hypervisor. [ 2323.689637] env[62824]: DEBUG nova.compute.manager [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2323.690413] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44d50a90-5cad-4402-a4ad-dcebca5672c0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2323.723379] env[62824]: DEBUG nova.network.neutron [-] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2323.749363] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-131f867c-a63e-4abd-973d-4d3aa64d06bb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2323.766234] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f84341f2-ded4-4dd2-b076-f224ada48eaa {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2323.792796] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Skipping network cache update for instance because it is being deleted. {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10462}} [ 2323.793061] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3] Skipping network cache update for instance because it is Building. {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10458}} [ 2323.793286] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Skipping network cache update for instance because it is Building. {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10458}} [ 2323.793481] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Skipping network cache update for instance because it is Building. {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10458}} [ 2323.825121] env[62824]: DEBUG nova.compute.manager [req-71d1fc48-029c-4aa5-abae-15d6d79c22bb req-b53932ee-8566-4032-81ef-cc67dd5518c2 service nova] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Detach interface failed, port_id=37654518-0a1a-44e2-8297-2169c44d0aad, reason: Instance 1805a6e7-48f2-47b4-b097-0644ba8217fa could not be found. {{(pid=62824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11482}} [ 2323.829254] env[62824]: DEBUG oslo_vmware.api [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146832, 'name': Rename_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2323.834959] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1f9e537a-d5e5-4905-854b-725e8ff9d4b4 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Acquiring lock "6d80ec06-8559-4964-8577-a2512aa366ed" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2323.835245] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1f9e537a-d5e5-4905-854b-725e8ff9d4b4 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Lock "6d80ec06-8559-4964-8577-a2512aa366ed" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2323.835471] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1f9e537a-d5e5-4905-854b-725e8ff9d4b4 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Acquiring lock "6d80ec06-8559-4964-8577-a2512aa366ed-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2323.835657] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1f9e537a-d5e5-4905-854b-725e8ff9d4b4 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Lock "6d80ec06-8559-4964-8577-a2512aa366ed-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2323.835840] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1f9e537a-d5e5-4905-854b-725e8ff9d4b4 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Lock "6d80ec06-8559-4964-8577-a2512aa366ed-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2323.838049] env[62824]: INFO nova.compute.manager [None req-1f9e537a-d5e5-4905-854b-725e8ff9d4b4 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: 6d80ec06-8559-4964-8577-a2512aa366ed] Terminating instance [ 2323.855495] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "refresh_cache-b3003c4b-ae5a-48df-8c12-a915a76253f4" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2323.855632] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquired lock "refresh_cache-b3003c4b-ae5a-48df-8c12-a915a76253f4" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2323.855795] env[62824]: DEBUG nova.network.neutron [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Forcefully refreshing network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 2323.855965] env[62824]: DEBUG nova.objects.instance [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lazy-loading 'info_cache' on Instance uuid b3003c4b-ae5a-48df-8c12-a915a76253f4 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2324.057208] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0195ffb6-10a3-4300-a333-0421c9cab82b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2324.067565] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-508cef80-a51a-4663-b353-9528f65385e7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2324.104182] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bed81ca3-3831-4f8d-b984-a0a184ade809 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2324.113278] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-029c66b2-74ea-4f46-8963-5f0db5c9b4b3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2324.131310] env[62824]: DEBUG nova.compute.provider_tree [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2324.138036] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9525db4b-d2c5-4539-886b-7604d1caf2b9 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lock "1db9ff97-e59c-43f6-984e-6405d8e993a4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.056s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2324.212976] env[62824]: INFO nova.compute.manager [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3] Took 22.80 seconds to build instance. [ 2324.226865] env[62824]: INFO nova.compute.manager [-] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Took 1.42 seconds to deallocate network for instance. [ 2324.296768] env[62824]: DEBUG oslo_vmware.api [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146832, 'name': Rename_Task, 'duration_secs': 0.689993} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2324.297067] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2324.297579] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f491aa1d-52b2-41f9-99f6-f55f01d45d27 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2324.305320] env[62824]: DEBUG oslo_vmware.api [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 2324.305320] env[62824]: value = "task-2146833" [ 2324.305320] env[62824]: _type = "Task" [ 2324.305320] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2324.314608] env[62824]: DEBUG oslo_vmware.api [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146833, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2324.342678] env[62824]: DEBUG nova.compute.manager [None req-1f9e537a-d5e5-4905-854b-725e8ff9d4b4 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: 6d80ec06-8559-4964-8577-a2512aa366ed] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2324.342905] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-1f9e537a-d5e5-4905-854b-725e8ff9d4b4 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: 6d80ec06-8559-4964-8577-a2512aa366ed] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2324.344238] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe10cfe2-3c88-41cf-8d30-535663cc21f8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2324.355025] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f9e537a-d5e5-4905-854b-725e8ff9d4b4 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: 6d80ec06-8559-4964-8577-a2512aa366ed] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2324.355025] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5bd2fdae-e12f-497b-a07f-c48f33d44eb6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2324.362222] env[62824]: DEBUG oslo_vmware.api [None req-1f9e537a-d5e5-4905-854b-725e8ff9d4b4 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for the task: (returnval){ [ 2324.362222] env[62824]: value = "task-2146834" [ 2324.362222] env[62824]: _type = "Task" [ 2324.362222] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2324.373483] env[62824]: DEBUG oslo_vmware.api [None req-1f9e537a-d5e5-4905-854b-725e8ff9d4b4 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2146834, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2324.634655] env[62824]: DEBUG nova.scheduler.client.report [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2324.718484] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3d5685ec-258c-45a7-95e0-896f64f70f78 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.312s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2324.739151] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9eadcb07-c2f2-4daf-ab0e-537cd63453af tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2324.817825] env[62824]: DEBUG oslo_vmware.api [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146833, 'name': PowerOnVM_Task, 'duration_secs': 0.487432} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2324.819637] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2324.819880] env[62824]: INFO nova.compute.manager [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Took 9.02 seconds to spawn the instance on the hypervisor. [ 2324.820098] env[62824]: DEBUG nova.compute.manager [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2324.821304] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a9266eb-8abb-460e-bc96-990754589ab4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2324.877040] env[62824]: DEBUG oslo_vmware.api [None req-1f9e537a-d5e5-4905-854b-725e8ff9d4b4 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2146834, 'name': PowerOffVM_Task, 'duration_secs': 0.318128} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2324.878277] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f9e537a-d5e5-4905-854b-725e8ff9d4b4 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: 6d80ec06-8559-4964-8577-a2512aa366ed] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2324.878465] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-1f9e537a-d5e5-4905-854b-725e8ff9d4b4 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: 6d80ec06-8559-4964-8577-a2512aa366ed] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2324.878739] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3db451c5-895d-4b6b-8de7-093430676eab {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2324.939152] env[62824]: DEBUG nova.network.neutron [req-418f046e-0701-44e1-adce-923cd7efecff req-4e5b8d60-f25a-41c7-9d4f-e1079cc9dbc5 service nova] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Updated VIF entry in instance network info cache for port d5c6ba87-e0c7-49c2-bbaf-b7c64f84cdf8. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2324.939706] env[62824]: DEBUG nova.network.neutron [req-418f046e-0701-44e1-adce-923cd7efecff req-4e5b8d60-f25a-41c7-9d4f-e1079cc9dbc5 service nova] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Updating instance_info_cache with network_info: [{"id": "d5c6ba87-e0c7-49c2-bbaf-b7c64f84cdf8", "address": "fa:16:3e:c4:9c:f4", "network": {"id": "aa1c09e5-8d72-43ad-a903-a0f5e711da80", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1643616572-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c987c5f844be42c99cc9d57a9ca84c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9bc2632-36f9-4912-8782-8bbb789f909d", "external-id": "nsx-vlan-transportzone-897", "segmentation_id": 897, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd5c6ba87-e0", "ovs_interfaceid": "d5c6ba87-e0c7-49c2-bbaf-b7c64f84cdf8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2325.141852] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.354s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2325.142407] env[62824]: DEBUG nova.compute.manager [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2325.149702] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9eadcb07-c2f2-4daf-ab0e-537cd63453af tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.411s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2325.149930] env[62824]: DEBUG nova.objects.instance [None req-9eadcb07-c2f2-4daf-ab0e-537cd63453af tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lazy-loading 'resources' on Instance uuid 1805a6e7-48f2-47b4-b097-0644ba8217fa {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2325.307896] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-1f9e537a-d5e5-4905-854b-725e8ff9d4b4 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: 6d80ec06-8559-4964-8577-a2512aa366ed] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2325.308143] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-1f9e537a-d5e5-4905-854b-725e8ff9d4b4 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: 6d80ec06-8559-4964-8577-a2512aa366ed] Deleting contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2325.308458] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-1f9e537a-d5e5-4905-854b-725e8ff9d4b4 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Deleting the datastore file [datastore2] 6d80ec06-8559-4964-8577-a2512aa366ed {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2325.308730] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-27b34c19-9fcb-43fc-8472-4a42c546f7d4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2325.317070] env[62824]: DEBUG oslo_vmware.api [None req-1f9e537a-d5e5-4905-854b-725e8ff9d4b4 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for the task: (returnval){ [ 2325.317070] env[62824]: value = "task-2146836" [ 2325.317070] env[62824]: _type = "Task" [ 2325.317070] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2325.327118] env[62824]: DEBUG oslo_vmware.api [None req-1f9e537a-d5e5-4905-854b-725e8ff9d4b4 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2146836, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2325.346321] env[62824]: INFO nova.compute.manager [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Took 15.16 seconds to build instance. [ 2325.398320] env[62824]: DEBUG nova.compute.manager [req-dbf65db9-17eb-464f-b0ff-d538924db10e req-0f59977f-1fd3-4301-b6e1-0d5cb81fff7b service nova] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Received event network-changed-aeeadfdd-d869-4555-bda2-391c435de1bb {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2325.398641] env[62824]: DEBUG nova.compute.manager [req-dbf65db9-17eb-464f-b0ff-d538924db10e req-0f59977f-1fd3-4301-b6e1-0d5cb81fff7b service nova] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Refreshing instance network info cache due to event network-changed-aeeadfdd-d869-4555-bda2-391c435de1bb. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2325.398886] env[62824]: DEBUG oslo_concurrency.lockutils [req-dbf65db9-17eb-464f-b0ff-d538924db10e req-0f59977f-1fd3-4301-b6e1-0d5cb81fff7b service nova] Acquiring lock "refresh_cache-1db9ff97-e59c-43f6-984e-6405d8e993a4" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2325.399767] env[62824]: DEBUG oslo_concurrency.lockutils [req-dbf65db9-17eb-464f-b0ff-d538924db10e req-0f59977f-1fd3-4301-b6e1-0d5cb81fff7b service nova] Acquired lock "refresh_cache-1db9ff97-e59c-43f6-984e-6405d8e993a4" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2325.399767] env[62824]: DEBUG nova.network.neutron [req-dbf65db9-17eb-464f-b0ff-d538924db10e req-0f59977f-1fd3-4301-b6e1-0d5cb81fff7b service nova] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Refreshing network info cache for port aeeadfdd-d869-4555-bda2-391c435de1bb {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2325.442210] env[62824]: DEBUG oslo_concurrency.lockutils [req-418f046e-0701-44e1-adce-923cd7efecff req-4e5b8d60-f25a-41c7-9d4f-e1079cc9dbc5 service nova] Releasing lock "refresh_cache-e1574266-e53b-4169-9db6-c66b84895edb" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2325.471587] env[62824]: DEBUG oslo_vmware.rw_handles [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52724aed-6614-2bc5-de89-88a16d737b6b/disk-0.vmdk. {{(pid=62824) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 2325.472538] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ad5f338-e6eb-4dc9-8aab-a1ba1043a342 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2325.482460] env[62824]: DEBUG oslo_vmware.rw_handles [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52724aed-6614-2bc5-de89-88a16d737b6b/disk-0.vmdk is in state: ready. {{(pid=62824) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 2325.482636] env[62824]: ERROR oslo_vmware.rw_handles [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52724aed-6614-2bc5-de89-88a16d737b6b/disk-0.vmdk due to incomplete transfer. [ 2325.482871] env[62824]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-0958e993-a05a-4799-ad04-b62efe45f87f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2325.491907] env[62824]: DEBUG oslo_vmware.rw_handles [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52724aed-6614-2bc5-de89-88a16d737b6b/disk-0.vmdk. {{(pid=62824) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 2325.492125] env[62824]: DEBUG nova.virt.vmwareapi.images [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Uploaded image 91de84ba-a43c-4db9-92a7-d2e20dada664 to the Glance image server {{(pid=62824) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 2325.494290] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Destroying the VM {{(pid=62824) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 2325.494554] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-ef544390-366c-44d9-9199-d3539801dfd6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2325.505719] env[62824]: DEBUG oslo_vmware.api [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Waiting for the task: (returnval){ [ 2325.505719] env[62824]: value = "task-2146837" [ 2325.505719] env[62824]: _type = "Task" [ 2325.505719] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2325.515846] env[62824]: DEBUG oslo_vmware.api [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146837, 'name': Destroy_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2325.646583] env[62824]: DEBUG nova.compute.manager [req-7df4e318-0490-459b-a6e5-da7f5d715943 req-ce12c3bc-78aa-4005-9e33-c25b0426eedd service nova] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Received event network-changed-3c1f3341-20a4-4b77-8dae-12ef5901a16e {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2325.646779] env[62824]: DEBUG nova.compute.manager [req-7df4e318-0490-459b-a6e5-da7f5d715943 req-ce12c3bc-78aa-4005-9e33-c25b0426eedd service nova] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Refreshing instance network info cache due to event network-changed-3c1f3341-20a4-4b77-8dae-12ef5901a16e. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2325.646991] env[62824]: DEBUG oslo_concurrency.lockutils [req-7df4e318-0490-459b-a6e5-da7f5d715943 req-ce12c3bc-78aa-4005-9e33-c25b0426eedd service nova] Acquiring lock "refresh_cache-511ef5bb-1dca-4eae-a0f1-19d1d7a17371" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2325.647250] env[62824]: DEBUG oslo_concurrency.lockutils [req-7df4e318-0490-459b-a6e5-da7f5d715943 req-ce12c3bc-78aa-4005-9e33-c25b0426eedd service nova] Acquired lock "refresh_cache-511ef5bb-1dca-4eae-a0f1-19d1d7a17371" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2325.647486] env[62824]: DEBUG nova.network.neutron [req-7df4e318-0490-459b-a6e5-da7f5d715943 req-ce12c3bc-78aa-4005-9e33-c25b0426eedd service nova] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Refreshing network info cache for port 3c1f3341-20a4-4b77-8dae-12ef5901a16e {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2325.652873] env[62824]: DEBUG nova.compute.utils [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2325.654194] env[62824]: DEBUG nova.compute.manager [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2325.654393] env[62824]: DEBUG nova.network.neutron [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2325.683318] env[62824]: DEBUG nova.network.neutron [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Updating instance_info_cache with network_info: [{"id": "4e0d51df-4419-4059-9dab-be44957e4f77", "address": "fa:16:3e:15:1a:78", "network": {"id": "a410f4f3-d072-4034-b797-8210e1a79093", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1482574391-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e10208040df34eb5adfe8dcbc76043d5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f4399275-8e92-4448-be9e-d4984e93e89c", "external-id": "nsx-vlan-transportzone-192", "segmentation_id": 192, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e0d51df-44", "ovs_interfaceid": "4e0d51df-4419-4059-9dab-be44957e4f77", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2325.700917] env[62824]: DEBUG oslo_concurrency.lockutils [None req-07530f12-f1bb-49d0-8972-cfc1629bba0b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquiring lock "68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2325.701202] env[62824]: DEBUG oslo_concurrency.lockutils [None req-07530f12-f1bb-49d0-8972-cfc1629bba0b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2325.701421] env[62824]: DEBUG oslo_concurrency.lockutils [None req-07530f12-f1bb-49d0-8972-cfc1629bba0b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquiring lock "68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2325.701606] env[62824]: DEBUG oslo_concurrency.lockutils [None req-07530f12-f1bb-49d0-8972-cfc1629bba0b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2325.701778] env[62824]: DEBUG oslo_concurrency.lockutils [None req-07530f12-f1bb-49d0-8972-cfc1629bba0b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2325.704682] env[62824]: DEBUG nova.policy [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e8c10fcbfe69448fa71cfad4a7c8e179', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4479921caf5f405b8fc49baad390a0e9', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2325.708228] env[62824]: INFO nova.compute.manager [None req-07530f12-f1bb-49d0-8972-cfc1629bba0b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3] Terminating instance [ 2325.835170] env[62824]: DEBUG oslo_vmware.api [None req-1f9e537a-d5e5-4905-854b-725e8ff9d4b4 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2146836, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.458873} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2325.835508] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-1f9e537a-d5e5-4905-854b-725e8ff9d4b4 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2325.835660] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-1f9e537a-d5e5-4905-854b-725e8ff9d4b4 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: 6d80ec06-8559-4964-8577-a2512aa366ed] Deleted contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2325.835880] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-1f9e537a-d5e5-4905-854b-725e8ff9d4b4 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: 6d80ec06-8559-4964-8577-a2512aa366ed] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2325.836072] env[62824]: INFO nova.compute.manager [None req-1f9e537a-d5e5-4905-854b-725e8ff9d4b4 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: 6d80ec06-8559-4964-8577-a2512aa366ed] Took 1.49 seconds to destroy the instance on the hypervisor. [ 2325.836345] env[62824]: DEBUG oslo.service.loopingcall [None req-1f9e537a-d5e5-4905-854b-725e8ff9d4b4 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2325.837302] env[62824]: DEBUG nova.compute.manager [-] [instance: 6d80ec06-8559-4964-8577-a2512aa366ed] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2325.837302] env[62824]: DEBUG nova.network.neutron [-] [instance: 6d80ec06-8559-4964-8577-a2512aa366ed] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2325.850985] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c9f8cb26-c57c-4b54-b2ce-c8334af2af51 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "c480f680-61fd-496b-8cf7-fa50f580b10d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.676s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2325.922212] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c91314bc-5561-47a4-9134-3c18c1f19036 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2325.935256] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2fbc832-2d63-4634-8449-3a89fd5e9c3b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2325.985398] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe65d54b-398a-46f3-9026-2ad8f55fd3bf {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2325.993733] env[62824]: DEBUG nova.network.neutron [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Successfully created port: 34ec0688-79ff-4219-afc1-498ec608f081 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2326.002285] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f324b7e1-2a40-4a0e-a7a4-e478799e7add {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2326.016611] env[62824]: DEBUG oslo_vmware.api [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146837, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2326.026766] env[62824]: DEBUG nova.compute.provider_tree [None req-9eadcb07-c2f2-4daf-ab0e-537cd63453af tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Updating inventory in ProviderTree for provider bf95157b-a274-42de-9ccf-9851128a44a1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 2326.160148] env[62824]: DEBUG nova.compute.manager [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2326.188883] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Releasing lock "refresh_cache-b3003c4b-ae5a-48df-8c12-a915a76253f4" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2326.188883] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Updated the network info_cache for instance {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10516}} [ 2326.189461] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2326.189461] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2326.189461] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2326.189743] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62824) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11064}} [ 2326.189743] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2326.189814] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Cleaning up deleted instances {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11745}} [ 2326.212353] env[62824]: DEBUG nova.compute.manager [None req-07530f12-f1bb-49d0-8972-cfc1629bba0b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2326.213138] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-07530f12-f1bb-49d0-8972-cfc1629bba0b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2326.214045] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c1a345b-cfd0-4dd7-a4b1-4976cdb3c7ce {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2326.232923] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-07530f12-f1bb-49d0-8972-cfc1629bba0b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2326.233455] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ffdf120c-f960-45d6-b939-ef811cb1bbbb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2326.245035] env[62824]: DEBUG oslo_vmware.api [None req-07530f12-f1bb-49d0-8972-cfc1629bba0b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2326.245035] env[62824]: value = "task-2146838" [ 2326.245035] env[62824]: _type = "Task" [ 2326.245035] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2326.256176] env[62824]: DEBUG oslo_vmware.api [None req-07530f12-f1bb-49d0-8972-cfc1629bba0b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146838, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2326.513544] env[62824]: DEBUG nova.network.neutron [req-7df4e318-0490-459b-a6e5-da7f5d715943 req-ce12c3bc-78aa-4005-9e33-c25b0426eedd service nova] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Updated VIF entry in instance network info cache for port 3c1f3341-20a4-4b77-8dae-12ef5901a16e. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2326.515396] env[62824]: DEBUG nova.network.neutron [req-7df4e318-0490-459b-a6e5-da7f5d715943 req-ce12c3bc-78aa-4005-9e33-c25b0426eedd service nova] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Updating instance_info_cache with network_info: [{"id": "3c1f3341-20a4-4b77-8dae-12ef5901a16e", "address": "fa:16:3e:1c:3a:8c", "network": {"id": "aa1c09e5-8d72-43ad-a903-a0f5e711da80", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1643616572-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c987c5f844be42c99cc9d57a9ca84c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9bc2632-36f9-4912-8782-8bbb789f909d", "external-id": "nsx-vlan-transportzone-897", "segmentation_id": 897, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3c1f3341-20", "ovs_interfaceid": "3c1f3341-20a4-4b77-8dae-12ef5901a16e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2326.523369] env[62824]: DEBUG oslo_vmware.api [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146837, 'name': Destroy_Task} progress is 100%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2326.528084] env[62824]: DEBUG nova.network.neutron [req-dbf65db9-17eb-464f-b0ff-d538924db10e req-0f59977f-1fd3-4301-b6e1-0d5cb81fff7b service nova] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Updated VIF entry in instance network info cache for port aeeadfdd-d869-4555-bda2-391c435de1bb. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2326.528493] env[62824]: DEBUG nova.network.neutron [req-dbf65db9-17eb-464f-b0ff-d538924db10e req-0f59977f-1fd3-4301-b6e1-0d5cb81fff7b service nova] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Updating instance_info_cache with network_info: [{"id": "aeeadfdd-d869-4555-bda2-391c435de1bb", "address": "fa:16:3e:43:2e:10", "network": {"id": "bba9b98c-c7eb-4890-8de5-eafacfdebd7d", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1928394728-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.176", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7fcdbfc29df64a1ba8d982bdcc667b64", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b49c5024-2ced-42ca-90cc-6066766d43e6", "external-id": "nsx-vlan-transportzone-239", "segmentation_id": 239, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaeeadfdd-d8", "ovs_interfaceid": "aeeadfdd-d869-4555-bda2-391c435de1bb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2326.560031] env[62824]: ERROR nova.scheduler.client.report [None req-9eadcb07-c2f2-4daf-ab0e-537cd63453af tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [req-21eb895b-1e34-4ee7-af56-01aa59b7dfc2] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID bf95157b-a274-42de-9ccf-9851128a44a1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-21eb895b-1e34-4ee7-af56-01aa59b7dfc2"}]} [ 2326.580916] env[62824]: DEBUG nova.scheduler.client.report [None req-9eadcb07-c2f2-4daf-ab0e-537cd63453af tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Refreshing inventories for resource provider bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 2326.596259] env[62824]: DEBUG nova.scheduler.client.report [None req-9eadcb07-c2f2-4daf-ab0e-537cd63453af tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Updating ProviderTree inventory for provider bf95157b-a274-42de-9ccf-9851128a44a1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 2326.596602] env[62824]: DEBUG nova.compute.provider_tree [None req-9eadcb07-c2f2-4daf-ab0e-537cd63453af tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Updating inventory in ProviderTree for provider bf95157b-a274-42de-9ccf-9851128a44a1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 2326.610374] env[62824]: DEBUG nova.scheduler.client.report [None req-9eadcb07-c2f2-4daf-ab0e-537cd63453af tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Refreshing aggregate associations for resource provider bf95157b-a274-42de-9ccf-9851128a44a1, aggregates: None {{(pid=62824) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 2326.631016] env[62824]: DEBUG nova.scheduler.client.report [None req-9eadcb07-c2f2-4daf-ab0e-537cd63453af tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Refreshing trait associations for resource provider bf95157b-a274-42de-9ccf-9851128a44a1, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_ISO {{(pid=62824) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 2326.706716] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] There are 59 instances to clean {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11754}} [ 2326.706908] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 19bcead9-b27c-471a-b217-cf7bfe101ed8] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2326.755381] env[62824]: DEBUG oslo_vmware.api [None req-07530f12-f1bb-49d0-8972-cfc1629bba0b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146838, 'name': PowerOffVM_Task, 'duration_secs': 0.359958} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2326.757996] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-07530f12-f1bb-49d0-8972-cfc1629bba0b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2326.758210] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-07530f12-f1bb-49d0-8972-cfc1629bba0b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2326.758651] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b1ca1162-969a-4466-b55e-7193f83f1026 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2326.865041] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4179ab2b-749d-4671-8c96-c5983bc0f2db {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2326.877753] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c3e67b6-2656-4257-af15-ea97d450251d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2326.908192] env[62824]: DEBUG nova.network.neutron [-] [instance: 6d80ec06-8559-4964-8577-a2512aa366ed] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2326.910595] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48d79c42-c5e2-4d96-9b9d-6a7483ff45fc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2326.919996] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d084218d-b836-4741-a70f-7cffc1fcd491 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2326.935731] env[62824]: DEBUG nova.compute.provider_tree [None req-9eadcb07-c2f2-4daf-ab0e-537cd63453af tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Updating inventory in ProviderTree for provider bf95157b-a274-42de-9ccf-9851128a44a1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 2327.021473] env[62824]: DEBUG oslo_concurrency.lockutils [req-7df4e318-0490-459b-a6e5-da7f5d715943 req-ce12c3bc-78aa-4005-9e33-c25b0426eedd service nova] Releasing lock "refresh_cache-511ef5bb-1dca-4eae-a0f1-19d1d7a17371" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2327.021853] env[62824]: DEBUG nova.compute.manager [req-7df4e318-0490-459b-a6e5-da7f5d715943 req-ce12c3bc-78aa-4005-9e33-c25b0426eedd service nova] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Received event network-changed-3c1f3341-20a4-4b77-8dae-12ef5901a16e {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2327.022104] env[62824]: DEBUG nova.compute.manager [req-7df4e318-0490-459b-a6e5-da7f5d715943 req-ce12c3bc-78aa-4005-9e33-c25b0426eedd service nova] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Refreshing instance network info cache due to event network-changed-3c1f3341-20a4-4b77-8dae-12ef5901a16e. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2327.022429] env[62824]: DEBUG oslo_concurrency.lockutils [req-7df4e318-0490-459b-a6e5-da7f5d715943 req-ce12c3bc-78aa-4005-9e33-c25b0426eedd service nova] Acquiring lock "refresh_cache-511ef5bb-1dca-4eae-a0f1-19d1d7a17371" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2327.022622] env[62824]: DEBUG oslo_concurrency.lockutils [req-7df4e318-0490-459b-a6e5-da7f5d715943 req-ce12c3bc-78aa-4005-9e33-c25b0426eedd service nova] Acquired lock "refresh_cache-511ef5bb-1dca-4eae-a0f1-19d1d7a17371" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2327.022846] env[62824]: DEBUG nova.network.neutron [req-7df4e318-0490-459b-a6e5-da7f5d715943 req-ce12c3bc-78aa-4005-9e33-c25b0426eedd service nova] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Refreshing network info cache for port 3c1f3341-20a4-4b77-8dae-12ef5901a16e {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2327.024378] env[62824]: DEBUG oslo_vmware.api [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146837, 'name': Destroy_Task} progress is 100%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2327.034891] env[62824]: DEBUG oslo_concurrency.lockutils [req-dbf65db9-17eb-464f-b0ff-d538924db10e req-0f59977f-1fd3-4301-b6e1-0d5cb81fff7b service nova] Releasing lock "refresh_cache-1db9ff97-e59c-43f6-984e-6405d8e993a4" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2327.170086] env[62824]: DEBUG nova.compute.manager [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2327.196860] env[62824]: DEBUG nova.virt.hardware [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2327.197117] env[62824]: DEBUG nova.virt.hardware [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2327.197279] env[62824]: DEBUG nova.virt.hardware [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2327.197465] env[62824]: DEBUG nova.virt.hardware [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2327.197613] env[62824]: DEBUG nova.virt.hardware [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2327.197848] env[62824]: DEBUG nova.virt.hardware [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2327.198276] env[62824]: DEBUG nova.virt.hardware [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2327.198551] env[62824]: DEBUG nova.virt.hardware [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2327.198821] env[62824]: DEBUG nova.virt.hardware [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2327.199298] env[62824]: DEBUG nova.virt.hardware [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2327.199533] env[62824]: DEBUG nova.virt.hardware [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2327.200492] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bb68470-9db7-40ab-8a4b-f1ba3808d927 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2327.209213] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c6e48c7-9247-42df-bd3a-67460e0446d6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2327.213501] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 71731b21-c302-4d66-a579-889165673712] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2327.233476] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-07530f12-f1bb-49d0-8972-cfc1629bba0b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2327.233765] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-07530f12-f1bb-49d0-8972-cfc1629bba0b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3] Deleting contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2327.233887] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-07530f12-f1bb-49d0-8972-cfc1629bba0b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Deleting the datastore file [datastore1] 68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2327.234162] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8d48a20f-2026-49df-9cda-1e51125c9d16 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2327.241668] env[62824]: DEBUG oslo_vmware.api [None req-07530f12-f1bb-49d0-8972-cfc1629bba0b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2327.241668] env[62824]: value = "task-2146840" [ 2327.241668] env[62824]: _type = "Task" [ 2327.241668] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2327.251046] env[62824]: DEBUG oslo_vmware.api [None req-07530f12-f1bb-49d0-8972-cfc1629bba0b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146840, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2327.382101] env[62824]: DEBUG nova.compute.manager [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Stashing vm_state: active {{(pid=62824) _prep_resize /opt/stack/nova/nova/compute/manager.py:6136}} [ 2327.414779] env[62824]: INFO nova.compute.manager [-] [instance: 6d80ec06-8559-4964-8577-a2512aa366ed] Took 1.58 seconds to deallocate network for instance. [ 2327.426186] env[62824]: DEBUG nova.compute.manager [req-45270fc0-5b78-4c8d-91de-2dabdad21ee9 req-9a54cd81-8039-4fc2-94cf-79086fb27445 service nova] [instance: 6d80ec06-8559-4964-8577-a2512aa366ed] Received event network-vif-deleted-24e47f2d-ba7e-48db-9b98-c4f3f6a78930 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2327.457561] env[62824]: ERROR nova.scheduler.client.report [None req-9eadcb07-c2f2-4daf-ab0e-537cd63453af tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [req-99080436-dc6e-47c9-b8f3-ee427a6d5457] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID bf95157b-a274-42de-9ccf-9851128a44a1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-99080436-dc6e-47c9-b8f3-ee427a6d5457"}]} [ 2327.475224] env[62824]: DEBUG nova.scheduler.client.report [None req-9eadcb07-c2f2-4daf-ab0e-537cd63453af tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Refreshing inventories for resource provider bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 2327.490940] env[62824]: DEBUG nova.scheduler.client.report [None req-9eadcb07-c2f2-4daf-ab0e-537cd63453af tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Updating ProviderTree inventory for provider bf95157b-a274-42de-9ccf-9851128a44a1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 2327.491184] env[62824]: DEBUG nova.compute.provider_tree [None req-9eadcb07-c2f2-4daf-ab0e-537cd63453af tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Updating inventory in ProviderTree for provider bf95157b-a274-42de-9ccf-9851128a44a1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 2327.503502] env[62824]: DEBUG nova.scheduler.client.report [None req-9eadcb07-c2f2-4daf-ab0e-537cd63453af tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Refreshing aggregate associations for resource provider bf95157b-a274-42de-9ccf-9851128a44a1, aggregates: None {{(pid=62824) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 2327.519837] env[62824]: DEBUG oslo_vmware.api [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146837, 'name': Destroy_Task, 'duration_secs': 1.946598} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2327.519995] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Destroyed the VM [ 2327.520148] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Deleting Snapshot of the VM instance {{(pid=62824) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 2327.520442] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-a577d00c-e45d-47d2-b7f9-99bbad795665 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2327.524564] env[62824]: DEBUG nova.scheduler.client.report [None req-9eadcb07-c2f2-4daf-ab0e-537cd63453af tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Refreshing trait associations for resource provider bf95157b-a274-42de-9ccf-9851128a44a1, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_ISO {{(pid=62824) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 2327.531129] env[62824]: DEBUG oslo_vmware.api [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Waiting for the task: (returnval){ [ 2327.531129] env[62824]: value = "task-2146841" [ 2327.531129] env[62824]: _type = "Task" [ 2327.531129] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2327.541863] env[62824]: DEBUG oslo_vmware.api [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146841, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2327.719358] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 815bd7ae-f461-4c74-9401-7c5fe43679ee] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2327.757992] env[62824]: DEBUG oslo_vmware.api [None req-07530f12-f1bb-49d0-8972-cfc1629bba0b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146840, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.200638} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2327.759729] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-07530f12-f1bb-49d0-8972-cfc1629bba0b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2327.760446] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-07530f12-f1bb-49d0-8972-cfc1629bba0b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3] Deleted contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2327.760657] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-07530f12-f1bb-49d0-8972-cfc1629bba0b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2327.761088] env[62824]: INFO nova.compute.manager [None req-07530f12-f1bb-49d0-8972-cfc1629bba0b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: 68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3] Took 1.55 seconds to destroy the instance on the hypervisor. [ 2327.761509] env[62824]: DEBUG oslo.service.loopingcall [None req-07530f12-f1bb-49d0-8972-cfc1629bba0b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2327.762831] env[62824]: DEBUG nova.compute.manager [req-ea40ff20-4db0-41f2-96fb-06dd4a5f8f68 req-0bf60e99-adb5-41ea-ba95-5d28c8b4e2b4 service nova] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Received event network-vif-plugged-34ec0688-79ff-4219-afc1-498ec608f081 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2327.763121] env[62824]: DEBUG oslo_concurrency.lockutils [req-ea40ff20-4db0-41f2-96fb-06dd4a5f8f68 req-0bf60e99-adb5-41ea-ba95-5d28c8b4e2b4 service nova] Acquiring lock "bf6e681d-b773-4787-886a-c90586bc81bf-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2327.763342] env[62824]: DEBUG oslo_concurrency.lockutils [req-ea40ff20-4db0-41f2-96fb-06dd4a5f8f68 req-0bf60e99-adb5-41ea-ba95-5d28c8b4e2b4 service nova] Lock "bf6e681d-b773-4787-886a-c90586bc81bf-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2327.763513] env[62824]: DEBUG oslo_concurrency.lockutils [req-ea40ff20-4db0-41f2-96fb-06dd4a5f8f68 req-0bf60e99-adb5-41ea-ba95-5d28c8b4e2b4 service nova] Lock "bf6e681d-b773-4787-886a-c90586bc81bf-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2327.763678] env[62824]: DEBUG nova.compute.manager [req-ea40ff20-4db0-41f2-96fb-06dd4a5f8f68 req-0bf60e99-adb5-41ea-ba95-5d28c8b4e2b4 service nova] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] No waiting events found dispatching network-vif-plugged-34ec0688-79ff-4219-afc1-498ec608f081 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2327.763839] env[62824]: WARNING nova.compute.manager [req-ea40ff20-4db0-41f2-96fb-06dd4a5f8f68 req-0bf60e99-adb5-41ea-ba95-5d28c8b4e2b4 service nova] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Received unexpected event network-vif-plugged-34ec0688-79ff-4219-afc1-498ec608f081 for instance with vm_state building and task_state spawning. [ 2327.766929] env[62824]: DEBUG nova.compute.manager [-] [instance: 68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2327.767034] env[62824]: DEBUG nova.network.neutron [-] [instance: 68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2327.788898] env[62824]: DEBUG nova.network.neutron [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Successfully updated port: 34ec0688-79ff-4219-afc1-498ec608f081 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2327.801086] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a8b8a83-baa3-4405-8f1f-2f7e877bb661 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2327.810679] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c46ad075-89a1-4204-bca4-e855ec1b1956 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2327.850110] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cadffac6-8e6e-43b7-b6b4-5ae017cb7352 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2327.858809] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47c66bb9-6aec-4edf-8840-fd68de2ccf35 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2327.873747] env[62824]: DEBUG nova.compute.provider_tree [None req-9eadcb07-c2f2-4daf-ab0e-537cd63453af tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Updating inventory in ProviderTree for provider bf95157b-a274-42de-9ccf-9851128a44a1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 2327.890570] env[62824]: DEBUG nova.network.neutron [req-7df4e318-0490-459b-a6e5-da7f5d715943 req-ce12c3bc-78aa-4005-9e33-c25b0426eedd service nova] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Updated VIF entry in instance network info cache for port 3c1f3341-20a4-4b77-8dae-12ef5901a16e. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2327.890913] env[62824]: DEBUG nova.network.neutron [req-7df4e318-0490-459b-a6e5-da7f5d715943 req-ce12c3bc-78aa-4005-9e33-c25b0426eedd service nova] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Updating instance_info_cache with network_info: [{"id": "3c1f3341-20a4-4b77-8dae-12ef5901a16e", "address": "fa:16:3e:1c:3a:8c", "network": {"id": "aa1c09e5-8d72-43ad-a903-a0f5e711da80", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1643616572-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c987c5f844be42c99cc9d57a9ca84c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9bc2632-36f9-4912-8782-8bbb789f909d", "external-id": "nsx-vlan-transportzone-897", "segmentation_id": 897, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3c1f3341-20", "ovs_interfaceid": "3c1f3341-20a4-4b77-8dae-12ef5901a16e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2327.899106] env[62824]: DEBUG oslo_concurrency.lockutils [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2327.928771] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1f9e537a-d5e5-4905-854b-725e8ff9d4b4 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2328.041850] env[62824]: DEBUG oslo_vmware.api [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146841, 'name': RemoveSnapshot_Task, 'duration_secs': 0.367758} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2328.041850] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Deleted Snapshot of the VM instance {{(pid=62824) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 2328.042179] env[62824]: DEBUG nova.compute.manager [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2328.043434] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-defef451-ee10-44b1-8de8-c053cf7f318a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2328.222455] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 3acdf86e-45e9-4353-a282-7272451389e7] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2328.294107] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquiring lock "refresh_cache-bf6e681d-b773-4787-886a-c90586bc81bf" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2328.294274] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquired lock "refresh_cache-bf6e681d-b773-4787-886a-c90586bc81bf" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2328.294433] env[62824]: DEBUG nova.network.neutron [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2328.393750] env[62824]: DEBUG oslo_concurrency.lockutils [req-7df4e318-0490-459b-a6e5-da7f5d715943 req-ce12c3bc-78aa-4005-9e33-c25b0426eedd service nova] Releasing lock "refresh_cache-511ef5bb-1dca-4eae-a0f1-19d1d7a17371" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2328.394141] env[62824]: DEBUG nova.compute.manager [req-7df4e318-0490-459b-a6e5-da7f5d715943 req-ce12c3bc-78aa-4005-9e33-c25b0426eedd service nova] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Received event network-changed-d5c6ba87-e0c7-49c2-bbaf-b7c64f84cdf8 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2328.394390] env[62824]: DEBUG nova.compute.manager [req-7df4e318-0490-459b-a6e5-da7f5d715943 req-ce12c3bc-78aa-4005-9e33-c25b0426eedd service nova] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Refreshing instance network info cache due to event network-changed-d5c6ba87-e0c7-49c2-bbaf-b7c64f84cdf8. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2328.394778] env[62824]: DEBUG oslo_concurrency.lockutils [req-7df4e318-0490-459b-a6e5-da7f5d715943 req-ce12c3bc-78aa-4005-9e33-c25b0426eedd service nova] Acquiring lock "refresh_cache-e1574266-e53b-4169-9db6-c66b84895edb" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2328.395052] env[62824]: DEBUG oslo_concurrency.lockutils [req-7df4e318-0490-459b-a6e5-da7f5d715943 req-ce12c3bc-78aa-4005-9e33-c25b0426eedd service nova] Acquired lock "refresh_cache-e1574266-e53b-4169-9db6-c66b84895edb" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2328.395281] env[62824]: DEBUG nova.network.neutron [req-7df4e318-0490-459b-a6e5-da7f5d715943 req-ce12c3bc-78aa-4005-9e33-c25b0426eedd service nova] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Refreshing network info cache for port d5c6ba87-e0c7-49c2-bbaf-b7c64f84cdf8 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2328.407114] env[62824]: DEBUG nova.scheduler.client.report [None req-9eadcb07-c2f2-4daf-ab0e-537cd63453af tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Updated inventory for provider bf95157b-a274-42de-9ccf-9851128a44a1 with generation 129 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 2328.407376] env[62824]: DEBUG nova.compute.provider_tree [None req-9eadcb07-c2f2-4daf-ab0e-537cd63453af tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Updating resource provider bf95157b-a274-42de-9ccf-9851128a44a1 generation from 129 to 130 during operation: update_inventory {{(pid=62824) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 2328.407566] env[62824]: DEBUG nova.compute.provider_tree [None req-9eadcb07-c2f2-4daf-ab0e-537cd63453af tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Updating inventory in ProviderTree for provider bf95157b-a274-42de-9ccf-9851128a44a1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 2328.530077] env[62824]: DEBUG nova.network.neutron [-] [instance: 68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2328.557028] env[62824]: INFO nova.compute.manager [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Shelve offloading [ 2328.725748] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: c5f0cfcc-5dbf-48fb-af2e-3d3b31c291a9] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2328.828806] env[62824]: DEBUG nova.network.neutron [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2328.912612] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9eadcb07-c2f2-4daf-ab0e-537cd63453af tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.763s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2328.914692] env[62824]: DEBUG oslo_concurrency.lockutils [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 1.016s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2328.937575] env[62824]: INFO nova.scheduler.client.report [None req-9eadcb07-c2f2-4daf-ab0e-537cd63453af tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Deleted allocations for instance 1805a6e7-48f2-47b4-b097-0644ba8217fa [ 2328.967190] env[62824]: DEBUG nova.network.neutron [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Updating instance_info_cache with network_info: [{"id": "34ec0688-79ff-4219-afc1-498ec608f081", "address": "fa:16:3e:6c:40:0f", "network": {"id": "12411632-b7f6-4544-b2e4-b04ae7337547", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-917106830-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4479921caf5f405b8fc49baad390a0e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "de7fa486-5f28-44ae-b0cf-72234ff87546", "external-id": "nsx-vlan-transportzone-229", "segmentation_id": 229, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap34ec0688-79", "ovs_interfaceid": "34ec0688-79ff-4219-afc1-498ec608f081", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2329.033142] env[62824]: INFO nova.compute.manager [-] [instance: 68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3] Took 1.27 seconds to deallocate network for instance. [ 2329.061394] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2329.061584] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6cfabad4-0a6f-459e-994c-880b18fff734 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2329.071062] env[62824]: DEBUG oslo_vmware.api [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Waiting for the task: (returnval){ [ 2329.071062] env[62824]: value = "task-2146842" [ 2329.071062] env[62824]: _type = "Task" [ 2329.071062] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2329.084548] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] VM already powered off {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 2329.084896] env[62824]: DEBUG nova.compute.manager [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2329.085997] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a530f55f-2eb7-483e-8c1a-3602545d1f46 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2329.098016] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Acquiring lock "refresh_cache-71e9733b-c84b-4501-8faf-a487f089b498" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2329.098016] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Acquired lock "refresh_cache-71e9733b-c84b-4501-8faf-a487f089b498" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2329.098201] env[62824]: DEBUG nova.network.neutron [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2329.128881] env[62824]: DEBUG nova.network.neutron [req-7df4e318-0490-459b-a6e5-da7f5d715943 req-ce12c3bc-78aa-4005-9e33-c25b0426eedd service nova] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Updated VIF entry in instance network info cache for port d5c6ba87-e0c7-49c2-bbaf-b7c64f84cdf8. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2329.129076] env[62824]: DEBUG nova.network.neutron [req-7df4e318-0490-459b-a6e5-da7f5d715943 req-ce12c3bc-78aa-4005-9e33-c25b0426eedd service nova] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Updating instance_info_cache with network_info: [{"id": "d5c6ba87-e0c7-49c2-bbaf-b7c64f84cdf8", "address": "fa:16:3e:c4:9c:f4", "network": {"id": "aa1c09e5-8d72-43ad-a903-a0f5e711da80", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1643616572-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.166", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c987c5f844be42c99cc9d57a9ca84c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9bc2632-36f9-4912-8782-8bbb789f909d", "external-id": "nsx-vlan-transportzone-897", "segmentation_id": 897, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd5c6ba87-e0", "ovs_interfaceid": "d5c6ba87-e0c7-49c2-bbaf-b7c64f84cdf8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2329.229579] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 204b866f-e9c5-441d-bf8e-77d324b39ecd] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2329.419964] env[62824]: INFO nova.compute.claims [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2329.444849] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9eadcb07-c2f2-4daf-ab0e-537cd63453af tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "1805a6e7-48f2-47b4-b097-0644ba8217fa" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.348s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2329.470133] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Releasing lock "refresh_cache-bf6e681d-b773-4787-886a-c90586bc81bf" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2329.470471] env[62824]: DEBUG nova.compute.manager [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Instance network_info: |[{"id": "34ec0688-79ff-4219-afc1-498ec608f081", "address": "fa:16:3e:6c:40:0f", "network": {"id": "12411632-b7f6-4544-b2e4-b04ae7337547", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-917106830-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4479921caf5f405b8fc49baad390a0e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "de7fa486-5f28-44ae-b0cf-72234ff87546", "external-id": "nsx-vlan-transportzone-229", "segmentation_id": 229, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap34ec0688-79", "ovs_interfaceid": "34ec0688-79ff-4219-afc1-498ec608f081", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2329.470873] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6c:40:0f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'de7fa486-5f28-44ae-b0cf-72234ff87546', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '34ec0688-79ff-4219-afc1-498ec608f081', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2329.478359] env[62824]: DEBUG oslo.service.loopingcall [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2329.478783] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2329.479014] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-87846d20-4474-40a0-a528-bc1396e207e3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2329.499716] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2329.499716] env[62824]: value = "task-2146843" [ 2329.499716] env[62824]: _type = "Task" [ 2329.499716] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2329.507801] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146843, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2329.539794] env[62824]: DEBUG oslo_concurrency.lockutils [None req-07530f12-f1bb-49d0-8972-cfc1629bba0b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2329.632281] env[62824]: DEBUG oslo_concurrency.lockutils [req-7df4e318-0490-459b-a6e5-da7f5d715943 req-ce12c3bc-78aa-4005-9e33-c25b0426eedd service nova] Releasing lock "refresh_cache-e1574266-e53b-4169-9db6-c66b84895edb" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2329.734055] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 0394d368-c846-4f3b-bfcf-2bc8e858052a] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2329.791548] env[62824]: DEBUG nova.compute.manager [req-6523fcfa-fc0b-4637-8f0e-0bc2b6b64c1f req-41c9d911-8f89-4366-8845-1707ef20f5f0 service nova] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Received event network-changed-34ec0688-79ff-4219-afc1-498ec608f081 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2329.791746] env[62824]: DEBUG nova.compute.manager [req-6523fcfa-fc0b-4637-8f0e-0bc2b6b64c1f req-41c9d911-8f89-4366-8845-1707ef20f5f0 service nova] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Refreshing instance network info cache due to event network-changed-34ec0688-79ff-4219-afc1-498ec608f081. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2329.791984] env[62824]: DEBUG oslo_concurrency.lockutils [req-6523fcfa-fc0b-4637-8f0e-0bc2b6b64c1f req-41c9d911-8f89-4366-8845-1707ef20f5f0 service nova] Acquiring lock "refresh_cache-bf6e681d-b773-4787-886a-c90586bc81bf" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2329.792226] env[62824]: DEBUG oslo_concurrency.lockutils [req-6523fcfa-fc0b-4637-8f0e-0bc2b6b64c1f req-41c9d911-8f89-4366-8845-1707ef20f5f0 service nova] Acquired lock "refresh_cache-bf6e681d-b773-4787-886a-c90586bc81bf" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2329.792447] env[62824]: DEBUG nova.network.neutron [req-6523fcfa-fc0b-4637-8f0e-0bc2b6b64c1f req-41c9d911-8f89-4366-8845-1707ef20f5f0 service nova] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Refreshing network info cache for port 34ec0688-79ff-4219-afc1-498ec608f081 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2329.866212] env[62824]: DEBUG nova.network.neutron [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Updating instance_info_cache with network_info: [{"id": "4214bf6e-29f0-4d92-be44-44fccf7ab149", "address": "fa:16:3e:a6:d8:c4", "network": {"id": "fc5d7128-fe9b-44ef-9d91-839190a97226", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-796405330-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35b7289f4746463e9b1cc47fe914a835", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0954fad3-d24d-496c-83e6-a09d3cb556fc", "external-id": "nsx-vlan-transportzone-216", "segmentation_id": 216, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4214bf6e-29", "ovs_interfaceid": "4214bf6e-29f0-4d92-be44-44fccf7ab149", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2329.925794] env[62824]: INFO nova.compute.resource_tracker [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Updating resource usage from migration e66ebab6-1da6-493b-b7f5-ef121278bf32 [ 2330.011549] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146843, 'name': CreateVM_Task, 'duration_secs': 0.362662} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2330.014194] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2330.015123] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2330.015371] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2330.015701] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2330.016019] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8f1cc871-1e4d-4e1a-8916-f902eabc71a5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2330.026218] env[62824]: DEBUG oslo_vmware.api [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2330.026218] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5284b9e0-0215-aaf3-ccf3-ebc7a4270ea9" [ 2330.026218] env[62824]: _type = "Task" [ 2330.026218] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2330.034776] env[62824]: DEBUG oslo_vmware.api [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5284b9e0-0215-aaf3-ccf3-ebc7a4270ea9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2330.124540] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3421867-0ba2-4afa-a85f-bc0aac1d8def {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2330.132753] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90d11009-1a11-4c9e-a9ff-ffffe92ba730 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2330.162874] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c561c547-3404-456c-9def-56be62bd394c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2330.170475] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a883c8e-f532-4ae1-b7cb-68b3cab2f161 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2330.183542] env[62824]: DEBUG nova.compute.provider_tree [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2330.237705] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: feda7def-7788-4a60-a56a-9353a1475c54] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2330.369363] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Releasing lock "refresh_cache-71e9733b-c84b-4501-8faf-a487f089b498" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2330.503110] env[62824]: DEBUG nova.network.neutron [req-6523fcfa-fc0b-4637-8f0e-0bc2b6b64c1f req-41c9d911-8f89-4366-8845-1707ef20f5f0 service nova] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Updated VIF entry in instance network info cache for port 34ec0688-79ff-4219-afc1-498ec608f081. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2330.503489] env[62824]: DEBUG nova.network.neutron [req-6523fcfa-fc0b-4637-8f0e-0bc2b6b64c1f req-41c9d911-8f89-4366-8845-1707ef20f5f0 service nova] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Updating instance_info_cache with network_info: [{"id": "34ec0688-79ff-4219-afc1-498ec608f081", "address": "fa:16:3e:6c:40:0f", "network": {"id": "12411632-b7f6-4544-b2e4-b04ae7337547", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-917106830-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4479921caf5f405b8fc49baad390a0e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "de7fa486-5f28-44ae-b0cf-72234ff87546", "external-id": "nsx-vlan-transportzone-229", "segmentation_id": 229, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap34ec0688-79", "ovs_interfaceid": "34ec0688-79ff-4219-afc1-498ec608f081", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2330.537431] env[62824]: DEBUG oslo_vmware.api [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5284b9e0-0215-aaf3-ccf3-ebc7a4270ea9, 'name': SearchDatastore_Task, 'duration_secs': 0.010185} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2330.537763] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2330.538022] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2330.538333] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2330.538483] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2330.538746] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2330.538917] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2bdc25e9-a81c-4b3f-a1c6-684c946a7f73 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2330.548087] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2330.548338] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2330.549063] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-af2f13b3-b538-436f-90b4-be7587f25193 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2330.554695] env[62824]: DEBUG oslo_vmware.api [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2330.554695] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]529f73b7-2cd2-491c-eef2-367fbbddfd35" [ 2330.554695] env[62824]: _type = "Task" [ 2330.554695] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2330.563470] env[62824]: DEBUG oslo_vmware.api [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]529f73b7-2cd2-491c-eef2-367fbbddfd35, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2330.645327] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2330.646549] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d115b47-7961-4f94-b43c-eb5cee10f959 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2330.655225] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2330.655479] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ff9d4fd9-a6b9-4ba2-abf7-818153f3ae48 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2330.686551] env[62824]: DEBUG nova.scheduler.client.report [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2330.740410] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: b0cfa2cb-0a3b-409b-8e8b-7adbf6fd746a] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2331.005912] env[62824]: DEBUG oslo_concurrency.lockutils [req-6523fcfa-fc0b-4637-8f0e-0bc2b6b64c1f req-41c9d911-8f89-4366-8845-1707ef20f5f0 service nova] Releasing lock "refresh_cache-bf6e681d-b773-4787-886a-c90586bc81bf" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2331.006242] env[62824]: DEBUG nova.compute.manager [req-6523fcfa-fc0b-4637-8f0e-0bc2b6b64c1f req-41c9d911-8f89-4366-8845-1707ef20f5f0 service nova] [instance: 68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3] Received event network-vif-deleted-2f4a7326-1463-4942-b137-573646230e57 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2331.065807] env[62824]: DEBUG oslo_vmware.api [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]529f73b7-2cd2-491c-eef2-367fbbddfd35, 'name': SearchDatastore_Task, 'duration_secs': 0.010206} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2331.066634] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-91a55ce0-449f-47fd-a1bc-fafee8fa9eee {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2331.072608] env[62824]: DEBUG oslo_vmware.api [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2331.072608] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]526de6f9-844a-9d6f-6ae7-9ad69cdc2b59" [ 2331.072608] env[62824]: _type = "Task" [ 2331.072608] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2331.081251] env[62824]: DEBUG oslo_vmware.api [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]526de6f9-844a-9d6f-6ae7-9ad69cdc2b59, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2331.193141] env[62824]: DEBUG oslo_concurrency.lockutils [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.278s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2331.193380] env[62824]: INFO nova.compute.manager [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Migrating [ 2331.200061] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1f9e537a-d5e5-4905-854b-725e8ff9d4b4 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.271s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2331.200197] env[62824]: DEBUG nova.objects.instance [None req-1f9e537a-d5e5-4905-854b-725e8ff9d4b4 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Lazy-loading 'resources' on Instance uuid 6d80ec06-8559-4964-8577-a2512aa366ed {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2331.243189] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 853f3098-881c-402c-8fd7-26540d1ecd88] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2331.374106] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ba7ba48-4cfd-416a-b960-cbd99686bd15 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2331.382323] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89223be3-9ac2-4633-8f3c-cd692ae21e5d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2331.413733] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3fed701-f347-4e36-aceb-6a2d180cc93a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2331.421333] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40696045-2e6a-4298-83e8-c6a690a9b27b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2331.434593] env[62824]: DEBUG nova.compute.provider_tree [None req-1f9e537a-d5e5-4905-854b-725e8ff9d4b4 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2331.583469] env[62824]: DEBUG oslo_vmware.api [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]526de6f9-844a-9d6f-6ae7-9ad69cdc2b59, 'name': SearchDatastore_Task, 'duration_secs': 0.010259} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2331.583758] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2331.584076] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] bf6e681d-b773-4787-886a-c90586bc81bf/bf6e681d-b773-4787-886a-c90586bc81bf.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2331.584346] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-36cf7747-f39e-4e19-87bf-ee750b568d78 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2331.592243] env[62824]: DEBUG oslo_vmware.api [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2331.592243] env[62824]: value = "task-2146845" [ 2331.592243] env[62824]: _type = "Task" [ 2331.592243] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2331.601669] env[62824]: DEBUG oslo_vmware.api [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146845, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2331.711482] env[62824]: DEBUG oslo_concurrency.lockutils [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquiring lock "refresh_cache-c480f680-61fd-496b-8cf7-fa50f580b10d" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2331.711653] env[62824]: DEBUG oslo_concurrency.lockutils [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquired lock "refresh_cache-c480f680-61fd-496b-8cf7-fa50f580b10d" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2331.711936] env[62824]: DEBUG nova.network.neutron [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2331.746239] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 7c44d956-53d1-4a5f-9016-c0f3f0f348b4] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2331.816262] env[62824]: DEBUG nova.compute.manager [req-e8e8387e-d8c0-4581-85df-30edc1787d00 req-6ac60986-9f2e-46b7-9242-5c5a6d0ba5a9 service nova] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Received event network-vif-unplugged-4214bf6e-29f0-4d92-be44-44fccf7ab149 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2331.816503] env[62824]: DEBUG oslo_concurrency.lockutils [req-e8e8387e-d8c0-4581-85df-30edc1787d00 req-6ac60986-9f2e-46b7-9242-5c5a6d0ba5a9 service nova] Acquiring lock "71e9733b-c84b-4501-8faf-a487f089b498-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2331.816718] env[62824]: DEBUG oslo_concurrency.lockutils [req-e8e8387e-d8c0-4581-85df-30edc1787d00 req-6ac60986-9f2e-46b7-9242-5c5a6d0ba5a9 service nova] Lock "71e9733b-c84b-4501-8faf-a487f089b498-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2331.816911] env[62824]: DEBUG oslo_concurrency.lockutils [req-e8e8387e-d8c0-4581-85df-30edc1787d00 req-6ac60986-9f2e-46b7-9242-5c5a6d0ba5a9 service nova] Lock "71e9733b-c84b-4501-8faf-a487f089b498-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2331.817118] env[62824]: DEBUG nova.compute.manager [req-e8e8387e-d8c0-4581-85df-30edc1787d00 req-6ac60986-9f2e-46b7-9242-5c5a6d0ba5a9 service nova] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] No waiting events found dispatching network-vif-unplugged-4214bf6e-29f0-4d92-be44-44fccf7ab149 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2331.817298] env[62824]: WARNING nova.compute.manager [req-e8e8387e-d8c0-4581-85df-30edc1787d00 req-6ac60986-9f2e-46b7-9242-5c5a6d0ba5a9 service nova] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Received unexpected event network-vif-unplugged-4214bf6e-29f0-4d92-be44-44fccf7ab149 for instance with vm_state shelved and task_state shelving_offloading. [ 2331.817462] env[62824]: DEBUG nova.compute.manager [req-e8e8387e-d8c0-4581-85df-30edc1787d00 req-6ac60986-9f2e-46b7-9242-5c5a6d0ba5a9 service nova] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Received event network-changed-4214bf6e-29f0-4d92-be44-44fccf7ab149 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2331.817616] env[62824]: DEBUG nova.compute.manager [req-e8e8387e-d8c0-4581-85df-30edc1787d00 req-6ac60986-9f2e-46b7-9242-5c5a6d0ba5a9 service nova] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Refreshing instance network info cache due to event network-changed-4214bf6e-29f0-4d92-be44-44fccf7ab149. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2331.817799] env[62824]: DEBUG oslo_concurrency.lockutils [req-e8e8387e-d8c0-4581-85df-30edc1787d00 req-6ac60986-9f2e-46b7-9242-5c5a6d0ba5a9 service nova] Acquiring lock "refresh_cache-71e9733b-c84b-4501-8faf-a487f089b498" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2331.817939] env[62824]: DEBUG oslo_concurrency.lockutils [req-e8e8387e-d8c0-4581-85df-30edc1787d00 req-6ac60986-9f2e-46b7-9242-5c5a6d0ba5a9 service nova] Acquired lock "refresh_cache-71e9733b-c84b-4501-8faf-a487f089b498" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2331.818102] env[62824]: DEBUG nova.network.neutron [req-e8e8387e-d8c0-4581-85df-30edc1787d00 req-6ac60986-9f2e-46b7-9242-5c5a6d0ba5a9 service nova] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Refreshing network info cache for port 4214bf6e-29f0-4d92-be44-44fccf7ab149 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2331.938528] env[62824]: DEBUG nova.scheduler.client.report [None req-1f9e537a-d5e5-4905-854b-725e8ff9d4b4 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2332.105066] env[62824]: DEBUG oslo_vmware.api [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146845, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2332.249522] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 1f162a63-8773-4c34-9c84-b2de1d5e9f50] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2332.423488] env[62824]: DEBUG nova.network.neutron [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Updating instance_info_cache with network_info: [{"id": "9190a574-76bc-4236-9f09-69d08897d634", "address": "fa:16:3e:a9:57:8f", "network": {"id": "b995934c-4934-4469-884c-73cb4de1796f", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1892307918-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c11aef7fb6247cb9b5272a6a063cd12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15165046-2de9-4ada-9e99-0126e20854a9", "external-id": "nsx-vlan-transportzone-974", "segmentation_id": 974, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9190a574-76", "ovs_interfaceid": "9190a574-76bc-4236-9f09-69d08897d634", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2332.443642] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1f9e537a-d5e5-4905-854b-725e8ff9d4b4 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.244s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2332.446464] env[62824]: DEBUG oslo_concurrency.lockutils [None req-07530f12-f1bb-49d0-8972-cfc1629bba0b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.907s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2332.446719] env[62824]: DEBUG nova.objects.instance [None req-07530f12-f1bb-49d0-8972-cfc1629bba0b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lazy-loading 'resources' on Instance uuid 68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2332.466601] env[62824]: INFO nova.scheduler.client.report [None req-1f9e537a-d5e5-4905-854b-725e8ff9d4b4 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Deleted allocations for instance 6d80ec06-8559-4964-8577-a2512aa366ed [ 2332.527758] env[62824]: DEBUG nova.network.neutron [req-e8e8387e-d8c0-4581-85df-30edc1787d00 req-6ac60986-9f2e-46b7-9242-5c5a6d0ba5a9 service nova] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Updated VIF entry in instance network info cache for port 4214bf6e-29f0-4d92-be44-44fccf7ab149. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2332.528117] env[62824]: DEBUG nova.network.neutron [req-e8e8387e-d8c0-4581-85df-30edc1787d00 req-6ac60986-9f2e-46b7-9242-5c5a6d0ba5a9 service nova] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Updating instance_info_cache with network_info: [{"id": "4214bf6e-29f0-4d92-be44-44fccf7ab149", "address": "fa:16:3e:a6:d8:c4", "network": {"id": "fc5d7128-fe9b-44ef-9d91-839190a97226", "bridge": null, "label": "tempest-ServersNegativeTestJSON-796405330-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35b7289f4746463e9b1cc47fe914a835", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap4214bf6e-29", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2332.605081] env[62824]: DEBUG oslo_vmware.api [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146845, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.718008} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2332.605363] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] bf6e681d-b773-4787-886a-c90586bc81bf/bf6e681d-b773-4787-886a-c90586bc81bf.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2332.605582] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2332.605843] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2193a740-8314-41f1-9e15-aade536ec573 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2332.614415] env[62824]: DEBUG oslo_vmware.api [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2332.614415] env[62824]: value = "task-2146846" [ 2332.614415] env[62824]: _type = "Task" [ 2332.614415] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2332.624264] env[62824]: DEBUG oslo_vmware.api [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146846, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2332.753563] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 0c9fb5cc-de69-46d8-9962-98e0a84e33c3] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2332.926759] env[62824]: DEBUG oslo_concurrency.lockutils [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Releasing lock "refresh_cache-c480f680-61fd-496b-8cf7-fa50f580b10d" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2332.974616] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1f9e537a-d5e5-4905-854b-725e8ff9d4b4 tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Lock "6d80ec06-8559-4964-8577-a2512aa366ed" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.139s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2333.031157] env[62824]: DEBUG oslo_concurrency.lockutils [req-e8e8387e-d8c0-4581-85df-30edc1787d00 req-6ac60986-9f2e-46b7-9242-5c5a6d0ba5a9 service nova] Releasing lock "refresh_cache-71e9733b-c84b-4501-8faf-a487f089b498" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2333.125720] env[62824]: DEBUG oslo_vmware.api [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146846, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.081786} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2333.126947] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2333.127915] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ba45889-54e2-436a-bb3a-368a79db9a8a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2333.130736] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-521f54b6-2a1b-4f7a-9ba7-34de886b6bb6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2333.154689] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Reconfiguring VM instance instance-0000006b to attach disk [datastore2] bf6e681d-b773-4787-886a-c90586bc81bf/bf6e681d-b773-4787-886a-c90586bc81bf.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2333.157202] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a91cffab-8727-4c5c-8548-922c1982677d tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Acquiring lock "b3003c4b-ae5a-48df-8c12-a915a76253f4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2333.157351] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a91cffab-8727-4c5c-8548-922c1982677d tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Lock "b3003c4b-ae5a-48df-8c12-a915a76253f4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2333.157531] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a91cffab-8727-4c5c-8548-922c1982677d tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Acquiring lock "b3003c4b-ae5a-48df-8c12-a915a76253f4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2333.157779] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a91cffab-8727-4c5c-8548-922c1982677d tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Lock "b3003c4b-ae5a-48df-8c12-a915a76253f4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2333.157925] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a91cffab-8727-4c5c-8548-922c1982677d tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Lock "b3003c4b-ae5a-48df-8c12-a915a76253f4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2333.159395] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c22894c6-500d-445f-aff8-5f895badf2be {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2333.173938] env[62824]: INFO nova.compute.manager [None req-a91cffab-8727-4c5c-8548-922c1982677d tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Terminating instance [ 2333.176416] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be19d86f-e999-4fa6-9532-3f40f6cb8b0e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2333.211408] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47325957-b751-4b0c-81c0-edf85f50aa7f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2333.214265] env[62824]: DEBUG oslo_vmware.api [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2333.214265] env[62824]: value = "task-2146847" [ 2333.214265] env[62824]: _type = "Task" [ 2333.214265] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2333.222028] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e29018f-21c3-41ab-8d63-dc505c487b09 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2333.229217] env[62824]: DEBUG oslo_vmware.api [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146847, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2333.239595] env[62824]: DEBUG nova.compute.provider_tree [None req-07530f12-f1bb-49d0-8972-cfc1629bba0b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Updating inventory in ProviderTree for provider bf95157b-a274-42de-9ccf-9851128a44a1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 2333.256458] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 8c2ac679-6915-4149-aa51-99170b0df685] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2333.682470] env[62824]: DEBUG nova.compute.manager [None req-a91cffab-8727-4c5c-8548-922c1982677d tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2333.682735] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-a91cffab-8727-4c5c-8548-922c1982677d tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2333.684138] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00226a12-5201-4b05-8f72-f77b70776f4f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2333.692791] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-a91cffab-8727-4c5c-8548-922c1982677d tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2333.693098] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-785307af-2287-4e47-b0e9-d97852573898 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2333.702812] env[62824]: DEBUG oslo_vmware.api [None req-a91cffab-8727-4c5c-8548-922c1982677d tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for the task: (returnval){ [ 2333.702812] env[62824]: value = "task-2146848" [ 2333.702812] env[62824]: _type = "Task" [ 2333.702812] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2333.711787] env[62824]: DEBUG oslo_vmware.api [None req-a91cffab-8727-4c5c-8548-922c1982677d tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2146848, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2333.726053] env[62824]: DEBUG oslo_vmware.api [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146847, 'name': ReconfigVM_Task, 'duration_secs': 0.535692} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2333.726344] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Reconfigured VM instance instance-0000006b to attach disk [datastore2] bf6e681d-b773-4787-886a-c90586bc81bf/bf6e681d-b773-4787-886a-c90586bc81bf.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2333.726991] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8096adf6-828a-4bc4-8851-b94f1a258a64 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2333.734051] env[62824]: DEBUG oslo_vmware.api [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2333.734051] env[62824]: value = "task-2146849" [ 2333.734051] env[62824]: _type = "Task" [ 2333.734051] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2333.746543] env[62824]: DEBUG oslo_vmware.api [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146849, 'name': Rename_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2333.759447] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: da46d1d5-f75f-4ef1-b571-fbebab89c2a6] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2333.775683] env[62824]: DEBUG nova.scheduler.client.report [None req-07530f12-f1bb-49d0-8972-cfc1629bba0b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Updated inventory for provider bf95157b-a274-42de-9ccf-9851128a44a1 with generation 130 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 2333.775952] env[62824]: DEBUG nova.compute.provider_tree [None req-07530f12-f1bb-49d0-8972-cfc1629bba0b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Updating resource provider bf95157b-a274-42de-9ccf-9851128a44a1 generation from 130 to 131 during operation: update_inventory {{(pid=62824) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 2333.776151] env[62824]: DEBUG nova.compute.provider_tree [None req-07530f12-f1bb-49d0-8972-cfc1629bba0b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Updating inventory in ProviderTree for provider bf95157b-a274-42de-9ccf-9851128a44a1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 2334.213583] env[62824]: DEBUG oslo_vmware.api [None req-a91cffab-8727-4c5c-8548-922c1982677d tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2146848, 'name': PowerOffVM_Task, 'duration_secs': 0.417412} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2334.214033] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-a91cffab-8727-4c5c-8548-922c1982677d tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2334.214033] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-a91cffab-8727-4c5c-8548-922c1982677d tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2334.214276] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b0a4a52f-9716-400b-880d-8f90c52c5b89 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2334.244863] env[62824]: DEBUG oslo_vmware.api [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146849, 'name': Rename_Task, 'duration_secs': 0.374044} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2334.245157] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2334.245441] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-caaa2729-d052-4c81-aff4-259d29831980 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2334.252026] env[62824]: DEBUG oslo_vmware.api [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2334.252026] env[62824]: value = "task-2146851" [ 2334.252026] env[62824]: _type = "Task" [ 2334.252026] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2334.260653] env[62824]: DEBUG oslo_vmware.api [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146851, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2334.263290] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: da1ad739-b252-4e29-a22a-ab3bcab173ec] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2334.282047] env[62824]: DEBUG oslo_concurrency.lockutils [None req-07530f12-f1bb-49d0-8972-cfc1629bba0b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.835s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2334.303463] env[62824]: INFO nova.scheduler.client.report [None req-07530f12-f1bb-49d0-8972-cfc1629bba0b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Deleted allocations for instance 68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3 [ 2334.441892] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5830fc0-c503-4917-b223-91833e99d34d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2334.461107] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Updating instance 'c480f680-61fd-496b-8cf7-fa50f580b10d' progress to 0 {{(pid=62824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 2334.762983] env[62824]: DEBUG oslo_vmware.api [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146851, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2334.766510] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: ab37b1a2-0012-40fb-9341-b613525e89cf] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2334.810763] env[62824]: DEBUG oslo_concurrency.lockutils [None req-07530f12-f1bb-49d0-8972-cfc1629bba0b tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.109s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2334.967824] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2334.968154] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a5cac335-62e1-4f2f-b273-e14f94687b98 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2334.977070] env[62824]: DEBUG oslo_vmware.api [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 2334.977070] env[62824]: value = "task-2146852" [ 2334.977070] env[62824]: _type = "Task" [ 2334.977070] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2334.985791] env[62824]: DEBUG oslo_vmware.api [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146852, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2335.264661] env[62824]: DEBUG oslo_vmware.api [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146851, 'name': PowerOnVM_Task, 'duration_secs': 0.622263} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2335.265048] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2335.265378] env[62824]: INFO nova.compute.manager [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Took 8.09 seconds to spawn the instance on the hypervisor. [ 2335.265704] env[62824]: DEBUG nova.compute.manager [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2335.266665] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76706952-9d6e-482a-9832-8f3e5694798b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2335.270906] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 2afd1f18-234c-40b5-9ede-7413ad30dafe] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2335.487875] env[62824]: DEBUG oslo_vmware.api [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146852, 'name': PowerOffVM_Task, 'duration_secs': 0.224111} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2335.488166] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2335.488374] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Updating instance 'c480f680-61fd-496b-8cf7-fa50f580b10d' progress to 17 {{(pid=62824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 2335.774790] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: bd7f2ed5-fa08-4a48-9755-60b61791ea0b] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2335.788042] env[62824]: INFO nova.compute.manager [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Took 13.10 seconds to build instance. [ 2335.995468] env[62824]: DEBUG nova.virt.hardware [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:24Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2335.995677] env[62824]: DEBUG nova.virt.hardware [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2335.995840] env[62824]: DEBUG nova.virt.hardware [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2335.996043] env[62824]: DEBUG nova.virt.hardware [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2335.996200] env[62824]: DEBUG nova.virt.hardware [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2335.996351] env[62824]: DEBUG nova.virt.hardware [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2335.996564] env[62824]: DEBUG nova.virt.hardware [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2335.996729] env[62824]: DEBUG nova.virt.hardware [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2335.996935] env[62824]: DEBUG nova.virt.hardware [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2335.997124] env[62824]: DEBUG nova.virt.hardware [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2335.997321] env[62824]: DEBUG nova.virt.hardware [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2336.002762] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b2507d29-a021-49c6-8665-515a9dd05300 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2336.020712] env[62824]: DEBUG oslo_vmware.api [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 2336.020712] env[62824]: value = "task-2146853" [ 2336.020712] env[62824]: _type = "Task" [ 2336.020712] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2336.034783] env[62824]: DEBUG oslo_vmware.api [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146853, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2336.202038] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquiring lock "fb9cfb53-285a-4586-acf0-cb08d56b93e7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2336.202243] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "fb9cfb53-285a-4586-acf0-cb08d56b93e7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2336.278662] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: fc77c44a-180c-46ff-9690-9072c6213c91] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2336.290478] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3786e68f-45ff-4a60-967d-f25bd263e444 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "bf6e681d-b773-4787-886a-c90586bc81bf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.614s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2336.534753] env[62824]: DEBUG oslo_vmware.api [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146853, 'name': ReconfigVM_Task, 'duration_secs': 0.291575} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2336.535175] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Updating instance 'c480f680-61fd-496b-8cf7-fa50f580b10d' progress to 33 {{(pid=62824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 2336.590717] env[62824]: INFO nova.compute.manager [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Rebuilding instance [ 2336.645796] env[62824]: DEBUG nova.compute.manager [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2336.647063] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-875b6dd6-fd02-46a4-80ce-2e73ebbd03b3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2336.705086] env[62824]: DEBUG nova.compute.manager [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: fb9cfb53-285a-4586-acf0-cb08d56b93e7] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2336.781706] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 20f93b46-5e7e-4256-8cc1-e0a1b16740d5] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2337.043882] env[62824]: DEBUG nova.virt.hardware [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2337.044156] env[62824]: DEBUG nova.virt.hardware [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2337.044319] env[62824]: DEBUG nova.virt.hardware [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2337.044566] env[62824]: DEBUG nova.virt.hardware [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2337.044749] env[62824]: DEBUG nova.virt.hardware [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2337.044904] env[62824]: DEBUG nova.virt.hardware [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2337.045128] env[62824]: DEBUG nova.virt.hardware [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2337.045292] env[62824]: DEBUG nova.virt.hardware [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2337.045468] env[62824]: DEBUG nova.virt.hardware [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2337.045637] env[62824]: DEBUG nova.virt.hardware [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2337.045813] env[62824]: DEBUG nova.virt.hardware [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2337.051521] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Reconfiguring VM instance instance-0000006a to detach disk 2000 {{(pid=62824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 2337.051811] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-824835a3-8b30-4f7f-b7a4-58fd39582ab6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2337.072557] env[62824]: DEBUG oslo_vmware.api [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 2337.072557] env[62824]: value = "task-2146854" [ 2337.072557] env[62824]: _type = "Task" [ 2337.072557] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2337.081442] env[62824]: DEBUG oslo_vmware.api [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146854, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2337.225882] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2337.226168] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2337.227770] env[62824]: INFO nova.compute.claims [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: fb9cfb53-285a-4586-acf0-cb08d56b93e7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2337.284403] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 7a2a51c4-e558-4f5f-b82c-718bc12c1df5] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2337.365893] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41d0194a-d6d3-4444-b80f-9b2f95a1efd4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquiring lock "interface-e1574266-e53b-4169-9db6-c66b84895edb-afcd2bdb-2bfc-40d7-a8e5-2a417be0e0e9" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2337.366157] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41d0194a-d6d3-4444-b80f-9b2f95a1efd4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lock "interface-e1574266-e53b-4169-9db6-c66b84895edb-afcd2bdb-2bfc-40d7-a8e5-2a417be0e0e9" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2337.366552] env[62824]: DEBUG nova.objects.instance [None req-41d0194a-d6d3-4444-b80f-9b2f95a1efd4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lazy-loading 'flavor' on Instance uuid e1574266-e53b-4169-9db6-c66b84895edb {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2337.584153] env[62824]: DEBUG oslo_vmware.api [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146854, 'name': ReconfigVM_Task, 'duration_secs': 0.248772} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2337.584393] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Reconfigured VM instance instance-0000006a to detach disk 2000 {{(pid=62824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 2337.585252] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b38c761e-2074-41e9-b5a3-6186bb2a4df5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2337.609342] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] c480f680-61fd-496b-8cf7-fa50f580b10d/c480f680-61fd-496b-8cf7-fa50f580b10d.vmdk or device None with type thin {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2337.609674] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b8d3fd15-a1e2-4a0c-99f9-105ed7791552 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2337.630844] env[62824]: DEBUG oslo_vmware.api [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 2337.630844] env[62824]: value = "task-2146855" [ 2337.630844] env[62824]: _type = "Task" [ 2337.630844] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2337.640918] env[62824]: DEBUG oslo_vmware.api [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146855, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2337.661792] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2337.662207] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e9da862f-5758-429a-b2ca-38a2aa32b499 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2337.671412] env[62824]: DEBUG oslo_vmware.api [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2337.671412] env[62824]: value = "task-2146856" [ 2337.671412] env[62824]: _type = "Task" [ 2337.671412] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2337.681556] env[62824]: DEBUG oslo_vmware.api [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146856, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2337.788340] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: dadf590a-7288-4dd2-90de-125fe272f4ba] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2338.027523] env[62824]: DEBUG nova.objects.instance [None req-41d0194a-d6d3-4444-b80f-9b2f95a1efd4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lazy-loading 'pci_requests' on Instance uuid e1574266-e53b-4169-9db6-c66b84895edb {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2338.142084] env[62824]: DEBUG oslo_vmware.api [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146855, 'name': ReconfigVM_Task, 'duration_secs': 0.299116} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2338.142411] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Reconfigured VM instance instance-0000006a to attach disk [datastore1] c480f680-61fd-496b-8cf7-fa50f580b10d/c480f680-61fd-496b-8cf7-fa50f580b10d.vmdk or device None with type thin {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2338.142786] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Updating instance 'c480f680-61fd-496b-8cf7-fa50f580b10d' progress to 50 {{(pid=62824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 2338.184739] env[62824]: DEBUG oslo_vmware.api [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146856, 'name': PowerOffVM_Task, 'duration_secs': 0.20823} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2338.185821] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2338.186020] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2338.186814] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdd7d514-f7c0-4597-a9d8-c8065cfce13c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2338.195845] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2338.196113] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-25d6eefb-8121-4d69-91cb-9208373611aa {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2338.293200] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: bc11c3c0-69c2-4b30-ac6f-3d21ad6ef49e] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2338.442342] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56d08459-7370-4237-96dd-1a0374714600 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2338.451112] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6c64f18-046d-4808-bdfb-115dada8406e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2338.481807] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-929c1237-8a3c-4eaa-8a6a-c9072ed8cd33 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2338.490351] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82cc0f5b-0c5b-4158-bc9d-7944ee169690 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2338.506807] env[62824]: DEBUG nova.compute.provider_tree [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2338.530473] env[62824]: DEBUG nova.objects.base [None req-41d0194a-d6d3-4444-b80f-9b2f95a1efd4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=62824) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 2338.530703] env[62824]: DEBUG nova.network.neutron [None req-41d0194a-d6d3-4444-b80f-9b2f95a1efd4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: e1574266-e53b-4169-9db6-c66b84895edb] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2338.630528] env[62824]: DEBUG nova.policy [None req-41d0194a-d6d3-4444-b80f-9b2f95a1efd4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b989eb89498b45dab6604116870d2fa1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c987c5f844be42c99cc9d57a9ca84c31', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2338.649593] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d18e416-d11c-4fd7-b2f9-55be073e04e6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2338.671337] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-000a8766-9348-494f-aacb-8d012e434f4b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2338.690884] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Updating instance 'c480f680-61fd-496b-8cf7-fa50f580b10d' progress to 67 {{(pid=62824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 2338.796524] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: d067fa97-cedc-4e3d-9be4-d860a79a7723] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2339.011571] env[62824]: DEBUG nova.scheduler.client.report [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2339.232600] env[62824]: DEBUG nova.network.neutron [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Port 9190a574-76bc-4236-9f09-69d08897d634 binding to destination host cpu-1 is already ACTIVE {{(pid=62824) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 2339.299837] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 4c2c2068-664d-404f-a99e-8fc7719f43e5] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2339.516196] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.290s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2339.516736] env[62824]: DEBUG nova.compute.manager [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: fb9cfb53-285a-4586-acf0-cb08d56b93e7] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2339.802983] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 29140542-dc4b-411a-ac10-7d84086eabbb] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2340.021511] env[62824]: DEBUG nova.compute.utils [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2340.022958] env[62824]: DEBUG nova.compute.manager [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: fb9cfb53-285a-4586-acf0-cb08d56b93e7] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2340.023141] env[62824]: DEBUG nova.network.neutron [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: fb9cfb53-285a-4586-acf0-cb08d56b93e7] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2340.073239] env[62824]: DEBUG nova.policy [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '959d127a7d144b33a0cae94db5c11846', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dfff08982dad4790bf4d555e2b4db5e4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2340.200278] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-a91cffab-8727-4c5c-8548-922c1982677d tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2340.200809] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-a91cffab-8727-4c5c-8548-922c1982677d tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Deleting contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2340.200809] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-a91cffab-8727-4c5c-8548-922c1982677d tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Deleting the datastore file [datastore2] b3003c4b-ae5a-48df-8c12-a915a76253f4 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2340.201743] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b0d067ff-05d2-4a61-a9f9-b2ff06e13d48 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2340.203904] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2340.204188] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Deleting contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2340.204294] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Deleting the datastore file [datastore2] bf6e681d-b773-4787-886a-c90586bc81bf {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2340.204594] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f855af33-8255-4a50-b94c-c9a52f13e4cb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2340.213359] env[62824]: DEBUG oslo_vmware.api [None req-a91cffab-8727-4c5c-8548-922c1982677d tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for the task: (returnval){ [ 2340.213359] env[62824]: value = "task-2146858" [ 2340.213359] env[62824]: _type = "Task" [ 2340.213359] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2340.214718] env[62824]: DEBUG oslo_vmware.api [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2340.214718] env[62824]: value = "task-2146859" [ 2340.214718] env[62824]: _type = "Task" [ 2340.214718] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2340.228509] env[62824]: DEBUG oslo_vmware.api [None req-a91cffab-8727-4c5c-8548-922c1982677d tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2146858, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2340.228715] env[62824]: DEBUG oslo_vmware.api [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146859, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2340.256601] env[62824]: DEBUG oslo_concurrency.lockutils [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquiring lock "c480f680-61fd-496b-8cf7-fa50f580b10d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2340.256842] env[62824]: DEBUG oslo_concurrency.lockutils [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "c480f680-61fd-496b-8cf7-fa50f580b10d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2340.257030] env[62824]: DEBUG oslo_concurrency.lockutils [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "c480f680-61fd-496b-8cf7-fa50f580b10d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2340.258929] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2340.259136] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Deleting contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2340.259319] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Deleting the datastore file [datastore2] 71e9733b-c84b-4501-8faf-a487f089b498 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2340.259722] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0c7592fc-1a7d-4eed-afb3-9e8f1e407b5c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2340.268181] env[62824]: DEBUG oslo_vmware.api [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Waiting for the task: (returnval){ [ 2340.268181] env[62824]: value = "task-2146860" [ 2340.268181] env[62824]: _type = "Task" [ 2340.268181] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2340.278615] env[62824]: DEBUG oslo_vmware.api [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146860, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2340.305729] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 49818910-61de-4fb8-bbab-d5d61d2f1ada] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2340.331675] env[62824]: DEBUG nova.network.neutron [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: fb9cfb53-285a-4586-acf0-cb08d56b93e7] Successfully created port: 5278435d-72ae-4fec-844c-0f1174154e30 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2340.526847] env[62824]: DEBUG nova.compute.manager [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: fb9cfb53-285a-4586-acf0-cb08d56b93e7] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2340.726541] env[62824]: DEBUG oslo_vmware.api [None req-a91cffab-8727-4c5c-8548-922c1982677d tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Task: {'id': task-2146858, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.155185} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2340.729480] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-a91cffab-8727-4c5c-8548-922c1982677d tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2340.729679] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-a91cffab-8727-4c5c-8548-922c1982677d tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Deleted contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2340.729858] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-a91cffab-8727-4c5c-8548-922c1982677d tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2340.730043] env[62824]: INFO nova.compute.manager [None req-a91cffab-8727-4c5c-8548-922c1982677d tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Took 7.05 seconds to destroy the instance on the hypervisor. [ 2340.730286] env[62824]: DEBUG oslo.service.loopingcall [None req-a91cffab-8727-4c5c-8548-922c1982677d tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2340.730522] env[62824]: DEBUG oslo_vmware.api [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146859, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.200465} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2340.730723] env[62824]: DEBUG nova.compute.manager [-] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2340.730820] env[62824]: DEBUG nova.network.neutron [-] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2340.732401] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2340.732582] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Deleted contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2340.732761] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2340.780589] env[62824]: DEBUG oslo_vmware.api [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146860, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.170304} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2340.780849] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2340.781092] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Deleted contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2340.781286] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2340.807507] env[62824]: INFO nova.scheduler.client.report [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Deleted allocations for instance 71e9733b-c84b-4501-8faf-a487f089b498 [ 2340.810271] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 363177c0-dc40-429a-a74b-e690da133edb] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2341.119409] env[62824]: DEBUG nova.compute.manager [req-d4d67296-146a-4fe9-9272-9482714242f8 req-2ff18ac0-8c79-4f1a-8d42-07326884fecc service nova] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Received event network-vif-deleted-4e0d51df-4419-4059-9dab-be44957e4f77 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2341.119725] env[62824]: INFO nova.compute.manager [req-d4d67296-146a-4fe9-9272-9482714242f8 req-2ff18ac0-8c79-4f1a-8d42-07326884fecc service nova] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Neutron deleted interface 4e0d51df-4419-4059-9dab-be44957e4f77; detaching it from the instance and deleting it from the info cache [ 2341.120016] env[62824]: DEBUG nova.network.neutron [req-d4d67296-146a-4fe9-9272-9482714242f8 req-2ff18ac0-8c79-4f1a-8d42-07326884fecc service nova] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2341.289670] env[62824]: DEBUG oslo_concurrency.lockutils [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquiring lock "refresh_cache-c480f680-61fd-496b-8cf7-fa50f580b10d" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2341.289879] env[62824]: DEBUG oslo_concurrency.lockutils [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquired lock "refresh_cache-c480f680-61fd-496b-8cf7-fa50f580b10d" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2341.290129] env[62824]: DEBUG nova.network.neutron [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2341.314643] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2341.314921] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2341.315827] env[62824]: DEBUG nova.objects.instance [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Lazy-loading 'resources' on Instance uuid 71e9733b-c84b-4501-8faf-a487f089b498 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2341.316199] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 350cfd12-a8d5-4a5f-b3de-d795fa179dfd] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2341.479033] env[62824]: DEBUG nova.network.neutron [-] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2341.536126] env[62824]: DEBUG nova.compute.manager [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: fb9cfb53-285a-4586-acf0-cb08d56b93e7] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2341.558766] env[62824]: DEBUG nova.virt.hardware [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2341.559013] env[62824]: DEBUG nova.virt.hardware [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2341.559182] env[62824]: DEBUG nova.virt.hardware [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2341.559369] env[62824]: DEBUG nova.virt.hardware [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2341.559517] env[62824]: DEBUG nova.virt.hardware [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2341.559669] env[62824]: DEBUG nova.virt.hardware [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2341.559877] env[62824]: DEBUG nova.virt.hardware [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2341.560047] env[62824]: DEBUG nova.virt.hardware [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2341.560222] env[62824]: DEBUG nova.virt.hardware [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2341.560387] env[62824]: DEBUG nova.virt.hardware [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2341.560606] env[62824]: DEBUG nova.virt.hardware [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2341.561638] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63723189-5c50-4bb1-890d-efe4080804e1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2341.571208] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c2bf4d1-200c-417a-bc91-b4961c08823a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2341.622697] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e4f869c2-8032-4921-9e6e-beb15fffd160 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2341.633951] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efc73862-8e74-452e-8169-f457b5c6e3ff {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2341.668018] env[62824]: DEBUG nova.compute.manager [req-d4d67296-146a-4fe9-9272-9482714242f8 req-2ff18ac0-8c79-4f1a-8d42-07326884fecc service nova] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Detach interface failed, port_id=4e0d51df-4419-4059-9dab-be44957e4f77, reason: Instance b3003c4b-ae5a-48df-8c12-a915a76253f4 could not be found. {{(pid=62824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11482}} [ 2341.766842] env[62824]: DEBUG nova.virt.hardware [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2341.766929] env[62824]: DEBUG nova.virt.hardware [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2341.767045] env[62824]: DEBUG nova.virt.hardware [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2341.767241] env[62824]: DEBUG nova.virt.hardware [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2341.767391] env[62824]: DEBUG nova.virt.hardware [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2341.767542] env[62824]: DEBUG nova.virt.hardware [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2341.767750] env[62824]: DEBUG nova.virt.hardware [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2341.767911] env[62824]: DEBUG nova.virt.hardware [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2341.768143] env[62824]: DEBUG nova.virt.hardware [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2341.768336] env[62824]: DEBUG nova.virt.hardware [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2341.768520] env[62824]: DEBUG nova.virt.hardware [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2341.769428] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ce6d537-04bd-4ad3-9197-c686fc0dbbe1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2341.778523] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5c9a2d4-1d76-4cee-bb6b-eef8c975c424 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2341.795999] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6c:40:0f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'de7fa486-5f28-44ae-b0cf-72234ff87546', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '34ec0688-79ff-4219-afc1-498ec608f081', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2341.804242] env[62824]: DEBUG oslo.service.loopingcall [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2341.804505] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2341.805055] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f801c2ac-a555-412e-8ac6-1b8ab51a2e5d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2341.820320] env[62824]: DEBUG nova.objects.instance [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Lazy-loading 'numa_topology' on Instance uuid 71e9733b-c84b-4501-8faf-a487f089b498 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2341.821560] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 1870b82a-9783-44ac-8de2-7b9ffc2a1bc8] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2341.833762] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2341.833762] env[62824]: value = "task-2146861" [ 2341.833762] env[62824]: _type = "Task" [ 2341.833762] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2341.845450] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146861, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2341.909068] env[62824]: DEBUG nova.compute.manager [req-8f2e4588-fdab-48de-83df-90789224e8e3 req-a7d8b316-aeb7-4e76-ad59-f0e053322c6e service nova] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Received event network-vif-plugged-afcd2bdb-2bfc-40d7-a8e5-2a417be0e0e9 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2341.909300] env[62824]: DEBUG oslo_concurrency.lockutils [req-8f2e4588-fdab-48de-83df-90789224e8e3 req-a7d8b316-aeb7-4e76-ad59-f0e053322c6e service nova] Acquiring lock "e1574266-e53b-4169-9db6-c66b84895edb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2341.909748] env[62824]: DEBUG oslo_concurrency.lockutils [req-8f2e4588-fdab-48de-83df-90789224e8e3 req-a7d8b316-aeb7-4e76-ad59-f0e053322c6e service nova] Lock "e1574266-e53b-4169-9db6-c66b84895edb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2341.909968] env[62824]: DEBUG oslo_concurrency.lockutils [req-8f2e4588-fdab-48de-83df-90789224e8e3 req-a7d8b316-aeb7-4e76-ad59-f0e053322c6e service nova] Lock "e1574266-e53b-4169-9db6-c66b84895edb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2341.910160] env[62824]: DEBUG nova.compute.manager [req-8f2e4588-fdab-48de-83df-90789224e8e3 req-a7d8b316-aeb7-4e76-ad59-f0e053322c6e service nova] [instance: e1574266-e53b-4169-9db6-c66b84895edb] No waiting events found dispatching network-vif-plugged-afcd2bdb-2bfc-40d7-a8e5-2a417be0e0e9 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2341.910354] env[62824]: WARNING nova.compute.manager [req-8f2e4588-fdab-48de-83df-90789224e8e3 req-a7d8b316-aeb7-4e76-ad59-f0e053322c6e service nova] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Received unexpected event network-vif-plugged-afcd2bdb-2bfc-40d7-a8e5-2a417be0e0e9 for instance with vm_state active and task_state None. [ 2341.935910] env[62824]: DEBUG nova.network.neutron [None req-41d0194a-d6d3-4444-b80f-9b2f95a1efd4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Successfully updated port: afcd2bdb-2bfc-40d7-a8e5-2a417be0e0e9 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2341.942619] env[62824]: DEBUG nova.network.neutron [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: fb9cfb53-285a-4586-acf0-cb08d56b93e7] Successfully updated port: 5278435d-72ae-4fec-844c-0f1174154e30 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2341.981872] env[62824]: INFO nova.compute.manager [-] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Took 1.25 seconds to deallocate network for instance. [ 2342.058840] env[62824]: DEBUG nova.network.neutron [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Updating instance_info_cache with network_info: [{"id": "9190a574-76bc-4236-9f09-69d08897d634", "address": "fa:16:3e:a9:57:8f", "network": {"id": "b995934c-4934-4469-884c-73cb4de1796f", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1892307918-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c11aef7fb6247cb9b5272a6a063cd12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15165046-2de9-4ada-9e99-0126e20854a9", "external-id": "nsx-vlan-transportzone-974", "segmentation_id": 974, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9190a574-76", "ovs_interfaceid": "9190a574-76bc-4236-9f09-69d08897d634", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2342.324379] env[62824]: DEBUG nova.objects.base [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Object Instance<71e9733b-c84b-4501-8faf-a487f089b498> lazy-loaded attributes: resources,numa_topology {{(pid=62824) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 2342.326925] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 02f34565-6675-4c79-ac47-b131ceba9df8] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2342.344596] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146861, 'name': CreateVM_Task, 'duration_secs': 0.344983} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2342.347302] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2342.348489] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2342.348685] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2342.349012] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2342.349278] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9ba0f388-c821-4aa9-9141-4838dfccb1ac {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2342.356617] env[62824]: DEBUG oslo_vmware.api [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2342.356617] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5213b684-ad47-df86-7026-fb8b3f26150d" [ 2342.356617] env[62824]: _type = "Task" [ 2342.356617] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2342.369499] env[62824]: DEBUG oslo_vmware.api [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5213b684-ad47-df86-7026-fb8b3f26150d, 'name': SearchDatastore_Task, 'duration_secs': 0.010653} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2342.369793] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2342.370039] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2342.370297] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2342.370476] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2342.370764] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2342.371058] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9bfa0df3-9849-4a3e-a02e-9c2c739bd507 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2342.382891] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2342.383090] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2342.383838] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d341e936-c4cc-43e9-a7a8-5c955c8d2e91 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2342.392025] env[62824]: DEBUG oslo_vmware.api [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2342.392025] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52f30a18-ab99-2160-cc4f-52f392c3ba9a" [ 2342.392025] env[62824]: _type = "Task" [ 2342.392025] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2342.403651] env[62824]: DEBUG oslo_vmware.api [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52f30a18-ab99-2160-cc4f-52f392c3ba9a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2342.438611] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41d0194a-d6d3-4444-b80f-9b2f95a1efd4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquiring lock "refresh_cache-e1574266-e53b-4169-9db6-c66b84895edb" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2342.438796] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41d0194a-d6d3-4444-b80f-9b2f95a1efd4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquired lock "refresh_cache-e1574266-e53b-4169-9db6-c66b84895edb" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2342.438970] env[62824]: DEBUG nova.network.neutron [None req-41d0194a-d6d3-4444-b80f-9b2f95a1efd4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2342.447375] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquiring lock "refresh_cache-fb9cfb53-285a-4586-acf0-cb08d56b93e7" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2342.447520] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquired lock "refresh_cache-fb9cfb53-285a-4586-acf0-cb08d56b93e7" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2342.447650] env[62824]: DEBUG nova.network.neutron [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: fb9cfb53-285a-4586-acf0-cb08d56b93e7] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2342.488316] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a91cffab-8727-4c5c-8548-922c1982677d tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2342.511996] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7c48a24-c59d-471d-98ef-d971ac45eae3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2342.520334] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50c76891-858b-42d2-a2f6-150148136795 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2342.550822] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f09b8ec2-ab90-42fb-9f8b-5e968ce1dfd3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2342.559113] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80737cff-132f-4518-93cc-a8d497149e2a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2342.563304] env[62824]: DEBUG oslo_concurrency.lockutils [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Releasing lock "refresh_cache-c480f680-61fd-496b-8cf7-fa50f580b10d" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2342.577140] env[62824]: DEBUG nova.compute.provider_tree [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Updating inventory in ProviderTree for provider bf95157b-a274-42de-9ccf-9851128a44a1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 2342.688765] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Acquiring lock "71e9733b-c84b-4501-8faf-a487f089b498" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2342.831718] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 15479062-af75-4925-99b3-77d6a49751ad] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2342.903330] env[62824]: DEBUG oslo_vmware.api [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52f30a18-ab99-2160-cc4f-52f392c3ba9a, 'name': SearchDatastore_Task, 'duration_secs': 0.008884} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2342.904177] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2029f83e-0802-4a39-b9be-b08471d7a804 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2342.909553] env[62824]: DEBUG oslo_vmware.api [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2342.909553] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52ce1957-4d67-dddb-a0ee-7b837c0762ec" [ 2342.909553] env[62824]: _type = "Task" [ 2342.909553] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2342.917390] env[62824]: DEBUG oslo_vmware.api [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52ce1957-4d67-dddb-a0ee-7b837c0762ec, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2342.987641] env[62824]: WARNING nova.network.neutron [None req-41d0194a-d6d3-4444-b80f-9b2f95a1efd4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: e1574266-e53b-4169-9db6-c66b84895edb] aa1c09e5-8d72-43ad-a903-a0f5e711da80 already exists in list: networks containing: ['aa1c09e5-8d72-43ad-a903-a0f5e711da80']. ignoring it [ 2342.997659] env[62824]: DEBUG nova.network.neutron [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: fb9cfb53-285a-4586-acf0-cb08d56b93e7] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2343.083852] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5df0a24b-eb83-43ca-99e7-ca1c31732c0e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2343.108305] env[62824]: ERROR nova.scheduler.client.report [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [req-c59a584d-2fd7-4d8c-996a-b7de140ee15c] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID bf95157b-a274-42de-9ccf-9851128a44a1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-c59a584d-2fd7-4d8c-996a-b7de140ee15c"}]} [ 2343.109347] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7c79190-665b-4358-b3d7-2eec9262ee49 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2343.119706] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Updating instance 'c480f680-61fd-496b-8cf7-fa50f580b10d' progress to 83 {{(pid=62824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 2343.132855] env[62824]: DEBUG nova.scheduler.client.report [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Refreshing inventories for resource provider bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 2343.151179] env[62824]: DEBUG nova.scheduler.client.report [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Updating ProviderTree inventory for provider bf95157b-a274-42de-9ccf-9851128a44a1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 2343.151179] env[62824]: DEBUG nova.compute.provider_tree [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Updating inventory in ProviderTree for provider bf95157b-a274-42de-9ccf-9851128a44a1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 2343.155255] env[62824]: DEBUG nova.compute.manager [req-755b0534-63bf-4b4e-99eb-9fd530de87f6 req-a58e1424-1757-4bd2-b4f6-effe823f03a8 service nova] [instance: fb9cfb53-285a-4586-acf0-cb08d56b93e7] Received event network-vif-plugged-5278435d-72ae-4fec-844c-0f1174154e30 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2343.155474] env[62824]: DEBUG oslo_concurrency.lockutils [req-755b0534-63bf-4b4e-99eb-9fd530de87f6 req-a58e1424-1757-4bd2-b4f6-effe823f03a8 service nova] Acquiring lock "fb9cfb53-285a-4586-acf0-cb08d56b93e7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2343.155679] env[62824]: DEBUG oslo_concurrency.lockutils [req-755b0534-63bf-4b4e-99eb-9fd530de87f6 req-a58e1424-1757-4bd2-b4f6-effe823f03a8 service nova] Lock "fb9cfb53-285a-4586-acf0-cb08d56b93e7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2343.155847] env[62824]: DEBUG oslo_concurrency.lockutils [req-755b0534-63bf-4b4e-99eb-9fd530de87f6 req-a58e1424-1757-4bd2-b4f6-effe823f03a8 service nova] Lock "fb9cfb53-285a-4586-acf0-cb08d56b93e7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2343.156033] env[62824]: DEBUG nova.compute.manager [req-755b0534-63bf-4b4e-99eb-9fd530de87f6 req-a58e1424-1757-4bd2-b4f6-effe823f03a8 service nova] [instance: fb9cfb53-285a-4586-acf0-cb08d56b93e7] No waiting events found dispatching network-vif-plugged-5278435d-72ae-4fec-844c-0f1174154e30 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2343.156249] env[62824]: WARNING nova.compute.manager [req-755b0534-63bf-4b4e-99eb-9fd530de87f6 req-a58e1424-1757-4bd2-b4f6-effe823f03a8 service nova] [instance: fb9cfb53-285a-4586-acf0-cb08d56b93e7] Received unexpected event network-vif-plugged-5278435d-72ae-4fec-844c-0f1174154e30 for instance with vm_state building and task_state spawning. [ 2343.156416] env[62824]: DEBUG nova.compute.manager [req-755b0534-63bf-4b4e-99eb-9fd530de87f6 req-a58e1424-1757-4bd2-b4f6-effe823f03a8 service nova] [instance: fb9cfb53-285a-4586-acf0-cb08d56b93e7] Received event network-changed-5278435d-72ae-4fec-844c-0f1174154e30 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2343.156761] env[62824]: DEBUG nova.compute.manager [req-755b0534-63bf-4b4e-99eb-9fd530de87f6 req-a58e1424-1757-4bd2-b4f6-effe823f03a8 service nova] [instance: fb9cfb53-285a-4586-acf0-cb08d56b93e7] Refreshing instance network info cache due to event network-changed-5278435d-72ae-4fec-844c-0f1174154e30. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2343.156956] env[62824]: DEBUG oslo_concurrency.lockutils [req-755b0534-63bf-4b4e-99eb-9fd530de87f6 req-a58e1424-1757-4bd2-b4f6-effe823f03a8 service nova] Acquiring lock "refresh_cache-fb9cfb53-285a-4586-acf0-cb08d56b93e7" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2343.162785] env[62824]: DEBUG nova.scheduler.client.report [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Refreshing aggregate associations for resource provider bf95157b-a274-42de-9ccf-9851128a44a1, aggregates: None {{(pid=62824) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 2343.180944] env[62824]: DEBUG nova.scheduler.client.report [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Refreshing trait associations for resource provider bf95157b-a274-42de-9ccf-9851128a44a1, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_ISO {{(pid=62824) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 2343.220663] env[62824]: DEBUG nova.network.neutron [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: fb9cfb53-285a-4586-acf0-cb08d56b93e7] Updating instance_info_cache with network_info: [{"id": "5278435d-72ae-4fec-844c-0f1174154e30", "address": "fa:16:3e:24:05:d7", "network": {"id": "0f512108-670f-4a85-9381-108e18b3b5ed", "bridge": "br-int", "label": "tempest-ServersTestJSON-766397782-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dfff08982dad4790bf4d555e2b4db5e4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f3a2eb5-353f-45c5-a73b-869626f4bb13", "external-id": "nsx-vlan-transportzone-411", "segmentation_id": 411, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5278435d-72", "ovs_interfaceid": "5278435d-72ae-4fec-844c-0f1174154e30", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2343.333360] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 919e8a45-7810-4a8d-a1aa-5046b5ab059c] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2343.349296] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e78b5d4d-56a0-478f-bffe-4385a81db18f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2343.357627] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ead27ad0-bf82-470c-902d-dba064dcc606 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2343.390818] env[62824]: DEBUG nova.network.neutron [None req-41d0194a-d6d3-4444-b80f-9b2f95a1efd4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Updating instance_info_cache with network_info: [{"id": "d5c6ba87-e0c7-49c2-bbaf-b7c64f84cdf8", "address": "fa:16:3e:c4:9c:f4", "network": {"id": "aa1c09e5-8d72-43ad-a903-a0f5e711da80", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1643616572-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.166", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c987c5f844be42c99cc9d57a9ca84c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9bc2632-36f9-4912-8782-8bbb789f909d", "external-id": "nsx-vlan-transportzone-897", "segmentation_id": 897, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd5c6ba87-e0", "ovs_interfaceid": "d5c6ba87-e0c7-49c2-bbaf-b7c64f84cdf8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "afcd2bdb-2bfc-40d7-a8e5-2a417be0e0e9", "address": "fa:16:3e:06:1e:c0", "network": {"id": "aa1c09e5-8d72-43ad-a903-a0f5e711da80", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1643616572-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c987c5f844be42c99cc9d57a9ca84c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9bc2632-36f9-4912-8782-8bbb789f909d", "external-id": "nsx-vlan-transportzone-897", "segmentation_id": 897, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapafcd2bdb-2b", "ovs_interfaceid": "afcd2bdb-2bfc-40d7-a8e5-2a417be0e0e9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2343.393622] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04746dd0-8824-4d70-abe3-0b1bfdf3f43f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2343.402169] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45d540c3-d0c9-47d3-aead-3c0a5921a12d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2343.419723] env[62824]: DEBUG nova.compute.provider_tree [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Updating inventory in ProviderTree for provider bf95157b-a274-42de-9ccf-9851128a44a1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 2343.431379] env[62824]: DEBUG oslo_vmware.api [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52ce1957-4d67-dddb-a0ee-7b837c0762ec, 'name': SearchDatastore_Task, 'duration_secs': 0.010154} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2343.431508] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2343.431705] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] bf6e681d-b773-4787-886a-c90586bc81bf/bf6e681d-b773-4787-886a-c90586bc81bf.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2343.431955] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-614c6e28-9cff-4137-baab-e4dbfabd8843 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2343.440301] env[62824]: DEBUG oslo_vmware.api [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2343.440301] env[62824]: value = "task-2146862" [ 2343.440301] env[62824]: _type = "Task" [ 2343.440301] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2343.448924] env[62824]: DEBUG oslo_vmware.api [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146862, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2343.627093] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2343.627370] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3ce20343-9608-480c-b355-0daa97ce266f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2343.636458] env[62824]: DEBUG oslo_vmware.api [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 2343.636458] env[62824]: value = "task-2146863" [ 2343.636458] env[62824]: _type = "Task" [ 2343.636458] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2343.647241] env[62824]: DEBUG oslo_vmware.api [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146863, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2343.723427] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Releasing lock "refresh_cache-fb9cfb53-285a-4586-acf0-cb08d56b93e7" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2343.723843] env[62824]: DEBUG nova.compute.manager [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: fb9cfb53-285a-4586-acf0-cb08d56b93e7] Instance network_info: |[{"id": "5278435d-72ae-4fec-844c-0f1174154e30", "address": "fa:16:3e:24:05:d7", "network": {"id": "0f512108-670f-4a85-9381-108e18b3b5ed", "bridge": "br-int", "label": "tempest-ServersTestJSON-766397782-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dfff08982dad4790bf4d555e2b4db5e4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f3a2eb5-353f-45c5-a73b-869626f4bb13", "external-id": "nsx-vlan-transportzone-411", "segmentation_id": 411, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5278435d-72", "ovs_interfaceid": "5278435d-72ae-4fec-844c-0f1174154e30", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2343.724276] env[62824]: DEBUG oslo_concurrency.lockutils [req-755b0534-63bf-4b4e-99eb-9fd530de87f6 req-a58e1424-1757-4bd2-b4f6-effe823f03a8 service nova] Acquired lock "refresh_cache-fb9cfb53-285a-4586-acf0-cb08d56b93e7" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2343.724490] env[62824]: DEBUG nova.network.neutron [req-755b0534-63bf-4b4e-99eb-9fd530de87f6 req-a58e1424-1757-4bd2-b4f6-effe823f03a8 service nova] [instance: fb9cfb53-285a-4586-acf0-cb08d56b93e7] Refreshing network info cache for port 5278435d-72ae-4fec-844c-0f1174154e30 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2343.726030] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: fb9cfb53-285a-4586-acf0-cb08d56b93e7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:24:05:d7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9f3a2eb5-353f-45c5-a73b-869626f4bb13', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5278435d-72ae-4fec-844c-0f1174154e30', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2343.735419] env[62824]: DEBUG oslo.service.loopingcall [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2343.739321] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fb9cfb53-285a-4586-acf0-cb08d56b93e7] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2343.739957] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a26294c4-2f64-4cfc-8a62-263c9a1a0d0a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2343.763379] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2343.763379] env[62824]: value = "task-2146864" [ 2343.763379] env[62824]: _type = "Task" [ 2343.763379] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2343.773072] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146864, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2343.838535] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 2f315f9e-94e4-47ef-9503-ee92b59e5452] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2343.893769] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41d0194a-d6d3-4444-b80f-9b2f95a1efd4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Releasing lock "refresh_cache-e1574266-e53b-4169-9db6-c66b84895edb" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2343.894576] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41d0194a-d6d3-4444-b80f-9b2f95a1efd4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquiring lock "e1574266-e53b-4169-9db6-c66b84895edb" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2343.894764] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41d0194a-d6d3-4444-b80f-9b2f95a1efd4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquired lock "e1574266-e53b-4169-9db6-c66b84895edb" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2343.895864] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4633209b-b1c5-45ff-82fa-f6159c2f62f3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2343.914157] env[62824]: DEBUG nova.virt.hardware [None req-41d0194a-d6d3-4444-b80f-9b2f95a1efd4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2343.914408] env[62824]: DEBUG nova.virt.hardware [None req-41d0194a-d6d3-4444-b80f-9b2f95a1efd4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2343.914571] env[62824]: DEBUG nova.virt.hardware [None req-41d0194a-d6d3-4444-b80f-9b2f95a1efd4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2343.914758] env[62824]: DEBUG nova.virt.hardware [None req-41d0194a-d6d3-4444-b80f-9b2f95a1efd4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2343.914907] env[62824]: DEBUG nova.virt.hardware [None req-41d0194a-d6d3-4444-b80f-9b2f95a1efd4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2343.915068] env[62824]: DEBUG nova.virt.hardware [None req-41d0194a-d6d3-4444-b80f-9b2f95a1efd4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2343.915279] env[62824]: DEBUG nova.virt.hardware [None req-41d0194a-d6d3-4444-b80f-9b2f95a1efd4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2343.915438] env[62824]: DEBUG nova.virt.hardware [None req-41d0194a-d6d3-4444-b80f-9b2f95a1efd4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2343.915603] env[62824]: DEBUG nova.virt.hardware [None req-41d0194a-d6d3-4444-b80f-9b2f95a1efd4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2343.915767] env[62824]: DEBUG nova.virt.hardware [None req-41d0194a-d6d3-4444-b80f-9b2f95a1efd4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2343.915948] env[62824]: DEBUG nova.virt.hardware [None req-41d0194a-d6d3-4444-b80f-9b2f95a1efd4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2343.922530] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-41d0194a-d6d3-4444-b80f-9b2f95a1efd4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Reconfiguring VM to attach interface {{(pid=62824) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 2343.927497] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e00cb1d7-3ac9-41e8-9816-4fd62a94e752 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2343.946957] env[62824]: DEBUG oslo_vmware.api [None req-41d0194a-d6d3-4444-b80f-9b2f95a1efd4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Waiting for the task: (returnval){ [ 2343.946957] env[62824]: value = "task-2146865" [ 2343.946957] env[62824]: _type = "Task" [ 2343.946957] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2343.950308] env[62824]: DEBUG oslo_vmware.api [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146862, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.444103} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2343.953937] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] bf6e681d-b773-4787-886a-c90586bc81bf/bf6e681d-b773-4787-886a-c90586bc81bf.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2343.953937] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2343.953937] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c56734bc-d4b4-46c4-aeaf-476ba5c23582 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2343.961372] env[62824]: DEBUG oslo_vmware.api [None req-41d0194a-d6d3-4444-b80f-9b2f95a1efd4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146865, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2343.962783] env[62824]: DEBUG oslo_vmware.api [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2343.962783] env[62824]: value = "task-2146866" [ 2343.962783] env[62824]: _type = "Task" [ 2343.962783] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2343.971531] env[62824]: DEBUG oslo_vmware.api [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146866, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2343.972761] env[62824]: DEBUG nova.scheduler.client.report [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Updated inventory for provider bf95157b-a274-42de-9ccf-9851128a44a1 with generation 133 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 2343.973152] env[62824]: DEBUG nova.compute.provider_tree [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Updating resource provider bf95157b-a274-42de-9ccf-9851128a44a1 generation from 133 to 134 during operation: update_inventory {{(pid=62824) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 2343.973450] env[62824]: DEBUG nova.compute.provider_tree [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Updating inventory in ProviderTree for provider bf95157b-a274-42de-9ccf-9851128a44a1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 2344.005258] env[62824]: DEBUG nova.compute.manager [req-c623d4d5-796e-4022-8ebb-bdbe1a78a46a req-b27fbf9e-69d9-415d-807e-6220b2ac8873 service nova] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Received event network-changed-afcd2bdb-2bfc-40d7-a8e5-2a417be0e0e9 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2344.005450] env[62824]: DEBUG nova.compute.manager [req-c623d4d5-796e-4022-8ebb-bdbe1a78a46a req-b27fbf9e-69d9-415d-807e-6220b2ac8873 service nova] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Refreshing instance network info cache due to event network-changed-afcd2bdb-2bfc-40d7-a8e5-2a417be0e0e9. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2344.005718] env[62824]: DEBUG oslo_concurrency.lockutils [req-c623d4d5-796e-4022-8ebb-bdbe1a78a46a req-b27fbf9e-69d9-415d-807e-6220b2ac8873 service nova] Acquiring lock "refresh_cache-e1574266-e53b-4169-9db6-c66b84895edb" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2344.005884] env[62824]: DEBUG oslo_concurrency.lockutils [req-c623d4d5-796e-4022-8ebb-bdbe1a78a46a req-b27fbf9e-69d9-415d-807e-6220b2ac8873 service nova] Acquired lock "refresh_cache-e1574266-e53b-4169-9db6-c66b84895edb" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2344.006078] env[62824]: DEBUG nova.network.neutron [req-c623d4d5-796e-4022-8ebb-bdbe1a78a46a req-b27fbf9e-69d9-415d-807e-6220b2ac8873 service nova] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Refreshing network info cache for port afcd2bdb-2bfc-40d7-a8e5-2a417be0e0e9 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2344.019323] env[62824]: DEBUG nova.network.neutron [req-755b0534-63bf-4b4e-99eb-9fd530de87f6 req-a58e1424-1757-4bd2-b4f6-effe823f03a8 service nova] [instance: fb9cfb53-285a-4586-acf0-cb08d56b93e7] Updated VIF entry in instance network info cache for port 5278435d-72ae-4fec-844c-0f1174154e30. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2344.019760] env[62824]: DEBUG nova.network.neutron [req-755b0534-63bf-4b4e-99eb-9fd530de87f6 req-a58e1424-1757-4bd2-b4f6-effe823f03a8 service nova] [instance: fb9cfb53-285a-4586-acf0-cb08d56b93e7] Updating instance_info_cache with network_info: [{"id": "5278435d-72ae-4fec-844c-0f1174154e30", "address": "fa:16:3e:24:05:d7", "network": {"id": "0f512108-670f-4a85-9381-108e18b3b5ed", "bridge": "br-int", "label": "tempest-ServersTestJSON-766397782-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dfff08982dad4790bf4d555e2b4db5e4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f3a2eb5-353f-45c5-a73b-869626f4bb13", "external-id": "nsx-vlan-transportzone-411", "segmentation_id": 411, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5278435d-72", "ovs_interfaceid": "5278435d-72ae-4fec-844c-0f1174154e30", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2344.149191] env[62824]: DEBUG oslo_vmware.api [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146863, 'name': PowerOnVM_Task, 'duration_secs': 0.437106} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2344.149532] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2344.149770] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-6bf8eb1d-ca89-4f2e-9975-0f27f2e5337d tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Updating instance 'c480f680-61fd-496b-8cf7-fa50f580b10d' progress to 100 {{(pid=62824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 2344.274500] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146864, 'name': CreateVM_Task, 'duration_secs': 0.426463} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2344.274732] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fb9cfb53-285a-4586-acf0-cb08d56b93e7] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2344.275446] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2344.275617] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2344.275949] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2344.276207] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4b4d163c-2964-47bc-8f69-db409182e70c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2344.281882] env[62824]: DEBUG oslo_vmware.api [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2344.281882] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]529ca040-db94-ed48-3ec8-f3aa5b8a406c" [ 2344.281882] env[62824]: _type = "Task" [ 2344.281882] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2344.290571] env[62824]: DEBUG oslo_vmware.api [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]529ca040-db94-ed48-3ec8-f3aa5b8a406c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2344.342610] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: bbfcb2e3-9326-4548-b15b-e054cbfd192e] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2344.461208] env[62824]: DEBUG oslo_vmware.api [None req-41d0194a-d6d3-4444-b80f-9b2f95a1efd4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146865, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2344.471704] env[62824]: DEBUG oslo_vmware.api [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146866, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071196} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2344.471992] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2344.472755] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8ae316e-bf1f-4df8-be82-87b5d9618aa6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2344.486843] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.172s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2344.497672] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Reconfiguring VM instance instance-0000006b to attach disk [datastore2] bf6e681d-b773-4787-886a-c90586bc81bf/bf6e681d-b773-4787-886a-c90586bc81bf.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2344.498217] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a91cffab-8727-4c5c-8548-922c1982677d tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.010s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2344.498438] env[62824]: DEBUG nova.objects.instance [None req-a91cffab-8727-4c5c-8548-922c1982677d tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Lazy-loading 'resources' on Instance uuid b3003c4b-ae5a-48df-8c12-a915a76253f4 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2344.499605] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0ae5e0cf-d1de-48a4-80f5-08bb5f85f502 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2344.521849] env[62824]: DEBUG oslo_concurrency.lockutils [req-755b0534-63bf-4b4e-99eb-9fd530de87f6 req-a58e1424-1757-4bd2-b4f6-effe823f03a8 service nova] Releasing lock "refresh_cache-fb9cfb53-285a-4586-acf0-cb08d56b93e7" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2344.525942] env[62824]: DEBUG oslo_vmware.api [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2344.525942] env[62824]: value = "task-2146867" [ 2344.525942] env[62824]: _type = "Task" [ 2344.525942] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2344.534617] env[62824]: DEBUG oslo_vmware.api [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146867, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2344.701332] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a339fc0-a7c8-4026-9d2b-35374ea8b4c9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2344.709854] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19825e7b-eb19-4434-9664-7a2e6edd9992 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2344.742079] env[62824]: DEBUG nova.network.neutron [req-c623d4d5-796e-4022-8ebb-bdbe1a78a46a req-b27fbf9e-69d9-415d-807e-6220b2ac8873 service nova] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Updated VIF entry in instance network info cache for port afcd2bdb-2bfc-40d7-a8e5-2a417be0e0e9. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2344.742514] env[62824]: DEBUG nova.network.neutron [req-c623d4d5-796e-4022-8ebb-bdbe1a78a46a req-b27fbf9e-69d9-415d-807e-6220b2ac8873 service nova] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Updating instance_info_cache with network_info: [{"id": "d5c6ba87-e0c7-49c2-bbaf-b7c64f84cdf8", "address": "fa:16:3e:c4:9c:f4", "network": {"id": "aa1c09e5-8d72-43ad-a903-a0f5e711da80", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1643616572-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.166", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c987c5f844be42c99cc9d57a9ca84c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9bc2632-36f9-4912-8782-8bbb789f909d", "external-id": "nsx-vlan-transportzone-897", "segmentation_id": 897, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd5c6ba87-e0", "ovs_interfaceid": "d5c6ba87-e0c7-49c2-bbaf-b7c64f84cdf8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "afcd2bdb-2bfc-40d7-a8e5-2a417be0e0e9", "address": "fa:16:3e:06:1e:c0", "network": {"id": "aa1c09e5-8d72-43ad-a903-a0f5e711da80", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1643616572-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c987c5f844be42c99cc9d57a9ca84c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9bc2632-36f9-4912-8782-8bbb789f909d", "external-id": "nsx-vlan-transportzone-897", "segmentation_id": 897, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapafcd2bdb-2b", "ovs_interfaceid": "afcd2bdb-2bfc-40d7-a8e5-2a417be0e0e9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2344.744591] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a449316-8799-4c41-aff1-c94751aeefd5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2344.752121] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11bb200d-d86b-474a-a46a-349540a8d004 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2344.769628] env[62824]: DEBUG nova.compute.provider_tree [None req-a91cffab-8727-4c5c-8548-922c1982677d tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2344.792046] env[62824]: DEBUG oslo_vmware.api [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]529ca040-db94-ed48-3ec8-f3aa5b8a406c, 'name': SearchDatastore_Task, 'duration_secs': 0.009404} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2344.792338] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2344.792568] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: fb9cfb53-285a-4586-acf0-cb08d56b93e7] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2344.792806] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2344.792964] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2344.793159] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2344.793418] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c91fee31-7813-40f5-9366-10fa666e23a5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2344.802579] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2344.802774] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2344.803493] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-51e3161b-623a-4b54-86e1-0eb56d1b8182 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2344.810831] env[62824]: DEBUG oslo_vmware.api [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2344.810831] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5295706b-d00a-9d96-8064-c3a7bafb1850" [ 2344.810831] env[62824]: _type = "Task" [ 2344.810831] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2344.819061] env[62824]: DEBUG oslo_vmware.api [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5295706b-d00a-9d96-8064-c3a7bafb1850, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2344.845648] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: b0456b8a-348a-4503-a92c-58e5ab455d1c] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2344.964120] env[62824]: DEBUG oslo_vmware.api [None req-41d0194a-d6d3-4444-b80f-9b2f95a1efd4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146865, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2345.003773] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e493306e-cfd3-46a0-8603-ee0f8049cfb0 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Lock "71e9733b-c84b-4501-8faf-a487f089b498" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 42.409s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2345.004423] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Lock "71e9733b-c84b-4501-8faf-a487f089b498" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 2.316s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2345.004608] env[62824]: INFO nova.compute.manager [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Unshelving [ 2345.034995] env[62824]: DEBUG oslo_vmware.api [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146867, 'name': ReconfigVM_Task, 'duration_secs': 0.286572} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2345.035312] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Reconfigured VM instance instance-0000006b to attach disk [datastore2] bf6e681d-b773-4787-886a-c90586bc81bf/bf6e681d-b773-4787-886a-c90586bc81bf.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2345.035964] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1334c29e-93ad-43ec-a4b8-43d811dbba4b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2345.044499] env[62824]: DEBUG oslo_vmware.api [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2345.044499] env[62824]: value = "task-2146868" [ 2345.044499] env[62824]: _type = "Task" [ 2345.044499] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2345.054177] env[62824]: DEBUG oslo_vmware.api [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146868, 'name': Rename_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2345.248496] env[62824]: DEBUG oslo_concurrency.lockutils [req-c623d4d5-796e-4022-8ebb-bdbe1a78a46a req-b27fbf9e-69d9-415d-807e-6220b2ac8873 service nova] Releasing lock "refresh_cache-e1574266-e53b-4169-9db6-c66b84895edb" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2345.272622] env[62824]: DEBUG nova.scheduler.client.report [None req-a91cffab-8727-4c5c-8548-922c1982677d tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2345.322548] env[62824]: DEBUG oslo_vmware.api [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5295706b-d00a-9d96-8064-c3a7bafb1850, 'name': SearchDatastore_Task, 'duration_secs': 0.044679} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2345.323310] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e5b9a2b0-0715-4a7b-aa16-f3f3cfeabcc8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2345.328409] env[62824]: DEBUG oslo_vmware.api [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2345.328409] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52c92fa6-4e4d-1ee1-ad27-3130a871387f" [ 2345.328409] env[62824]: _type = "Task" [ 2345.328409] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2345.336413] env[62824]: DEBUG oslo_vmware.api [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52c92fa6-4e4d-1ee1-ad27-3130a871387f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2345.349017] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: ed7d3e95-f26e-40fd-a370-e17922bbff8e] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2345.462748] env[62824]: DEBUG oslo_vmware.api [None req-41d0194a-d6d3-4444-b80f-9b2f95a1efd4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146865, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2345.556149] env[62824]: DEBUG oslo_vmware.api [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146868, 'name': Rename_Task, 'duration_secs': 0.142281} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2345.556465] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2345.556738] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-064275e4-fd72-4009-9b03-c58b23c8f4d3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2345.565563] env[62824]: DEBUG oslo_vmware.api [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2345.565563] env[62824]: value = "task-2146869" [ 2345.565563] env[62824]: _type = "Task" [ 2345.565563] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2345.574492] env[62824]: DEBUG oslo_vmware.api [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146869, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2345.778185] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a91cffab-8727-4c5c-8548-922c1982677d tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.280s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2345.803240] env[62824]: INFO nova.scheduler.client.report [None req-a91cffab-8727-4c5c-8548-922c1982677d tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Deleted allocations for instance b3003c4b-ae5a-48df-8c12-a915a76253f4 [ 2345.840316] env[62824]: DEBUG oslo_vmware.api [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52c92fa6-4e4d-1ee1-ad27-3130a871387f, 'name': SearchDatastore_Task, 'duration_secs': 0.009794} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2345.840651] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2345.840936] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] fb9cfb53-285a-4586-acf0-cb08d56b93e7/fb9cfb53-285a-4586-acf0-cb08d56b93e7.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2345.841212] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-67c488b4-9011-4cf6-b3e2-e91c603cf794 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2345.850237] env[62824]: DEBUG oslo_vmware.api [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2345.850237] env[62824]: value = "task-2146870" [ 2345.850237] env[62824]: _type = "Task" [ 2345.850237] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2345.853963] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 21c05725-3dad-4965-98d0-0622ebcc8ece] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2345.861390] env[62824]: DEBUG oslo_vmware.api [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146870, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2345.963502] env[62824]: DEBUG oslo_vmware.api [None req-41d0194a-d6d3-4444-b80f-9b2f95a1efd4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146865, 'name': ReconfigVM_Task, 'duration_secs': 1.718813} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2345.964020] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41d0194a-d6d3-4444-b80f-9b2f95a1efd4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Releasing lock "e1574266-e53b-4169-9db6-c66b84895edb" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2345.964259] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-41d0194a-d6d3-4444-b80f-9b2f95a1efd4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Reconfigured VM to attach interface {{(pid=62824) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 2346.029114] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2346.029400] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2346.029638] env[62824]: DEBUG nova.objects.instance [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Lazy-loading 'pci_requests' on Instance uuid 71e9733b-c84b-4501-8faf-a487f089b498 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2346.082794] env[62824]: DEBUG oslo_vmware.api [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146869, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2346.245393] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9c8f0adb-9b23-432f-8854-cd75806cadd6 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquiring lock "c480f680-61fd-496b-8cf7-fa50f580b10d" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2346.245669] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9c8f0adb-9b23-432f-8854-cd75806cadd6 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "c480f680-61fd-496b-8cf7-fa50f580b10d" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2346.245905] env[62824]: DEBUG nova.compute.manager [None req-9c8f0adb-9b23-432f-8854-cd75806cadd6 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Going to confirm migration 2 {{(pid=62824) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5250}} [ 2346.312081] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a91cffab-8727-4c5c-8548-922c1982677d tempest-ServerRescueNegativeTestJSON-1621829740 tempest-ServerRescueNegativeTestJSON-1621829740-project-member] Lock "b3003c4b-ae5a-48df-8c12-a915a76253f4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.155s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2346.356861] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 6a724332-a165-4a2b-9dd8-85f27e7b7637] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2346.362124] env[62824]: DEBUG oslo_vmware.api [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146870, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.510501} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2346.363190] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] fb9cfb53-285a-4586-acf0-cb08d56b93e7/fb9cfb53-285a-4586-acf0-cb08d56b93e7.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2346.363415] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: fb9cfb53-285a-4586-acf0-cb08d56b93e7] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2346.363671] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0cd64e63-57ec-4a38-a603-f214fd927600 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2346.372683] env[62824]: DEBUG oslo_vmware.api [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2346.372683] env[62824]: value = "task-2146871" [ 2346.372683] env[62824]: _type = "Task" [ 2346.372683] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2346.382571] env[62824]: DEBUG oslo_vmware.api [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146871, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2346.469391] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41d0194a-d6d3-4444-b80f-9b2f95a1efd4 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lock "interface-e1574266-e53b-4169-9db6-c66b84895edb-afcd2bdb-2bfc-40d7-a8e5-2a417be0e0e9" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 9.103s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2346.533386] env[62824]: DEBUG nova.objects.instance [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Lazy-loading 'numa_topology' on Instance uuid 71e9733b-c84b-4501-8faf-a487f089b498 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2346.578063] env[62824]: DEBUG oslo_vmware.api [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146869, 'name': PowerOnVM_Task, 'duration_secs': 0.534529} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2346.578347] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2346.578555] env[62824]: DEBUG nova.compute.manager [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2346.579393] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-986aa953-7a9d-4a89-affc-1cd67f025738 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2346.818533] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9c8f0adb-9b23-432f-8854-cd75806cadd6 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquiring lock "refresh_cache-c480f680-61fd-496b-8cf7-fa50f580b10d" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2346.818757] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9c8f0adb-9b23-432f-8854-cd75806cadd6 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquired lock "refresh_cache-c480f680-61fd-496b-8cf7-fa50f580b10d" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2346.818955] env[62824]: DEBUG nova.network.neutron [None req-9c8f0adb-9b23-432f-8854-cd75806cadd6 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2346.819160] env[62824]: DEBUG nova.objects.instance [None req-9c8f0adb-9b23-432f-8854-cd75806cadd6 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lazy-loading 'info_cache' on Instance uuid c480f680-61fd-496b-8cf7-fa50f580b10d {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2346.863553] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 88bad2dd-dce7-41df-b56c-93a5d054c11e] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2346.884058] env[62824]: DEBUG oslo_vmware.api [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146871, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.412725} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2346.884058] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: fb9cfb53-285a-4586-acf0-cb08d56b93e7] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2346.884879] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7116bece-db2c-40ee-a115-56a654d656c6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2346.908021] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: fb9cfb53-285a-4586-acf0-cb08d56b93e7] Reconfiguring VM instance instance-0000006c to attach disk [datastore2] fb9cfb53-285a-4586-acf0-cb08d56b93e7/fb9cfb53-285a-4586-acf0-cb08d56b93e7.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2346.908502] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8c7c4dc4-42b4-4663-baf4-0914506a6292 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2346.928938] env[62824]: DEBUG oslo_vmware.api [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2346.928938] env[62824]: value = "task-2146872" [ 2346.928938] env[62824]: _type = "Task" [ 2346.928938] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2346.937534] env[62824]: DEBUG oslo_vmware.api [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146872, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2347.035687] env[62824]: INFO nova.compute.claims [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2347.097035] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2347.367020] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 4423d172-acdd-4c69-b3b8-ff166e1b8548] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2347.441252] env[62824]: DEBUG oslo_vmware.api [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146872, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2347.651686] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a1b6b2ca-987e-4682-8655-53ca831108d2 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquiring lock "3d294ee9-92f6-4115-95db-9d4b66562b56" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2347.653126] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a1b6b2ca-987e-4682-8655-53ca831108d2 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "3d294ee9-92f6-4115-95db-9d4b66562b56" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2347.870946] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: c65ccd20-b682-420a-9c1a-47a45959197e] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2347.940291] env[62824]: DEBUG oslo_vmware.api [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146872, 'name': ReconfigVM_Task, 'duration_secs': 0.669305} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2347.940612] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: fb9cfb53-285a-4586-acf0-cb08d56b93e7] Reconfigured VM instance instance-0000006c to attach disk [datastore2] fb9cfb53-285a-4586-acf0-cb08d56b93e7/fb9cfb53-285a-4586-acf0-cb08d56b93e7.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2347.941228] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0b470733-a6ee-48c0-9444-94b25c424692 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2347.948474] env[62824]: DEBUG oslo_vmware.api [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2347.948474] env[62824]: value = "task-2146873" [ 2347.948474] env[62824]: _type = "Task" [ 2347.948474] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2347.959304] env[62824]: DEBUG oslo_vmware.api [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146873, 'name': Rename_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2348.021741] env[62824]: DEBUG nova.network.neutron [None req-9c8f0adb-9b23-432f-8854-cd75806cadd6 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Updating instance_info_cache with network_info: [{"id": "9190a574-76bc-4236-9f09-69d08897d634", "address": "fa:16:3e:a9:57:8f", "network": {"id": "b995934c-4934-4469-884c-73cb4de1796f", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1892307918-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c11aef7fb6247cb9b5272a6a063cd12", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "15165046-2de9-4ada-9e99-0126e20854a9", "external-id": "nsx-vlan-transportzone-974", "segmentation_id": 974, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9190a574-76", "ovs_interfaceid": "9190a574-76bc-4236-9f09-69d08897d634", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2348.154678] env[62824]: DEBUG nova.compute.utils [None req-a1b6b2ca-987e-4682-8655-53ca831108d2 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2348.216073] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de662157-941a-44fe-a8e9-f3ea2ee5f94a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2348.226049] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76af5bf6-ce72-47ff-aac5-1bd121673bdb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2348.259010] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df222f1c-13c5-4419-827a-a889bd5ad975 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2348.266748] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b1f0ba5-8267-49c0-9633-99c4ee178310 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2348.280415] env[62824]: DEBUG nova.compute.provider_tree [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Updating inventory in ProviderTree for provider bf95157b-a274-42de-9ccf-9851128a44a1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 2348.374021] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 02405b17-7022-4e2c-a357-415de88f63d0] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2348.461078] env[62824]: DEBUG oslo_vmware.api [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146873, 'name': Rename_Task, 'duration_secs': 0.273305} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2348.461353] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: fb9cfb53-285a-4586-acf0-cb08d56b93e7] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2348.461615] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3f9e5454-1f60-4872-a074-741dc1ab722e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2348.469335] env[62824]: DEBUG oslo_vmware.api [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2348.469335] env[62824]: value = "task-2146874" [ 2348.469335] env[62824]: _type = "Task" [ 2348.469335] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2348.479043] env[62824]: DEBUG oslo_vmware.api [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146874, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2348.524981] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9c8f0adb-9b23-432f-8854-cd75806cadd6 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Releasing lock "refresh_cache-c480f680-61fd-496b-8cf7-fa50f580b10d" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2348.525299] env[62824]: DEBUG nova.objects.instance [None req-9c8f0adb-9b23-432f-8854-cd75806cadd6 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lazy-loading 'migration_context' on Instance uuid c480f680-61fd-496b-8cf7-fa50f580b10d {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2348.660363] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a1b6b2ca-987e-4682-8655-53ca831108d2 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "3d294ee9-92f6-4115-95db-9d4b66562b56" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.008s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2348.815771] env[62824]: DEBUG nova.scheduler.client.report [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Updated inventory for provider bf95157b-a274-42de-9ccf-9851128a44a1 with generation 134 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 2348.816068] env[62824]: DEBUG nova.compute.provider_tree [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Updating resource provider bf95157b-a274-42de-9ccf-9851128a44a1 generation from 134 to 135 during operation: update_inventory {{(pid=62824) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 2348.816271] env[62824]: DEBUG nova.compute.provider_tree [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Updating inventory in ProviderTree for provider bf95157b-a274-42de-9ccf-9851128a44a1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 2348.877095] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: b9f30f87-3594-4468-9d29-70890d8761e3] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2348.980367] env[62824]: DEBUG oslo_vmware.api [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146874, 'name': PowerOnVM_Task} progress is 90%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2349.028640] env[62824]: DEBUG nova.objects.base [None req-9c8f0adb-9b23-432f-8854-cd75806cadd6 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=62824) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 2349.029630] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cdb24d2-8ea5-4cc2-a46a-933f99a4e802 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2349.051237] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-11eb16cf-dc3d-46b2-8c1b-e01ca78f7417 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2349.057144] env[62824]: DEBUG oslo_vmware.api [None req-9c8f0adb-9b23-432f-8854-cd75806cadd6 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 2349.057144] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52e62bbe-1e07-4897-3eb7-ad654f92b0a3" [ 2349.057144] env[62824]: _type = "Task" [ 2349.057144] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2349.065918] env[62824]: DEBUG oslo_vmware.api [None req-9c8f0adb-9b23-432f-8854-cd75806cadd6 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52e62bbe-1e07-4897-3eb7-ad654f92b0a3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2349.321247] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.292s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2349.323369] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 2.227s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2349.323560] env[62824]: DEBUG nova.objects.instance [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62824) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 2349.354058] env[62824]: INFO nova.network.neutron [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Updating port 4214bf6e-29f0-4d92-be44-44fccf7ab149 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 2349.380507] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 35f0435c-fe46-46ab-bc71-c21bb0ad3d55] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2349.482472] env[62824]: DEBUG oslo_vmware.api [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146874, 'name': PowerOnVM_Task, 'duration_secs': 0.632047} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2349.482763] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: fb9cfb53-285a-4586-acf0-cb08d56b93e7] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2349.482978] env[62824]: INFO nova.compute.manager [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: fb9cfb53-285a-4586-acf0-cb08d56b93e7] Took 7.95 seconds to spawn the instance on the hypervisor. [ 2349.483174] env[62824]: DEBUG nova.compute.manager [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: fb9cfb53-285a-4586-acf0-cb08d56b93e7] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2349.483952] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93c163ce-a976-4db0-bbbf-1a597f15efed {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2349.568949] env[62824]: DEBUG oslo_vmware.api [None req-9c8f0adb-9b23-432f-8854-cd75806cadd6 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52e62bbe-1e07-4897-3eb7-ad654f92b0a3, 'name': SearchDatastore_Task, 'duration_secs': 0.022993} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2349.569265] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9c8f0adb-9b23-432f-8854-cd75806cadd6 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2349.720778] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a1b6b2ca-987e-4682-8655-53ca831108d2 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquiring lock "3d294ee9-92f6-4115-95db-9d4b66562b56" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2349.720778] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a1b6b2ca-987e-4682-8655-53ca831108d2 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "3d294ee9-92f6-4115-95db-9d4b66562b56" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2349.720778] env[62824]: INFO nova.compute.manager [None req-a1b6b2ca-987e-4682-8655-53ca831108d2 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Attaching volume 42b59773-27a4-492d-b430-c629b14a2500 to /dev/sdb [ 2349.753702] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ccc0caa-a798-4db1-b186-fb6a71b22420 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2349.760477] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d765875e-e71e-4ba5-bf41-949b91377548 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2349.779836] env[62824]: DEBUG nova.virt.block_device [None req-a1b6b2ca-987e-4682-8655-53ca831108d2 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Updating existing volume attachment record: a970ffff-f407-4ca9-b3fe-84ba1482b67b {{(pid=62824) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 2349.883693] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 2afecec5-763d-4616-b690-41d3101cfc47] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2350.002750] env[62824]: INFO nova.compute.manager [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: fb9cfb53-285a-4586-acf0-cb08d56b93e7] Took 12.79 seconds to build instance. [ 2350.337024] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ea0678d3-a61f-4bad-bad9-6c36781ab881 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.010s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2350.337024] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9c8f0adb-9b23-432f-8854-cd75806cadd6 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.765s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2350.387122] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 8bb90a4d-93c9-4f54-b15e-48fe966e6c97] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2350.425687] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2d178aa0-6a19-48fd-a5cc-3661ef7c3914 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquiring lock "interface-e1574266-e53b-4169-9db6-c66b84895edb-afcd2bdb-2bfc-40d7-a8e5-2a417be0e0e9" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2350.426681] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2d178aa0-6a19-48fd-a5cc-3661ef7c3914 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lock "interface-e1574266-e53b-4169-9db6-c66b84895edb-afcd2bdb-2bfc-40d7-a8e5-2a417be0e0e9" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2350.505220] env[62824]: DEBUG oslo_concurrency.lockutils [None req-e0c3a621-8537-4399-9d95-5931478d5369 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "fb9cfb53-285a-4586-acf0-cb08d56b93e7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.303s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2350.534254] env[62824]: DEBUG oslo_concurrency.lockutils [None req-43cd9d71-795f-434c-b349-420c672b26c6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquiring lock "bf6e681d-b773-4787-886a-c90586bc81bf" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2350.534608] env[62824]: DEBUG oslo_concurrency.lockutils [None req-43cd9d71-795f-434c-b349-420c672b26c6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "bf6e681d-b773-4787-886a-c90586bc81bf" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2350.535273] env[62824]: DEBUG oslo_concurrency.lockutils [None req-43cd9d71-795f-434c-b349-420c672b26c6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquiring lock "bf6e681d-b773-4787-886a-c90586bc81bf-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2350.535660] env[62824]: DEBUG oslo_concurrency.lockutils [None req-43cd9d71-795f-434c-b349-420c672b26c6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "bf6e681d-b773-4787-886a-c90586bc81bf-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2350.536529] env[62824]: DEBUG oslo_concurrency.lockutils [None req-43cd9d71-795f-434c-b349-420c672b26c6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "bf6e681d-b773-4787-886a-c90586bc81bf-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2350.539261] env[62824]: INFO nova.compute.manager [None req-43cd9d71-795f-434c-b349-420c672b26c6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Terminating instance [ 2350.891918] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Acquiring lock "refresh_cache-71e9733b-c84b-4501-8faf-a487f089b498" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2350.891918] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Acquired lock "refresh_cache-71e9733b-c84b-4501-8faf-a487f089b498" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2350.891918] env[62824]: DEBUG nova.network.neutron [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2350.896413] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 387b7f14-f3c1-43ed-8c4e-6de3ce9822e2] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2350.929517] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2d178aa0-6a19-48fd-a5cc-3661ef7c3914 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquiring lock "e1574266-e53b-4169-9db6-c66b84895edb" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2350.930260] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2d178aa0-6a19-48fd-a5cc-3661ef7c3914 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquired lock "e1574266-e53b-4169-9db6-c66b84895edb" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2350.936021] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac76bfbe-05a8-4ad2-b615-4b5d8b3939aa {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2350.957321] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2caf43a1-2aee-4bf2-91b8-80faff766123 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2350.987798] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-2d178aa0-6a19-48fd-a5cc-3661ef7c3914 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Reconfiguring VM to detach interface {{(pid=62824) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 2350.990754] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-07791001-0980-48ec-b475-a581d534f09d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2351.019028] env[62824]: DEBUG oslo_vmware.api [None req-2d178aa0-6a19-48fd-a5cc-3661ef7c3914 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Waiting for the task: (returnval){ [ 2351.019028] env[62824]: value = "task-2146878" [ 2351.019028] env[62824]: _type = "Task" [ 2351.019028] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2351.028646] env[62824]: DEBUG oslo_vmware.api [None req-2d178aa0-6a19-48fd-a5cc-3661ef7c3914 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146878, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2351.043520] env[62824]: DEBUG nova.compute.manager [None req-43cd9d71-795f-434c-b349-420c672b26c6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2351.044127] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-43cd9d71-795f-434c-b349-420c672b26c6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2351.045202] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2916b463-9744-409c-959b-2cf181ead8d4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2351.058034] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-43cd9d71-795f-434c-b349-420c672b26c6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2351.058386] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4d59508f-1bdb-41f1-96c4-e3c335d3ef10 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2351.067543] env[62824]: DEBUG oslo_vmware.api [None req-43cd9d71-795f-434c-b349-420c672b26c6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2351.067543] env[62824]: value = "task-2146879" [ 2351.067543] env[62824]: _type = "Task" [ 2351.067543] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2351.087804] env[62824]: DEBUG oslo_vmware.api [None req-43cd9d71-795f-434c-b349-420c672b26c6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146879, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2351.099815] env[62824]: DEBUG nova.compute.manager [req-4075bcab-5810-43b8-ba5f-59f7485179a5 req-4aea82b8-5cb0-47ba-8e0b-0b96a04a1bf0 service nova] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Received event network-vif-plugged-4214bf6e-29f0-4d92-be44-44fccf7ab149 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2351.099913] env[62824]: DEBUG oslo_concurrency.lockutils [req-4075bcab-5810-43b8-ba5f-59f7485179a5 req-4aea82b8-5cb0-47ba-8e0b-0b96a04a1bf0 service nova] Acquiring lock "71e9733b-c84b-4501-8faf-a487f089b498-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2351.100721] env[62824]: DEBUG oslo_concurrency.lockutils [req-4075bcab-5810-43b8-ba5f-59f7485179a5 req-4aea82b8-5cb0-47ba-8e0b-0b96a04a1bf0 service nova] Lock "71e9733b-c84b-4501-8faf-a487f089b498-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2351.100984] env[62824]: DEBUG oslo_concurrency.lockutils [req-4075bcab-5810-43b8-ba5f-59f7485179a5 req-4aea82b8-5cb0-47ba-8e0b-0b96a04a1bf0 service nova] Lock "71e9733b-c84b-4501-8faf-a487f089b498-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2351.101373] env[62824]: DEBUG nova.compute.manager [req-4075bcab-5810-43b8-ba5f-59f7485179a5 req-4aea82b8-5cb0-47ba-8e0b-0b96a04a1bf0 service nova] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] No waiting events found dispatching network-vif-plugged-4214bf6e-29f0-4d92-be44-44fccf7ab149 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2351.101543] env[62824]: WARNING nova.compute.manager [req-4075bcab-5810-43b8-ba5f-59f7485179a5 req-4aea82b8-5cb0-47ba-8e0b-0b96a04a1bf0 service nova] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Received unexpected event network-vif-plugged-4214bf6e-29f0-4d92-be44-44fccf7ab149 for instance with vm_state shelved_offloaded and task_state spawning. [ 2351.159348] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcb95ac0-4121-497c-a901-a622cbac3a2f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2351.168531] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7a46531-18f8-4e51-877f-2e268a2c0b6e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2351.201599] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0fe3e30-1827-4a0a-9461-2390025e80a2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2351.210242] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b3f1264-89cf-4b17-80df-0b171e453890 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2351.228906] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquiring lock "21a7eed6-c146-4911-9a0a-a1e1644e8e9d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2351.229202] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "21a7eed6-c146-4911-9a0a-a1e1644e8e9d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2351.230916] env[62824]: DEBUG nova.compute.provider_tree [None req-9c8f0adb-9b23-432f-8854-cd75806cadd6 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2351.399137] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 7febb8b4-f984-4d79-a888-b2829f2a9df6] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2351.526692] env[62824]: DEBUG oslo_vmware.api [None req-2d178aa0-6a19-48fd-a5cc-3661ef7c3914 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146878, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2351.581052] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9468b42a-7240-4676-aa48-105eb6ebeca9 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquiring lock "fb9cfb53-285a-4586-acf0-cb08d56b93e7" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2351.581387] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9468b42a-7240-4676-aa48-105eb6ebeca9 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "fb9cfb53-285a-4586-acf0-cb08d56b93e7" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2351.581476] env[62824]: DEBUG nova.compute.manager [None req-9468b42a-7240-4676-aa48-105eb6ebeca9 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: fb9cfb53-285a-4586-acf0-cb08d56b93e7] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2351.581802] env[62824]: DEBUG oslo_vmware.api [None req-43cd9d71-795f-434c-b349-420c672b26c6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146879, 'name': PowerOffVM_Task, 'duration_secs': 0.267331} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2351.582602] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d654b58d-c638-4f0c-b1d7-26d3720f108f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2351.585127] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-43cd9d71-795f-434c-b349-420c672b26c6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2351.585308] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-43cd9d71-795f-434c-b349-420c672b26c6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2351.585552] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d262640f-3cae-4750-97e2-9671dc2ff29f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2351.592050] env[62824]: DEBUG nova.compute.manager [None req-9468b42a-7240-4676-aa48-105eb6ebeca9 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: fb9cfb53-285a-4586-acf0-cb08d56b93e7] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62824) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3404}} [ 2351.592621] env[62824]: DEBUG nova.objects.instance [None req-9468b42a-7240-4676-aa48-105eb6ebeca9 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lazy-loading 'flavor' on Instance uuid fb9cfb53-285a-4586-acf0-cb08d56b93e7 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2351.616738] env[62824]: DEBUG nova.network.neutron [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Updating instance_info_cache with network_info: [{"id": "4214bf6e-29f0-4d92-be44-44fccf7ab149", "address": "fa:16:3e:a6:d8:c4", "network": {"id": "fc5d7128-fe9b-44ef-9d91-839190a97226", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-796405330-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35b7289f4746463e9b1cc47fe914a835", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0954fad3-d24d-496c-83e6-a09d3cb556fc", "external-id": "nsx-vlan-transportzone-216", "segmentation_id": 216, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4214bf6e-29", "ovs_interfaceid": "4214bf6e-29f0-4d92-be44-44fccf7ab149", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2351.734037] env[62824]: DEBUG nova.compute.manager [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2351.737052] env[62824]: DEBUG nova.scheduler.client.report [None req-9c8f0adb-9b23-432f-8854-cd75806cadd6 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2351.902228] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: bbf1440b-8681-48c8-a178-9a83b925c695] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2351.926489] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-43cd9d71-795f-434c-b349-420c672b26c6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2351.926602] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-43cd9d71-795f-434c-b349-420c672b26c6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Deleting contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2351.927296] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-43cd9d71-795f-434c-b349-420c672b26c6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Deleting the datastore file [datastore2] bf6e681d-b773-4787-886a-c90586bc81bf {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2351.927296] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-12e70715-d4da-4046-8100-e6c6194d760b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2351.935713] env[62824]: DEBUG oslo_vmware.api [None req-43cd9d71-795f-434c-b349-420c672b26c6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2351.935713] env[62824]: value = "task-2146881" [ 2351.935713] env[62824]: _type = "Task" [ 2351.935713] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2351.944838] env[62824]: DEBUG oslo_vmware.api [None req-43cd9d71-795f-434c-b349-420c672b26c6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146881, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2352.028443] env[62824]: DEBUG oslo_vmware.api [None req-2d178aa0-6a19-48fd-a5cc-3661ef7c3914 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146878, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2352.120328] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Releasing lock "refresh_cache-71e9733b-c84b-4501-8faf-a487f089b498" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2352.155836] env[62824]: DEBUG nova.virt.hardware [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='910ba0dcea8e9315e95c30da610205e2',container_format='bare',created_at=2025-01-16T14:58:26Z,direct_url=,disk_format='vmdk',id=91de84ba-a43c-4db9-92a7-d2e20dada664,min_disk=1,min_ram=0,name='tempest-ServersNegativeTestJSON-server-1434454935-shelved',owner='35b7289f4746463e9b1cc47fe914a835',properties=ImageMetaProps,protected=,size=31661568,status='active',tags=,updated_at=2025-01-16T14:58:49Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2352.156701] env[62824]: DEBUG nova.virt.hardware [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2352.156701] env[62824]: DEBUG nova.virt.hardware [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2352.156701] env[62824]: DEBUG nova.virt.hardware [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2352.156899] env[62824]: DEBUG nova.virt.hardware [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2352.157034] env[62824]: DEBUG nova.virt.hardware [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2352.157198] env[62824]: DEBUG nova.virt.hardware [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2352.157387] env[62824]: DEBUG nova.virt.hardware [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2352.157558] env[62824]: DEBUG nova.virt.hardware [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2352.157723] env[62824]: DEBUG nova.virt.hardware [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2352.157898] env[62824]: DEBUG nova.virt.hardware [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2352.158948] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9223a30e-9b1e-4f50-984c-9f3496d48438 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2352.167276] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffa9d3f0-9b33-4354-a2e3-a842c3359769 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2352.184074] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a6:d8:c4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0954fad3-d24d-496c-83e6-a09d3cb556fc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4214bf6e-29f0-4d92-be44-44fccf7ab149', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2352.191471] env[62824]: DEBUG oslo.service.loopingcall [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2352.191723] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2352.191943] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7cfed7ea-c7a1-4528-a522-4d16d7039fe9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2352.211385] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2352.211385] env[62824]: value = "task-2146882" [ 2352.211385] env[62824]: _type = "Task" [ 2352.211385] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2352.219571] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146882, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2352.266288] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2352.405846] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 59b5b883-4188-471c-8862-444f3ce08cb0] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2352.447524] env[62824]: DEBUG oslo_vmware.api [None req-43cd9d71-795f-434c-b349-420c672b26c6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146881, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.294389} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2352.448247] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-43cd9d71-795f-434c-b349-420c672b26c6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2352.448247] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-43cd9d71-795f-434c-b349-420c672b26c6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Deleted contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2352.448416] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-43cd9d71-795f-434c-b349-420c672b26c6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2352.448549] env[62824]: INFO nova.compute.manager [None req-43cd9d71-795f-434c-b349-420c672b26c6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Took 1.40 seconds to destroy the instance on the hypervisor. [ 2352.448846] env[62824]: DEBUG oslo.service.loopingcall [None req-43cd9d71-795f-434c-b349-420c672b26c6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2352.449130] env[62824]: DEBUG nova.compute.manager [-] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2352.449413] env[62824]: DEBUG nova.network.neutron [-] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2352.530016] env[62824]: DEBUG oslo_vmware.api [None req-2d178aa0-6a19-48fd-a5cc-3661ef7c3914 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146878, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2352.603468] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-9468b42a-7240-4676-aa48-105eb6ebeca9 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: fb9cfb53-285a-4586-acf0-cb08d56b93e7] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2352.603839] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-764d137e-afbe-410d-9f14-b40e37920f29 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2352.612846] env[62824]: DEBUG oslo_vmware.api [None req-9468b42a-7240-4676-aa48-105eb6ebeca9 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2352.612846] env[62824]: value = "task-2146884" [ 2352.612846] env[62824]: _type = "Task" [ 2352.612846] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2352.623707] env[62824]: DEBUG oslo_vmware.api [None req-9468b42a-7240-4676-aa48-105eb6ebeca9 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146884, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2352.722257] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146882, 'name': CreateVM_Task, 'duration_secs': 0.442156} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2352.722492] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2352.723139] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/91de84ba-a43c-4db9-92a7-d2e20dada664" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2352.723324] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Acquired lock "[datastore2] devstack-image-cache_base/91de84ba-a43c-4db9-92a7-d2e20dada664" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2352.723725] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/91de84ba-a43c-4db9-92a7-d2e20dada664" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2352.723985] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-122df2fe-f171-4a33-b9b2-80011234d5b6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2352.729707] env[62824]: DEBUG oslo_vmware.api [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Waiting for the task: (returnval){ [ 2352.729707] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]521da9db-3827-319f-6c8f-5cacdc6fa867" [ 2352.729707] env[62824]: _type = "Task" [ 2352.729707] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2352.738070] env[62824]: DEBUG oslo_vmware.api [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]521da9db-3827-319f-6c8f-5cacdc6fa867, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2352.751601] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9c8f0adb-9b23-432f-8854-cd75806cadd6 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.417s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2352.754355] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.488s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2352.755797] env[62824]: INFO nova.compute.claims [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2352.910844] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: c36ece43-3d70-4e67-a740-9057f413c722] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2353.028696] env[62824]: DEBUG oslo_vmware.api [None req-2d178aa0-6a19-48fd-a5cc-3661ef7c3914 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146878, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2353.123363] env[62824]: DEBUG oslo_vmware.api [None req-9468b42a-7240-4676-aa48-105eb6ebeca9 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146884, 'name': PowerOffVM_Task, 'duration_secs': 0.286926} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2353.123363] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-9468b42a-7240-4676-aa48-105eb6ebeca9 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: fb9cfb53-285a-4586-acf0-cb08d56b93e7] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2353.123524] env[62824]: DEBUG nova.compute.manager [None req-9468b42a-7240-4676-aa48-105eb6ebeca9 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: fb9cfb53-285a-4586-acf0-cb08d56b93e7] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2353.124301] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bf0ef1f-ccc0-4f70-b318-f6282e3ae283 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2353.195235] env[62824]: DEBUG nova.compute.manager [req-eeefe87b-d332-4d5d-ad09-bfe0f8da8922 req-14ff753c-fd68-45b5-b96c-7d3b46159dad service nova] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Received event network-changed-4214bf6e-29f0-4d92-be44-44fccf7ab149 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2353.195428] env[62824]: DEBUG nova.compute.manager [req-eeefe87b-d332-4d5d-ad09-bfe0f8da8922 req-14ff753c-fd68-45b5-b96c-7d3b46159dad service nova] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Refreshing instance network info cache due to event network-changed-4214bf6e-29f0-4d92-be44-44fccf7ab149. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2353.195651] env[62824]: DEBUG oslo_concurrency.lockutils [req-eeefe87b-d332-4d5d-ad09-bfe0f8da8922 req-14ff753c-fd68-45b5-b96c-7d3b46159dad service nova] Acquiring lock "refresh_cache-71e9733b-c84b-4501-8faf-a487f089b498" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2353.195832] env[62824]: DEBUG oslo_concurrency.lockutils [req-eeefe87b-d332-4d5d-ad09-bfe0f8da8922 req-14ff753c-fd68-45b5-b96c-7d3b46159dad service nova] Acquired lock "refresh_cache-71e9733b-c84b-4501-8faf-a487f089b498" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2353.195948] env[62824]: DEBUG nova.network.neutron [req-eeefe87b-d332-4d5d-ad09-bfe0f8da8922 req-14ff753c-fd68-45b5-b96c-7d3b46159dad service nova] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Refreshing network info cache for port 4214bf6e-29f0-4d92-be44-44fccf7ab149 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2353.240152] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Releasing lock "[datastore2] devstack-image-cache_base/91de84ba-a43c-4db9-92a7-d2e20dada664" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2353.240428] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Processing image 91de84ba-a43c-4db9-92a7-d2e20dada664 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2353.240688] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/91de84ba-a43c-4db9-92a7-d2e20dada664/91de84ba-a43c-4db9-92a7-d2e20dada664.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2353.240847] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Acquired lock "[datastore2] devstack-image-cache_base/91de84ba-a43c-4db9-92a7-d2e20dada664/91de84ba-a43c-4db9-92a7-d2e20dada664.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2353.241043] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2353.241345] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9fc5b6e5-8bac-4486-923c-ed45c8c8f162 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2353.252700] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2353.252899] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2353.253681] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-49dc7fa1-e223-4f41-a7a8-1f7121064886 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2353.263855] env[62824]: DEBUG oslo_vmware.api [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Waiting for the task: (returnval){ [ 2353.263855] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]523eb48b-d93a-0e72-6104-d7c3ab727f89" [ 2353.263855] env[62824]: _type = "Task" [ 2353.263855] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2353.277878] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Preparing fetch location {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 2353.278161] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Fetch image to [datastore2] OSTACK_IMG_bb242f32-2f4c-4ad0-a044-984e9dd0b356/OSTACK_IMG_bb242f32-2f4c-4ad0-a044-984e9dd0b356.vmdk {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 2353.278362] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Downloading stream optimized image 91de84ba-a43c-4db9-92a7-d2e20dada664 to [datastore2] OSTACK_IMG_bb242f32-2f4c-4ad0-a044-984e9dd0b356/OSTACK_IMG_bb242f32-2f4c-4ad0-a044-984e9dd0b356.vmdk on the data store datastore2 as vApp {{(pid=62824) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 2353.278536] env[62824]: DEBUG nova.virt.vmwareapi.images [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Downloading image file data 91de84ba-a43c-4db9-92a7-d2e20dada664 to the ESX as VM named 'OSTACK_IMG_bb242f32-2f4c-4ad0-a044-984e9dd0b356' {{(pid=62824) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 2353.325503] env[62824]: INFO nova.scheduler.client.report [None req-9c8f0adb-9b23-432f-8854-cd75806cadd6 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Deleted allocation for migration e66ebab6-1da6-493b-b7f5-ef121278bf32 [ 2353.369575] env[62824]: DEBUG oslo_vmware.rw_handles [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 2353.369575] env[62824]: value = "resgroup-9" [ 2353.369575] env[62824]: _type = "ResourcePool" [ 2353.369575] env[62824]: }. {{(pid=62824) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 2353.369929] env[62824]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-ec6e7804-a8e8-4cb2-b110-ca2c10785f78 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2353.393981] env[62824]: DEBUG oslo_vmware.rw_handles [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Lease: (returnval){ [ 2353.393981] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]524f788c-060c-c7a0-a549-9901229e6ed8" [ 2353.393981] env[62824]: _type = "HttpNfcLease" [ 2353.393981] env[62824]: } obtained for vApp import into resource pool (val){ [ 2353.393981] env[62824]: value = "resgroup-9" [ 2353.393981] env[62824]: _type = "ResourcePool" [ 2353.393981] env[62824]: }. {{(pid=62824) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 2353.393981] env[62824]: DEBUG oslo_vmware.api [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Waiting for the lease: (returnval){ [ 2353.393981] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]524f788c-060c-c7a0-a549-9901229e6ed8" [ 2353.393981] env[62824]: _type = "HttpNfcLease" [ 2353.393981] env[62824]: } to be ready. {{(pid=62824) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 2353.402451] env[62824]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 2353.402451] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]524f788c-060c-c7a0-a549-9901229e6ed8" [ 2353.402451] env[62824]: _type = "HttpNfcLease" [ 2353.402451] env[62824]: } is initializing. {{(pid=62824) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 2353.413650] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: f3740d9a-b21a-4a79-9e28-2a89ecd08bb6] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2353.448658] env[62824]: DEBUG nova.network.neutron [-] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2353.529096] env[62824]: DEBUG oslo_vmware.api [None req-2d178aa0-6a19-48fd-a5cc-3661ef7c3914 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146878, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2353.638900] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9468b42a-7240-4676-aa48-105eb6ebeca9 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "fb9cfb53-285a-4586-acf0-cb08d56b93e7" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.057s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2353.832800] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9c8f0adb-9b23-432f-8854-cd75806cadd6 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "c480f680-61fd-496b-8cf7-fa50f580b10d" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 7.587s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2353.906944] env[62824]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 2353.906944] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]524f788c-060c-c7a0-a549-9901229e6ed8" [ 2353.906944] env[62824]: _type = "HttpNfcLease" [ 2353.906944] env[62824]: } is initializing. {{(pid=62824) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 2353.917804] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 07ce931f-45ef-409b-b714-9f1cd47a3a88] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2353.951832] env[62824]: INFO nova.compute.manager [-] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Took 1.50 seconds to deallocate network for instance. [ 2354.011043] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59541697-4e45-4c7f-a3c2-faab8d44eef6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2354.019321] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c572aef0-bb34-4622-a41f-9a3fa0ab8fb3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2354.033933] env[62824]: DEBUG oslo_vmware.api [None req-2d178aa0-6a19-48fd-a5cc-3661ef7c3914 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146878, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2354.060819] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d93639a5-e9c2-4fed-b4e9-bb63fcc81f52 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2354.070223] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a799488d-091a-4b6e-8c49-9df96be23b5e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2354.085539] env[62824]: DEBUG nova.compute.provider_tree [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2354.088170] env[62824]: DEBUG nova.network.neutron [req-eeefe87b-d332-4d5d-ad09-bfe0f8da8922 req-14ff753c-fd68-45b5-b96c-7d3b46159dad service nova] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Updated VIF entry in instance network info cache for port 4214bf6e-29f0-4d92-be44-44fccf7ab149. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2354.088479] env[62824]: DEBUG nova.network.neutron [req-eeefe87b-d332-4d5d-ad09-bfe0f8da8922 req-14ff753c-fd68-45b5-b96c-7d3b46159dad service nova] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Updating instance_info_cache with network_info: [{"id": "4214bf6e-29f0-4d92-be44-44fccf7ab149", "address": "fa:16:3e:a6:d8:c4", "network": {"id": "fc5d7128-fe9b-44ef-9d91-839190a97226", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-796405330-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35b7289f4746463e9b1cc47fe914a835", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0954fad3-d24d-496c-83e6-a09d3cb556fc", "external-id": "nsx-vlan-transportzone-216", "segmentation_id": 216, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4214bf6e-29", "ovs_interfaceid": "4214bf6e-29f0-4d92-be44-44fccf7ab149", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2354.391695] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b70c134a-661f-4db0-98ce-919ec5cda15d tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquiring lock "fb9cfb53-285a-4586-acf0-cb08d56b93e7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2354.392141] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b70c134a-661f-4db0-98ce-919ec5cda15d tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "fb9cfb53-285a-4586-acf0-cb08d56b93e7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2354.392813] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b70c134a-661f-4db0-98ce-919ec5cda15d tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquiring lock "fb9cfb53-285a-4586-acf0-cb08d56b93e7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2354.392813] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b70c134a-661f-4db0-98ce-919ec5cda15d tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "fb9cfb53-285a-4586-acf0-cb08d56b93e7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2354.392813] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b70c134a-661f-4db0-98ce-919ec5cda15d tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "fb9cfb53-285a-4586-acf0-cb08d56b93e7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2354.395497] env[62824]: INFO nova.compute.manager [None req-b70c134a-661f-4db0-98ce-919ec5cda15d tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: fb9cfb53-285a-4586-acf0-cb08d56b93e7] Terminating instance [ 2354.406955] env[62824]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 2354.406955] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]524f788c-060c-c7a0-a549-9901229e6ed8" [ 2354.406955] env[62824]: _type = "HttpNfcLease" [ 2354.406955] env[62824]: } is initializing. {{(pid=62824) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 2354.425749] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: dbfa6f8d-6c4d-474d-820a-3a6ebe86fcda] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2354.460581] env[62824]: DEBUG oslo_concurrency.lockutils [None req-43cd9d71-795f-434c-b349-420c672b26c6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2354.533353] env[62824]: DEBUG oslo_vmware.api [None req-2d178aa0-6a19-48fd-a5cc-3661ef7c3914 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146878, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2354.591231] env[62824]: DEBUG oslo_concurrency.lockutils [req-eeefe87b-d332-4d5d-ad09-bfe0f8da8922 req-14ff753c-fd68-45b5-b96c-7d3b46159dad service nova] Releasing lock "refresh_cache-71e9733b-c84b-4501-8faf-a487f089b498" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2354.591496] env[62824]: DEBUG nova.compute.manager [req-eeefe87b-d332-4d5d-ad09-bfe0f8da8922 req-14ff753c-fd68-45b5-b96c-7d3b46159dad service nova] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Received event network-vif-deleted-34ec0688-79ff-4219-afc1-498ec608f081 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2354.592322] env[62824]: INFO nova.compute.manager [req-eeefe87b-d332-4d5d-ad09-bfe0f8da8922 req-14ff753c-fd68-45b5-b96c-7d3b46159dad service nova] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Neutron deleted interface 34ec0688-79ff-4219-afc1-498ec608f081; detaching it from the instance and deleting it from the info cache [ 2354.592566] env[62824]: DEBUG nova.network.neutron [req-eeefe87b-d332-4d5d-ad09-bfe0f8da8922 req-14ff753c-fd68-45b5-b96c-7d3b46159dad service nova] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2354.594475] env[62824]: DEBUG nova.scheduler.client.report [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2354.830603] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1b6b2ca-987e-4682-8655-53ca831108d2 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Volume attach. Driver type: vmdk {{(pid=62824) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 2354.830881] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1b6b2ca-987e-4682-8655-53ca831108d2 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-438761', 'volume_id': '42b59773-27a4-492d-b430-c629b14a2500', 'name': 'volume-42b59773-27a4-492d-b430-c629b14a2500', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '3d294ee9-92f6-4115-95db-9d4b66562b56', 'attached_at': '', 'detached_at': '', 'volume_id': '42b59773-27a4-492d-b430-c629b14a2500', 'serial': '42b59773-27a4-492d-b430-c629b14a2500'} {{(pid=62824) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 2354.831794] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-621dafea-14d1-4a3f-a17f-e3dce2bff525 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2354.850886] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f007c471-5833-49fa-81d8-85a04b6f282c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2354.877713] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1b6b2ca-987e-4682-8655-53ca831108d2 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Reconfiguring VM instance instance-00000066 to attach disk [datastore2] volume-42b59773-27a4-492d-b430-c629b14a2500/volume-42b59773-27a4-492d-b430-c629b14a2500.vmdk or device None with type thin {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2354.878082] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2890a8d5-37b5-4f49-ba73-c6b38b47911e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2354.899464] env[62824]: DEBUG nova.compute.manager [None req-b70c134a-661f-4db0-98ce-919ec5cda15d tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: fb9cfb53-285a-4586-acf0-cb08d56b93e7] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2354.899609] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-b70c134a-661f-4db0-98ce-919ec5cda15d tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: fb9cfb53-285a-4586-acf0-cb08d56b93e7] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2354.901891] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5ffc4f8-dfb6-40b5-b7be-540be883cc12 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2354.908051] env[62824]: DEBUG oslo_vmware.api [None req-a1b6b2ca-987e-4682-8655-53ca831108d2 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2354.908051] env[62824]: value = "task-2146886" [ 2354.908051] env[62824]: _type = "Task" [ 2354.908051] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2354.915460] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-b70c134a-661f-4db0-98ce-919ec5cda15d tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: fb9cfb53-285a-4586-acf0-cb08d56b93e7] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2354.916052] env[62824]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 2354.916052] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]524f788c-060c-c7a0-a549-9901229e6ed8" [ 2354.916052] env[62824]: _type = "HttpNfcLease" [ 2354.916052] env[62824]: } is initializing. {{(pid=62824) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 2354.916760] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-848deabf-66a9-4875-a53b-a84ce8b2b001 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2354.921536] env[62824]: DEBUG oslo_vmware.api [None req-a1b6b2ca-987e-4682-8655-53ca831108d2 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146886, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2354.929341] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: f4d63a93-23af-470c-b36a-662af81dc386] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2355.034792] env[62824]: DEBUG oslo_vmware.api [None req-2d178aa0-6a19-48fd-a5cc-3661ef7c3914 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146878, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2355.103027] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.346s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2355.103027] env[62824]: DEBUG nova.compute.manager [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2355.104831] env[62824]: DEBUG oslo_concurrency.lockutils [None req-43cd9d71-795f-434c-b349-420c672b26c6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.644s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2355.105241] env[62824]: DEBUG nova.objects.instance [None req-43cd9d71-795f-434c-b349-420c672b26c6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lazy-loading 'resources' on Instance uuid bf6e681d-b773-4787-886a-c90586bc81bf {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2355.108167] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c5ed5c78-a557-498d-8197-442415d7c7e5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2355.120135] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2333256-2ae0-403c-94a0-d61972d1fc42 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2355.160383] env[62824]: DEBUG nova.compute.manager [req-eeefe87b-d332-4d5d-ad09-bfe0f8da8922 req-14ff753c-fd68-45b5-b96c-7d3b46159dad service nova] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Detach interface failed, port_id=34ec0688-79ff-4219-afc1-498ec608f081, reason: Instance bf6e681d-b773-4787-886a-c90586bc81bf could not be found. {{(pid=62824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11482}} [ 2355.274211] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9c8f0adb-9b23-432f-8854-cd75806cadd6 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquiring lock "c480f680-61fd-496b-8cf7-fa50f580b10d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2355.274524] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9c8f0adb-9b23-432f-8854-cd75806cadd6 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "c480f680-61fd-496b-8cf7-fa50f580b10d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2355.274727] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9c8f0adb-9b23-432f-8854-cd75806cadd6 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquiring lock "c480f680-61fd-496b-8cf7-fa50f580b10d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2355.274911] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9c8f0adb-9b23-432f-8854-cd75806cadd6 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "c480f680-61fd-496b-8cf7-fa50f580b10d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2355.275619] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9c8f0adb-9b23-432f-8854-cd75806cadd6 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "c480f680-61fd-496b-8cf7-fa50f580b10d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2355.277534] env[62824]: INFO nova.compute.manager [None req-9c8f0adb-9b23-432f-8854-cd75806cadd6 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Terminating instance [ 2355.409689] env[62824]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 2355.409689] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]524f788c-060c-c7a0-a549-9901229e6ed8" [ 2355.409689] env[62824]: _type = "HttpNfcLease" [ 2355.409689] env[62824]: } is ready. {{(pid=62824) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 2355.412988] env[62824]: DEBUG oslo_vmware.rw_handles [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 2355.412988] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]524f788c-060c-c7a0-a549-9901229e6ed8" [ 2355.412988] env[62824]: _type = "HttpNfcLease" [ 2355.412988] env[62824]: }. {{(pid=62824) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 2355.415613] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dca20d4c-319d-4fc7-a171-0ba80a2138f2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2355.424712] env[62824]: DEBUG oslo_vmware.rw_handles [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ccfa91-30ab-517d-d001-cdc7a317cc18/disk-0.vmdk from lease info. {{(pid=62824) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 2355.424911] env[62824]: DEBUG oslo_vmware.rw_handles [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Creating HTTP connection to write to file with size = 31661568 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ccfa91-30ab-517d-d001-cdc7a317cc18/disk-0.vmdk. {{(pid=62824) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 2355.429843] env[62824]: DEBUG oslo_vmware.api [None req-a1b6b2ca-987e-4682-8655-53ca831108d2 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146886, 'name': ReconfigVM_Task, 'duration_secs': 0.375338} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2355.431091] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1b6b2ca-987e-4682-8655-53ca831108d2 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Reconfigured VM instance instance-00000066 to attach disk [datastore2] volume-42b59773-27a4-492d-b430-c629b14a2500/volume-42b59773-27a4-492d-b430-c629b14a2500.vmdk or device None with type thin {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2355.499253] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 82382932-7302-4441-a6f8-9aa2300ec0f6] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2355.501062] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3f5a8a50-a169-4f1b-827a-8ee8597b6e1f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2355.517019] env[62824]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-ae2191de-2894-4dff-95ca-9ec7e89e79cc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2355.522201] env[62824]: DEBUG oslo_vmware.api [None req-a1b6b2ca-987e-4682-8655-53ca831108d2 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2355.522201] env[62824]: value = "task-2146888" [ 2355.522201] env[62824]: _type = "Task" [ 2355.522201] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2355.535262] env[62824]: DEBUG oslo_vmware.api [None req-a1b6b2ca-987e-4682-8655-53ca831108d2 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146888, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2355.538359] env[62824]: DEBUG oslo_vmware.api [None req-2d178aa0-6a19-48fd-a5cc-3661ef7c3914 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146878, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2355.612369] env[62824]: DEBUG nova.compute.utils [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2355.616458] env[62824]: DEBUG nova.compute.manager [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2355.616625] env[62824]: DEBUG nova.network.neutron [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2355.664083] env[62824]: DEBUG nova.policy [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e8c10fcbfe69448fa71cfad4a7c8e179', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4479921caf5f405b8fc49baad390a0e9', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2355.783121] env[62824]: DEBUG nova.compute.manager [None req-9c8f0adb-9b23-432f-8854-cd75806cadd6 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2355.783121] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-9c8f0adb-9b23-432f-8854-cd75806cadd6 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2355.784024] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae900f91-af56-43a8-a7ca-5ab284f12bb6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2355.796099] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c8f0adb-9b23-432f-8854-cd75806cadd6 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2355.796099] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-74f0fdb4-6583-46fb-a29f-269672f1c74b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2355.806263] env[62824]: DEBUG oslo_vmware.api [None req-9c8f0adb-9b23-432f-8854-cd75806cadd6 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 2355.806263] env[62824]: value = "task-2146889" [ 2355.806263] env[62824]: _type = "Task" [ 2355.806263] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2355.815653] env[62824]: DEBUG oslo_vmware.api [None req-9c8f0adb-9b23-432f-8854-cd75806cadd6 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146889, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2355.854134] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f487918-e4ec-48ec-87bb-cf4b8d8ae743 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2355.863348] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-773b11c2-2fdf-4468-88f4-f3d7e6c25209 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2355.906342] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48a472c8-f626-40be-85e9-164493a62c19 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2355.919192] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0ba7dc7-27dd-4f00-baec-468c310ab369 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2355.935218] env[62824]: DEBUG nova.compute.provider_tree [None req-43cd9d71-795f-434c-b349-420c672b26c6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Updating inventory in ProviderTree for provider bf95157b-a274-42de-9ccf-9851128a44a1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 2355.978737] env[62824]: DEBUG nova.network.neutron [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Successfully created port: edaf5d3c-004d-4ba0-b037-e7eef2acbe44 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2356.012610] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 29b8dd5f-1855-490c-a01b-54840073a753] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2356.041100] env[62824]: DEBUG oslo_vmware.api [None req-a1b6b2ca-987e-4682-8655-53ca831108d2 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146888, 'name': ReconfigVM_Task, 'duration_secs': 0.165261} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2356.048577] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1b6b2ca-987e-4682-8655-53ca831108d2 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-438761', 'volume_id': '42b59773-27a4-492d-b430-c629b14a2500', 'name': 'volume-42b59773-27a4-492d-b430-c629b14a2500', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '3d294ee9-92f6-4115-95db-9d4b66562b56', 'attached_at': '', 'detached_at': '', 'volume_id': '42b59773-27a4-492d-b430-c629b14a2500', 'serial': '42b59773-27a4-492d-b430-c629b14a2500'} {{(pid=62824) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 2356.050052] env[62824]: DEBUG oslo_vmware.api [None req-2d178aa0-6a19-48fd-a5cc-3661ef7c3914 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146878, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2356.118901] env[62824]: DEBUG nova.compute.manager [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2356.318604] env[62824]: DEBUG oslo_vmware.api [None req-9c8f0adb-9b23-432f-8854-cd75806cadd6 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146889, 'name': PowerOffVM_Task, 'duration_secs': 0.267363} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2356.318895] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c8f0adb-9b23-432f-8854-cd75806cadd6 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2356.319098] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-9c8f0adb-9b23-432f-8854-cd75806cadd6 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2356.319480] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f1e67709-c0b4-41eb-a83a-3b2e92dd17e5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2356.465464] env[62824]: ERROR nova.scheduler.client.report [None req-43cd9d71-795f-434c-b349-420c672b26c6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [req-9aa4f4df-38c6-4355-8432-ab6ef2609e90] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID bf95157b-a274-42de-9ccf-9851128a44a1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-9aa4f4df-38c6-4355-8432-ab6ef2609e90"}]} [ 2356.483569] env[62824]: DEBUG nova.scheduler.client.report [None req-43cd9d71-795f-434c-b349-420c672b26c6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Refreshing inventories for resource provider bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 2356.500036] env[62824]: DEBUG nova.scheduler.client.report [None req-43cd9d71-795f-434c-b349-420c672b26c6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Updating ProviderTree inventory for provider bf95157b-a274-42de-9ccf-9851128a44a1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 2356.500309] env[62824]: DEBUG nova.compute.provider_tree [None req-43cd9d71-795f-434c-b349-420c672b26c6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Updating inventory in ProviderTree for provider bf95157b-a274-42de-9ccf-9851128a44a1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 2356.513203] env[62824]: DEBUG nova.scheduler.client.report [None req-43cd9d71-795f-434c-b349-420c672b26c6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Refreshing aggregate associations for resource provider bf95157b-a274-42de-9ccf-9851128a44a1, aggregates: None {{(pid=62824) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 2356.515856] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2356.535147] env[62824]: DEBUG nova.scheduler.client.report [None req-43cd9d71-795f-434c-b349-420c672b26c6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Refreshing trait associations for resource provider bf95157b-a274-42de-9ccf-9851128a44a1, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_ISO {{(pid=62824) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 2356.541880] env[62824]: DEBUG oslo_vmware.api [None req-2d178aa0-6a19-48fd-a5cc-3661ef7c3914 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146878, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2356.775399] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5edf42a-ac5a-4420-85b9-f422612d0eeb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2356.781996] env[62824]: DEBUG oslo_vmware.rw_handles [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Completed reading data from the image iterator. {{(pid=62824) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 2356.782242] env[62824]: DEBUG oslo_vmware.rw_handles [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ccfa91-30ab-517d-d001-cdc7a317cc18/disk-0.vmdk. {{(pid=62824) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 2356.783077] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82fe305b-591a-442f-a763-273bc61481fd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2356.790016] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d7e67de-6d0b-4054-9a8b-e982bffd3d75 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2356.796103] env[62824]: DEBUG oslo_vmware.rw_handles [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ccfa91-30ab-517d-d001-cdc7a317cc18/disk-0.vmdk is in state: ready. {{(pid=62824) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 2356.797833] env[62824]: DEBUG oslo_vmware.rw_handles [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ccfa91-30ab-517d-d001-cdc7a317cc18/disk-0.vmdk. {{(pid=62824) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 2356.797833] env[62824]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-a046484b-dfab-47dd-85b3-eb3645f8c1e2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2356.829069] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8cc8cdd-9de8-4560-bbc2-5c1c68c7963e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2356.838511] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e804fdcb-f02a-4e96-a73c-3224c845a052 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2356.854435] env[62824]: DEBUG nova.compute.provider_tree [None req-43cd9d71-795f-434c-b349-420c672b26c6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Updating inventory in ProviderTree for provider bf95157b-a274-42de-9ccf-9851128a44a1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 2357.038037] env[62824]: DEBUG oslo_vmware.api [None req-2d178aa0-6a19-48fd-a5cc-3661ef7c3914 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146878, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2357.102335] env[62824]: DEBUG nova.objects.instance [None req-a1b6b2ca-987e-4682-8655-53ca831108d2 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lazy-loading 'flavor' on Instance uuid 3d294ee9-92f6-4115-95db-9d4b66562b56 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2357.131965] env[62824]: DEBUG nova.compute.manager [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2357.162531] env[62824]: DEBUG nova.virt.hardware [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2357.162931] env[62824]: DEBUG nova.virt.hardware [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2357.163169] env[62824]: DEBUG nova.virt.hardware [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2357.163374] env[62824]: DEBUG nova.virt.hardware [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2357.163545] env[62824]: DEBUG nova.virt.hardware [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2357.163704] env[62824]: DEBUG nova.virt.hardware [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2357.163912] env[62824]: DEBUG nova.virt.hardware [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2357.164084] env[62824]: DEBUG nova.virt.hardware [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2357.164262] env[62824]: DEBUG nova.virt.hardware [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2357.164428] env[62824]: DEBUG nova.virt.hardware [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2357.164604] env[62824]: DEBUG nova.virt.hardware [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2357.165486] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d805a6d-e718-43f4-a156-d14525e234ba {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2357.174298] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e4f8b1c-3cca-49c8-8390-1e3750920faf {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2357.221349] env[62824]: DEBUG oslo_vmware.rw_handles [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ccfa91-30ab-517d-d001-cdc7a317cc18/disk-0.vmdk. {{(pid=62824) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 2357.222032] env[62824]: INFO nova.virt.vmwareapi.images [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Downloaded image file data 91de84ba-a43c-4db9-92a7-d2e20dada664 [ 2357.222709] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e8dde3a-ad59-4ea4-b09d-d8c101f0acd5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2357.243408] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a1d2637d-a7f7-4dd9-8b24-9503a1fda98e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2357.279021] env[62824]: INFO nova.virt.vmwareapi.images [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] The imported VM was unregistered [ 2357.280034] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Caching image {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 2357.280316] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Creating directory with path [datastore2] devstack-image-cache_base/91de84ba-a43c-4db9-92a7-d2e20dada664 {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2357.280605] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-14d5af4f-54fc-4c93-9ce4-5d9321fb24fa {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2357.295764] env[62824]: DEBUG oslo_concurrency.lockutils [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Acquiring lock "c70880c9-a69a-49b9-b028-189c04276623" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2357.296059] env[62824]: DEBUG oslo_concurrency.lockutils [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Lock "c70880c9-a69a-49b9-b028-189c04276623" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2357.303582] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Created directory with path [datastore2] devstack-image-cache_base/91de84ba-a43c-4db9-92a7-d2e20dada664 {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2357.303746] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_bb242f32-2f4c-4ad0-a044-984e9dd0b356/OSTACK_IMG_bb242f32-2f4c-4ad0-a044-984e9dd0b356.vmdk to [datastore2] devstack-image-cache_base/91de84ba-a43c-4db9-92a7-d2e20dada664/91de84ba-a43c-4db9-92a7-d2e20dada664.vmdk. {{(pid=62824) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 2357.304014] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-9282cec9-8a65-42a5-94bb-dcd6d73ec91d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2357.312989] env[62824]: DEBUG oslo_vmware.api [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Waiting for the task: (returnval){ [ 2357.312989] env[62824]: value = "task-2146892" [ 2357.312989] env[62824]: _type = "Task" [ 2357.312989] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2357.321783] env[62824]: DEBUG oslo_vmware.api [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146892, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2357.379554] env[62824]: ERROR nova.scheduler.client.report [None req-43cd9d71-795f-434c-b349-420c672b26c6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [req-2ff339e0-a65d-4490-b3f2-e621816b850c] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID bf95157b-a274-42de-9ccf-9851128a44a1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-2ff339e0-a65d-4490-b3f2-e621816b850c"}]} [ 2357.396882] env[62824]: DEBUG nova.scheduler.client.report [None req-43cd9d71-795f-434c-b349-420c672b26c6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Refreshing inventories for resource provider bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 2357.412189] env[62824]: DEBUG nova.scheduler.client.report [None req-43cd9d71-795f-434c-b349-420c672b26c6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Updating ProviderTree inventory for provider bf95157b-a274-42de-9ccf-9851128a44a1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 2357.412462] env[62824]: DEBUG nova.compute.provider_tree [None req-43cd9d71-795f-434c-b349-420c672b26c6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Updating inventory in ProviderTree for provider bf95157b-a274-42de-9ccf-9851128a44a1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 2357.424396] env[62824]: DEBUG nova.scheduler.client.report [None req-43cd9d71-795f-434c-b349-420c672b26c6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Refreshing aggregate associations for resource provider bf95157b-a274-42de-9ccf-9851128a44a1, aggregates: None {{(pid=62824) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 2357.444775] env[62824]: DEBUG nova.scheduler.client.report [None req-43cd9d71-795f-434c-b349-420c672b26c6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Refreshing trait associations for resource provider bf95157b-a274-42de-9ccf-9851128a44a1, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_ISO {{(pid=62824) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 2357.540766] env[62824]: DEBUG oslo_vmware.api [None req-2d178aa0-6a19-48fd-a5cc-3661ef7c3914 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146878, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2357.609037] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a1b6b2ca-987e-4682-8655-53ca831108d2 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "3d294ee9-92f6-4115-95db-9d4b66562b56" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.889s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2357.655605] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac3d92b4-145c-4347-a52b-108a5efd967f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2357.664563] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84792260-1647-430f-a524-e686852be225 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2357.701349] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1724146a-3787-4428-8728-44ace9c163a9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2357.711445] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f790d55-74ec-4d7c-b32e-f5e680a8aae2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2357.729832] env[62824]: DEBUG nova.compute.provider_tree [None req-43cd9d71-795f-434c-b349-420c672b26c6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Updating inventory in ProviderTree for provider bf95157b-a274-42de-9ccf-9851128a44a1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 2357.799014] env[62824]: DEBUG nova.compute.manager [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c70880c9-a69a-49b9-b028-189c04276623] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2357.826038] env[62824]: DEBUG oslo_vmware.api [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146892, 'name': MoveVirtualDisk_Task} progress is 24%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2358.041620] env[62824]: DEBUG oslo_vmware.api [None req-2d178aa0-6a19-48fd-a5cc-3661ef7c3914 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146878, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2358.265464] env[62824]: DEBUG nova.scheduler.client.report [None req-43cd9d71-795f-434c-b349-420c672b26c6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Updated inventory for provider bf95157b-a274-42de-9ccf-9851128a44a1 with generation 138 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 2358.265737] env[62824]: DEBUG nova.compute.provider_tree [None req-43cd9d71-795f-434c-b349-420c672b26c6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Updating resource provider bf95157b-a274-42de-9ccf-9851128a44a1 generation from 138 to 139 during operation: update_inventory {{(pid=62824) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 2358.265923] env[62824]: DEBUG nova.compute.provider_tree [None req-43cd9d71-795f-434c-b349-420c672b26c6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Updating inventory in ProviderTree for provider bf95157b-a274-42de-9ccf-9851128a44a1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 2358.321143] env[62824]: DEBUG oslo_concurrency.lockutils [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2358.324940] env[62824]: DEBUG oslo_vmware.api [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146892, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2358.541880] env[62824]: DEBUG oslo_vmware.api [None req-2d178aa0-6a19-48fd-a5cc-3661ef7c3914 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146878, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2358.553072] env[62824]: INFO nova.compute.manager [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Rebuilding instance [ 2358.597106] env[62824]: DEBUG nova.compute.manager [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2358.598071] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12f92557-67aa-424d-a143-ca94ab6b6548 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2358.773054] env[62824]: DEBUG oslo_concurrency.lockutils [None req-43cd9d71-795f-434c-b349-420c672b26c6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.667s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2358.774727] env[62824]: DEBUG oslo_concurrency.lockutils [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.454s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2358.776367] env[62824]: INFO nova.compute.claims [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c70880c9-a69a-49b9-b028-189c04276623] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2358.795383] env[62824]: INFO nova.scheduler.client.report [None req-43cd9d71-795f-434c-b349-420c672b26c6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Deleted allocations for instance bf6e681d-b773-4787-886a-c90586bc81bf [ 2358.825469] env[62824]: DEBUG oslo_vmware.api [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146892, 'name': MoveVirtualDisk_Task} progress is 66%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2359.042329] env[62824]: DEBUG oslo_vmware.api [None req-2d178aa0-6a19-48fd-a5cc-3661ef7c3914 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146878, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2359.302438] env[62824]: DEBUG oslo_concurrency.lockutils [None req-43cd9d71-795f-434c-b349-420c672b26c6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "bf6e681d-b773-4787-886a-c90586bc81bf" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.768s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2359.325907] env[62824]: DEBUG oslo_vmware.api [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146892, 'name': MoveVirtualDisk_Task} progress is 88%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2359.542177] env[62824]: DEBUG oslo_vmware.api [None req-2d178aa0-6a19-48fd-a5cc-3661ef7c3914 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146878, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2359.613598] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2359.613909] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a628973e-18f5-4ea5-96bb-79d3cdb040c4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2359.623066] env[62824]: DEBUG oslo_vmware.api [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2359.623066] env[62824]: value = "task-2146893" [ 2359.623066] env[62824]: _type = "Task" [ 2359.623066] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2359.632734] env[62824]: DEBUG oslo_vmware.api [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146893, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2359.828261] env[62824]: DEBUG oslo_vmware.api [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146892, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.28823} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2359.828614] env[62824]: INFO nova.virt.vmwareapi.ds_util [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_bb242f32-2f4c-4ad0-a044-984e9dd0b356/OSTACK_IMG_bb242f32-2f4c-4ad0-a044-984e9dd0b356.vmdk to [datastore2] devstack-image-cache_base/91de84ba-a43c-4db9-92a7-d2e20dada664/91de84ba-a43c-4db9-92a7-d2e20dada664.vmdk. [ 2359.828723] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Cleaning up location [datastore2] OSTACK_IMG_bb242f32-2f4c-4ad0-a044-984e9dd0b356 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 2359.828890] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_bb242f32-2f4c-4ad0-a044-984e9dd0b356 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2359.829169] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8a8908b8-7b2d-426a-b822-ad6d4530059e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2359.836076] env[62824]: DEBUG oslo_vmware.api [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Waiting for the task: (returnval){ [ 2359.836076] env[62824]: value = "task-2146894" [ 2359.836076] env[62824]: _type = "Task" [ 2359.836076] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2359.844836] env[62824]: DEBUG oslo_vmware.api [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146894, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2359.962415] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdd0e7cf-02aa-4ed6-872a-6e667f8bc7c4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2359.970306] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03b536aa-2c98-4e8c-9d30-e27d5cdf4a6c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2360.000286] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20cd86cb-699d-4351-acea-c6202ad56935 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2360.009995] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e08410ee-cf8a-44a1-af64-455385355e91 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2360.024905] env[62824]: DEBUG nova.compute.provider_tree [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Updating inventory in ProviderTree for provider bf95157b-a274-42de-9ccf-9851128a44a1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 2360.042745] env[62824]: DEBUG oslo_vmware.api [None req-2d178aa0-6a19-48fd-a5cc-3661ef7c3914 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146878, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2360.134085] env[62824]: DEBUG oslo_vmware.api [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146893, 'name': PowerOffVM_Task, 'duration_secs': 0.220771} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2360.134085] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2360.188400] env[62824]: INFO nova.compute.manager [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Detaching volume 42b59773-27a4-492d-b430-c629b14a2500 [ 2360.218656] env[62824]: INFO nova.virt.block_device [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Attempting to driver detach volume 42b59773-27a4-492d-b430-c629b14a2500 from mountpoint /dev/sdb [ 2360.218897] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Volume detach. Driver type: vmdk {{(pid=62824) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 2360.219099] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-438761', 'volume_id': '42b59773-27a4-492d-b430-c629b14a2500', 'name': 'volume-42b59773-27a4-492d-b430-c629b14a2500', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '3d294ee9-92f6-4115-95db-9d4b66562b56', 'attached_at': '', 'detached_at': '', 'volume_id': '42b59773-27a4-492d-b430-c629b14a2500', 'serial': '42b59773-27a4-492d-b430-c629b14a2500'} {{(pid=62824) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 2360.220024] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb31fd36-ef99-495e-9201-81dc19ca1992 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2360.243581] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-466a5b3b-1aaf-43ff-b6fb-4e8bbd1193eb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2360.250110] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41194580-b628-4215-bb71-856ef553c0c3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2360.271047] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3be60af-2c05-4f33-a32f-6cb031d84953 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2360.286709] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] The volume has not been displaced from its original location: [datastore2] volume-42b59773-27a4-492d-b430-c629b14a2500/volume-42b59773-27a4-492d-b430-c629b14a2500.vmdk. No consolidation needed. {{(pid=62824) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 2360.292064] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Reconfiguring VM instance instance-00000066 to detach disk 2001 {{(pid=62824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 2360.292379] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ee8940e1-fec6-44af-ba16-a0d745ad5f35 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2360.313108] env[62824]: DEBUG oslo_vmware.api [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2360.313108] env[62824]: value = "task-2146895" [ 2360.313108] env[62824]: _type = "Task" [ 2360.313108] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2360.346349] env[62824]: DEBUG oslo_vmware.api [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146894, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.039553} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2360.346815] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2360.347063] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Releasing lock "[datastore2] devstack-image-cache_base/91de84ba-a43c-4db9-92a7-d2e20dada664/91de84ba-a43c-4db9-92a7-d2e20dada664.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2360.347254] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/91de84ba-a43c-4db9-92a7-d2e20dada664/91de84ba-a43c-4db9-92a7-d2e20dada664.vmdk to [datastore2] 71e9733b-c84b-4501-8faf-a487f089b498/71e9733b-c84b-4501-8faf-a487f089b498.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2360.347512] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-73256a36-416b-442f-9bf9-65beb920f6d4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2360.375205] env[62824]: DEBUG oslo_vmware.api [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Waiting for the task: (returnval){ [ 2360.375205] env[62824]: value = "task-2146896" [ 2360.375205] env[62824]: _type = "Task" [ 2360.375205] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2360.383972] env[62824]: DEBUG oslo_vmware.api [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146896, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2360.546932] env[62824]: DEBUG oslo_vmware.api [None req-2d178aa0-6a19-48fd-a5cc-3661ef7c3914 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146878, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2360.559561] env[62824]: DEBUG nova.scheduler.client.report [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Updated inventory for provider bf95157b-a274-42de-9ccf-9851128a44a1 with generation 139 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 2360.559856] env[62824]: DEBUG nova.compute.provider_tree [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Updating resource provider bf95157b-a274-42de-9ccf-9851128a44a1 generation from 139 to 140 during operation: update_inventory {{(pid=62824) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 2360.560076] env[62824]: DEBUG nova.compute.provider_tree [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Updating inventory in ProviderTree for provider bf95157b-a274-42de-9ccf-9851128a44a1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 2360.823942] env[62824]: DEBUG oslo_vmware.api [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146895, 'name': ReconfigVM_Task, 'duration_secs': 0.272833} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2360.824268] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Reconfigured VM instance instance-00000066 to detach disk 2001 {{(pid=62824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 2360.829227] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a3c8c247-60eb-4c52-9a5d-65f7b36970f2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2360.847374] env[62824]: DEBUG oslo_vmware.api [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2360.847374] env[62824]: value = "task-2146897" [ 2360.847374] env[62824]: _type = "Task" [ 2360.847374] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2360.857689] env[62824]: DEBUG oslo_vmware.api [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146897, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2360.885863] env[62824]: DEBUG oslo_vmware.api [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146896, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2361.046300] env[62824]: DEBUG oslo_vmware.api [None req-2d178aa0-6a19-48fd-a5cc-3661ef7c3914 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146878, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2361.065602] env[62824]: DEBUG oslo_concurrency.lockutils [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.291s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2361.066252] env[62824]: DEBUG nova.compute.manager [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c70880c9-a69a-49b9-b028-189c04276623] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2361.095803] env[62824]: INFO nova.compute.manager [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Rebuilding instance [ 2361.147649] env[62824]: DEBUG nova.compute.manager [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2361.148590] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c725414-41d8-43f2-8b75-fb5ee49003c8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2361.358551] env[62824]: DEBUG oslo_vmware.api [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146897, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2361.386560] env[62824]: DEBUG oslo_vmware.api [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146896, 'name': CopyVirtualDisk_Task} progress is 43%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2361.546325] env[62824]: DEBUG oslo_vmware.api [None req-2d178aa0-6a19-48fd-a5cc-3661ef7c3914 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146878, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2361.572424] env[62824]: DEBUG nova.compute.utils [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2361.574091] env[62824]: DEBUG nova.compute.manager [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c70880c9-a69a-49b9-b028-189c04276623] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2361.574334] env[62824]: DEBUG nova.network.neutron [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c70880c9-a69a-49b9-b028-189c04276623] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2361.619355] env[62824]: DEBUG nova.policy [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cb36c701db8a471bbd65d2b68ce3becf', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c95aa87980de4ec8a2170c2640423301', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2361.858613] env[62824]: DEBUG oslo_vmware.api [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146897, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2361.889502] env[62824]: DEBUG oslo_vmware.api [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146896, 'name': CopyVirtualDisk_Task} progress is 63%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2361.908825] env[62824]: DEBUG nova.network.neutron [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c70880c9-a69a-49b9-b028-189c04276623] Successfully created port: f71d9cbc-c160-47f9-9771-af38641d3683 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2362.047640] env[62824]: DEBUG oslo_vmware.api [None req-2d178aa0-6a19-48fd-a5cc-3661ef7c3914 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146878, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2362.077496] env[62824]: DEBUG nova.compute.manager [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c70880c9-a69a-49b9-b028-189c04276623] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2362.164559] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2362.164957] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-497bf598-4f64-4887-b117-6e3cc47cad4b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2362.174686] env[62824]: DEBUG oslo_vmware.api [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2362.174686] env[62824]: value = "task-2146898" [ 2362.174686] env[62824]: _type = "Task" [ 2362.174686] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2362.185327] env[62824]: DEBUG oslo_vmware.api [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2146898, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2362.361063] env[62824]: DEBUG oslo_vmware.api [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146897, 'name': ReconfigVM_Task, 'duration_secs': 1.366082} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2362.361413] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-438761', 'volume_id': '42b59773-27a4-492d-b430-c629b14a2500', 'name': 'volume-42b59773-27a4-492d-b430-c629b14a2500', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '3d294ee9-92f6-4115-95db-9d4b66562b56', 'attached_at': '', 'detached_at': '', 'volume_id': '42b59773-27a4-492d-b430-c629b14a2500', 'serial': '42b59773-27a4-492d-b430-c629b14a2500'} {{(pid=62824) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 2362.391054] env[62824]: DEBUG oslo_vmware.api [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146896, 'name': CopyVirtualDisk_Task} progress is 85%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2362.504708] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2362.549161] env[62824]: DEBUG oslo_vmware.api [None req-2d178aa0-6a19-48fd-a5cc-3661ef7c3914 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146878, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2362.686358] env[62824]: DEBUG oslo_vmware.api [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2146898, 'name': PowerOffVM_Task, 'duration_secs': 0.260135} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2362.686718] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2362.686930] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2362.687803] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-659ea462-4f32-4df1-900e-91e5ec7ad382 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2362.697142] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2362.697440] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7f58f2db-3d40-4fb0-bca2-514c7fafb5ea {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2362.889486] env[62824]: DEBUG oslo_vmware.api [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146896, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.399451} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2362.889771] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/91de84ba-a43c-4db9-92a7-d2e20dada664/91de84ba-a43c-4db9-92a7-d2e20dada664.vmdk to [datastore2] 71e9733b-c84b-4501-8faf-a487f089b498/71e9733b-c84b-4501-8faf-a487f089b498.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2362.890752] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ece928bc-6695-4d0a-80d6-3594d2ca492e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2362.915046] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Reconfiguring VM instance instance-0000005c to attach disk [datastore2] 71e9733b-c84b-4501-8faf-a487f089b498/71e9733b-c84b-4501-8faf-a487f089b498.vmdk or device None with type streamOptimized {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2362.915046] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8aa967c2-c73a-421c-ac47-cf2cf882ac2e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2362.938799] env[62824]: DEBUG oslo_vmware.api [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Waiting for the task: (returnval){ [ 2362.938799] env[62824]: value = "task-2146900" [ 2362.938799] env[62824]: _type = "Task" [ 2362.938799] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2362.947784] env[62824]: DEBUG oslo_vmware.api [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146900, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2363.012050] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._cleanup_running_deleted_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2363.012253] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-41517485-76de-4e8f-a181-a012faec848d None None] Getting list of instances from cluster (obj){ [ 2363.012253] env[62824]: value = "domain-c8" [ 2363.012253] env[62824]: _type = "ClusterComputeResource" [ 2363.012253] env[62824]: } {{(pid=62824) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 2363.013553] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a13c00e4-3b1f-4757-9fb5-d40c85586619 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2363.034720] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-41517485-76de-4e8f-a181-a012faec848d None None] Got total of 11 instances {{(pid=62824) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 2363.046553] env[62824]: DEBUG oslo_vmware.api [None req-2d178aa0-6a19-48fd-a5cc-3661ef7c3914 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146878, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2363.087732] env[62824]: DEBUG nova.compute.manager [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c70880c9-a69a-49b9-b028-189c04276623] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2363.113553] env[62824]: DEBUG nova.virt.hardware [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2363.113800] env[62824]: DEBUG nova.virt.hardware [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2363.113962] env[62824]: DEBUG nova.virt.hardware [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2363.114165] env[62824]: DEBUG nova.virt.hardware [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2363.114312] env[62824]: DEBUG nova.virt.hardware [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2363.114458] env[62824]: DEBUG nova.virt.hardware [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2363.114664] env[62824]: DEBUG nova.virt.hardware [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2363.114826] env[62824]: DEBUG nova.virt.hardware [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2363.114996] env[62824]: DEBUG nova.virt.hardware [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2363.115182] env[62824]: DEBUG nova.virt.hardware [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2363.115356] env[62824]: DEBUG nova.virt.hardware [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2363.116239] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2b15e24-f53c-482b-b962-26e2bd624547 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2363.124943] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8b410f3-8344-49cc-aba9-295d81243d77 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2363.434892] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2363.435254] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-873d0e09-8b77-4d76-b058-24fbc0fbfea6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2363.443853] env[62824]: DEBUG oslo_vmware.api [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2363.443853] env[62824]: value = "task-2146901" [ 2363.443853] env[62824]: _type = "Task" [ 2363.443853] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2363.451263] env[62824]: DEBUG oslo_vmware.api [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146900, 'name': ReconfigVM_Task, 'duration_secs': 0.272672} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2363.452107] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Reconfigured VM instance instance-0000005c to attach disk [datastore2] 71e9733b-c84b-4501-8faf-a487f089b498/71e9733b-c84b-4501-8faf-a487f089b498.vmdk or device None with type streamOptimized {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2363.452775] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2eeed9f2-a0f3-404b-b821-be83b55187da {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2363.457867] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] VM already powered off {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 2363.458103] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Volume detach. Driver type: vmdk {{(pid=62824) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 2363.458320] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-438761', 'volume_id': '42b59773-27a4-492d-b430-c629b14a2500', 'name': 'volume-42b59773-27a4-492d-b430-c629b14a2500', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '3d294ee9-92f6-4115-95db-9d4b66562b56', 'attached_at': '', 'detached_at': '', 'volume_id': '42b59773-27a4-492d-b430-c629b14a2500', 'serial': '42b59773-27a4-492d-b430-c629b14a2500'} {{(pid=62824) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 2363.459083] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d21a4a9d-9455-40a5-b99d-7796f2b0d562 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2363.463141] env[62824]: DEBUG oslo_vmware.api [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Waiting for the task: (returnval){ [ 2363.463141] env[62824]: value = "task-2146902" [ 2363.463141] env[62824]: _type = "Task" [ 2363.463141] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2363.479767] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fe1a4b3-0fb4-4116-9369-d485d0a2f621 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2363.485711] env[62824]: DEBUG oslo_vmware.api [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146902, 'name': Rename_Task} progress is 10%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2363.489899] env[62824]: WARNING nova.virt.vmwareapi.driver [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] The volume None does not exist!: nova.exception.DiskNotFound: Unable to find volume [ 2363.490217] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2363.491008] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-202b43c6-e291-4905-bbad-7bf7ebf0c3dc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2363.498027] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2363.498109] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9fbb251c-556c-4cbc-a213-160b3ae04d58 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2363.551154] env[62824]: DEBUG oslo_vmware.api [None req-2d178aa0-6a19-48fd-a5cc-3661ef7c3914 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146878, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2363.973783] env[62824]: DEBUG oslo_vmware.api [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146902, 'name': Rename_Task, 'duration_secs': 0.244532} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2363.974135] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2363.974315] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1e37f439-4fe7-4363-9160-2bd8ecab3549 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2363.982347] env[62824]: DEBUG oslo_vmware.api [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Waiting for the task: (returnval){ [ 2363.982347] env[62824]: value = "task-2146904" [ 2363.982347] env[62824]: _type = "Task" [ 2363.982347] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2363.991806] env[62824]: DEBUG oslo_vmware.api [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146904, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2364.049640] env[62824]: DEBUG oslo_vmware.api [None req-2d178aa0-6a19-48fd-a5cc-3661ef7c3914 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146878, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2364.493405] env[62824]: DEBUG oslo_vmware.api [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146904, 'name': PowerOnVM_Task} progress is 95%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2364.551234] env[62824]: DEBUG oslo_vmware.api [None req-2d178aa0-6a19-48fd-a5cc-3661ef7c3914 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146878, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2364.836418] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2364.836650] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Deleting contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2364.836839] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Deleting the datastore file [datastore1] 1db9ff97-e59c-43f6-984e-6405d8e993a4 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2364.837234] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9be3cc29-e34c-4358-8e7b-82b6576c6cd1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2364.840387] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2364.840644] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Deleting contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2364.840897] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Deleting the datastore file [datastore2] 3d294ee9-92f6-4115-95db-9d4b66562b56 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2364.841578] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-647e9656-8d4a-4461-971f-c24288edcdc5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2364.848658] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-9c8f0adb-9b23-432f-8854-cd75806cadd6 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2364.848968] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-9c8f0adb-9b23-432f-8854-cd75806cadd6 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Deleting contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2364.849246] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-9c8f0adb-9b23-432f-8854-cd75806cadd6 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Deleting the datastore file [datastore1] c480f680-61fd-496b-8cf7-fa50f580b10d {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2364.851386] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e2dbf5fd-584f-42be-b12d-48d1fea1e3af {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2364.853279] env[62824]: DEBUG oslo_vmware.api [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2364.853279] env[62824]: value = "task-2146906" [ 2364.853279] env[62824]: _type = "Task" [ 2364.853279] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2364.853620] env[62824]: DEBUG oslo_vmware.api [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2364.853620] env[62824]: value = "task-2146905" [ 2364.853620] env[62824]: _type = "Task" [ 2364.853620] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2364.861106] env[62824]: DEBUG oslo_vmware.api [None req-9c8f0adb-9b23-432f-8854-cd75806cadd6 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for the task: (returnval){ [ 2364.861106] env[62824]: value = "task-2146907" [ 2364.861106] env[62824]: _type = "Task" [ 2364.861106] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2364.870492] env[62824]: DEBUG oslo_vmware.api [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2146905, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2364.870741] env[62824]: DEBUG oslo_vmware.api [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146906, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2364.875942] env[62824]: DEBUG oslo_vmware.api [None req-9c8f0adb-9b23-432f-8854-cd75806cadd6 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146907, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2364.993623] env[62824]: DEBUG oslo_vmware.api [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146904, 'name': PowerOnVM_Task, 'duration_secs': 0.553374} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2364.993996] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2364.998104] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-b70c134a-661f-4db0-98ce-919ec5cda15d tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: fb9cfb53-285a-4586-acf0-cb08d56b93e7] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2364.998364] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-b70c134a-661f-4db0-98ce-919ec5cda15d tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: fb9cfb53-285a-4586-acf0-cb08d56b93e7] Deleting contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2364.998516] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-b70c134a-661f-4db0-98ce-919ec5cda15d tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Deleting the datastore file [datastore2] fb9cfb53-285a-4586-acf0-cb08d56b93e7 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2364.998779] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ab6c0842-b835-4d59-9a45-e793ea973a44 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2365.022028] env[62824]: DEBUG oslo_vmware.api [None req-b70c134a-661f-4db0-98ce-919ec5cda15d tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2365.022028] env[62824]: value = "task-2146908" [ 2365.022028] env[62824]: _type = "Task" [ 2365.022028] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2365.032687] env[62824]: DEBUG oslo_vmware.api [None req-b70c134a-661f-4db0-98ce-919ec5cda15d tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146908, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2365.051995] env[62824]: DEBUG oslo_vmware.api [None req-2d178aa0-6a19-48fd-a5cc-3661ef7c3914 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146878, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2365.130373] env[62824]: DEBUG nova.compute.manager [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2365.132622] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c1eb597-72e3-4509-9f3c-aae67a505909 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2365.370886] env[62824]: DEBUG oslo_vmware.api [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146906, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.164176} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2365.371168] env[62824]: DEBUG oslo_vmware.api [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2146905, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.178212} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2365.371707] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2365.371918] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Deleted contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2365.372111] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2365.374509] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2365.374693] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Deleted contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2365.374870] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2365.379714] env[62824]: DEBUG oslo_vmware.api [None req-9c8f0adb-9b23-432f-8854-cd75806cadd6 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Task: {'id': task-2146907, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.168878} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2365.380119] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-9c8f0adb-9b23-432f-8854-cd75806cadd6 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2365.380303] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-9c8f0adb-9b23-432f-8854-cd75806cadd6 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Deleted contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2365.380478] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-9c8f0adb-9b23-432f-8854-cd75806cadd6 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2365.380648] env[62824]: INFO nova.compute.manager [None req-9c8f0adb-9b23-432f-8854-cd75806cadd6 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Took 9.60 seconds to destroy the instance on the hypervisor. [ 2365.380901] env[62824]: DEBUG oslo.service.loopingcall [None req-9c8f0adb-9b23-432f-8854-cd75806cadd6 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2365.381109] env[62824]: DEBUG nova.compute.manager [-] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2365.381206] env[62824]: DEBUG nova.network.neutron [-] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2365.532448] env[62824]: DEBUG oslo_vmware.api [None req-b70c134a-661f-4db0-98ce-919ec5cda15d tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146908, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.167741} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2365.532701] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-b70c134a-661f-4db0-98ce-919ec5cda15d tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2365.532886] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-b70c134a-661f-4db0-98ce-919ec5cda15d tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: fb9cfb53-285a-4586-acf0-cb08d56b93e7] Deleted contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2365.533081] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-b70c134a-661f-4db0-98ce-919ec5cda15d tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: fb9cfb53-285a-4586-acf0-cb08d56b93e7] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2365.533268] env[62824]: INFO nova.compute.manager [None req-b70c134a-661f-4db0-98ce-919ec5cda15d tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: fb9cfb53-285a-4586-acf0-cb08d56b93e7] Took 10.63 seconds to destroy the instance on the hypervisor. [ 2365.533503] env[62824]: DEBUG oslo.service.loopingcall [None req-b70c134a-661f-4db0-98ce-919ec5cda15d tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2365.533692] env[62824]: DEBUG nova.compute.manager [-] [instance: fb9cfb53-285a-4586-acf0-cb08d56b93e7] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2365.533788] env[62824]: DEBUG nova.network.neutron [-] [instance: fb9cfb53-285a-4586-acf0-cb08d56b93e7] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2365.550485] env[62824]: DEBUG oslo_vmware.api [None req-2d178aa0-6a19-48fd-a5cc-3661ef7c3914 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146878, 'name': ReconfigVM_Task, 'duration_secs': 14.189987} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2365.550717] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2d178aa0-6a19-48fd-a5cc-3661ef7c3914 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Releasing lock "e1574266-e53b-4169-9db6-c66b84895edb" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2365.550994] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-2d178aa0-6a19-48fd-a5cc-3661ef7c3914 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Reconfigured VM to detach interface {{(pid=62824) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 2365.651971] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f45dfa2e-cb2e-4aef-996d-90c94e97321a tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Lock "71e9733b-c84b-4501-8faf-a487f089b498" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 20.647s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2365.718391] env[62824]: DEBUG nova.compute.manager [req-9fe7d255-977d-40d9-a90d-54def8dd3287 req-7385dfa7-93d5-42bc-8fd6-dfbb9d4690f2 service nova] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Received event network-vif-deleted-9190a574-76bc-4236-9f09-69d08897d634 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2365.718637] env[62824]: INFO nova.compute.manager [req-9fe7d255-977d-40d9-a90d-54def8dd3287 req-7385dfa7-93d5-42bc-8fd6-dfbb9d4690f2 service nova] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Neutron deleted interface 9190a574-76bc-4236-9f09-69d08897d634; detaching it from the instance and deleting it from the info cache [ 2365.718755] env[62824]: DEBUG nova.network.neutron [req-9fe7d255-977d-40d9-a90d-54def8dd3287 req-7385dfa7-93d5-42bc-8fd6-dfbb9d4690f2 service nova] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2365.879593] env[62824]: INFO nova.virt.block_device [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Booting with volume 42b59773-27a4-492d-b430-c629b14a2500 at /dev/sdb [ 2365.923453] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-80352be0-952c-44eb-9bc6-9bec61a09bb5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2365.933547] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7e9783a-e79a-4d07-a809-4235344a1077 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2365.964351] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-79e1d247-0c7b-4d85-a584-ba12d27d5564 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2365.973996] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-630a14bc-79f1-4c50-9bfe-8e3452e137d1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2366.004886] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1dbc03c-87ac-426a-a4e3-585e58caaf0f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2366.012327] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-815d9bf0-1823-4080-87d4-c6e4838b7160 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2366.027258] env[62824]: DEBUG nova.virt.block_device [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Updating existing volume attachment record: 12710dd9-9937-4c84-8856-b560c4a74c35 {{(pid=62824) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 2366.194441] env[62824]: DEBUG nova.network.neutron [-] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2366.220770] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-65cdb86d-04a0-40af-b086-2b32066ae33e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2366.231124] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ed6c9af-4a54-4c20-b92f-bc2eebb0e266 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2366.266568] env[62824]: DEBUG nova.compute.manager [req-9fe7d255-977d-40d9-a90d-54def8dd3287 req-7385dfa7-93d5-42bc-8fd6-dfbb9d4690f2 service nova] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Detach interface failed, port_id=9190a574-76bc-4236-9f09-69d08897d634, reason: Instance c480f680-61fd-496b-8cf7-fa50f580b10d could not be found. {{(pid=62824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11482}} [ 2366.330840] env[62824]: DEBUG nova.network.neutron [-] [instance: fb9cfb53-285a-4586-acf0-cb08d56b93e7] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2366.414804] env[62824]: DEBUG nova.virt.hardware [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2366.415071] env[62824]: DEBUG nova.virt.hardware [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2366.415232] env[62824]: DEBUG nova.virt.hardware [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2366.415415] env[62824]: DEBUG nova.virt.hardware [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2366.415563] env[62824]: DEBUG nova.virt.hardware [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2366.415740] env[62824]: DEBUG nova.virt.hardware [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2366.415980] env[62824]: DEBUG nova.virt.hardware [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2366.416164] env[62824]: DEBUG nova.virt.hardware [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2366.416335] env[62824]: DEBUG nova.virt.hardware [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2366.416499] env[62824]: DEBUG nova.virt.hardware [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2366.416687] env[62824]: DEBUG nova.virt.hardware [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2366.417540] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64f869ab-4ac9-4387-9ac9-2c29f18204d2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2366.427498] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0810a481-bf7b-4948-a235-b25016499fbc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2366.441924] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:43:2e:10', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b49c5024-2ced-42ca-90cc-6066766d43e6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'aeeadfdd-d869-4555-bda2-391c435de1bb', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2366.449300] env[62824]: DEBUG oslo.service.loopingcall [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2366.449554] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2366.449770] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5bdb402f-e496-4864-bf81-1e2432e387d2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2366.470442] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2366.470442] env[62824]: value = "task-2146909" [ 2366.470442] env[62824]: _type = "Task" [ 2366.470442] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2366.478660] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146909, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2366.513814] env[62824]: DEBUG nova.compute.manager [req-191c8cbb-4aaf-4e0f-9a76-88e9c2ecae20 req-9df775a0-4d2f-47c6-9ec2-86e9a0bc1861 service nova] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Received event network-vif-plugged-edaf5d3c-004d-4ba0-b037-e7eef2acbe44 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2366.514122] env[62824]: DEBUG oslo_concurrency.lockutils [req-191c8cbb-4aaf-4e0f-9a76-88e9c2ecae20 req-9df775a0-4d2f-47c6-9ec2-86e9a0bc1861 service nova] Acquiring lock "21a7eed6-c146-4911-9a0a-a1e1644e8e9d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2366.514377] env[62824]: DEBUG oslo_concurrency.lockutils [req-191c8cbb-4aaf-4e0f-9a76-88e9c2ecae20 req-9df775a0-4d2f-47c6-9ec2-86e9a0bc1861 service nova] Lock "21a7eed6-c146-4911-9a0a-a1e1644e8e9d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2366.514604] env[62824]: DEBUG oslo_concurrency.lockutils [req-191c8cbb-4aaf-4e0f-9a76-88e9c2ecae20 req-9df775a0-4d2f-47c6-9ec2-86e9a0bc1861 service nova] Lock "21a7eed6-c146-4911-9a0a-a1e1644e8e9d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2366.514908] env[62824]: DEBUG nova.compute.manager [req-191c8cbb-4aaf-4e0f-9a76-88e9c2ecae20 req-9df775a0-4d2f-47c6-9ec2-86e9a0bc1861 service nova] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] No waiting events found dispatching network-vif-plugged-edaf5d3c-004d-4ba0-b037-e7eef2acbe44 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2366.515112] env[62824]: WARNING nova.compute.manager [req-191c8cbb-4aaf-4e0f-9a76-88e9c2ecae20 req-9df775a0-4d2f-47c6-9ec2-86e9a0bc1861 service nova] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Received unexpected event network-vif-plugged-edaf5d3c-004d-4ba0-b037-e7eef2acbe44 for instance with vm_state building and task_state spawning. [ 2366.609393] env[62824]: DEBUG nova.network.neutron [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Successfully updated port: edaf5d3c-004d-4ba0-b037-e7eef2acbe44 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2366.699734] env[62824]: INFO nova.compute.manager [-] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Took 1.32 seconds to deallocate network for instance. [ 2366.838139] env[62824]: INFO nova.compute.manager [-] [instance: fb9cfb53-285a-4586-acf0-cb08d56b93e7] Took 1.30 seconds to deallocate network for instance. [ 2366.856158] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2d178aa0-6a19-48fd-a5cc-3661ef7c3914 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquiring lock "refresh_cache-e1574266-e53b-4169-9db6-c66b84895edb" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2366.856449] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2d178aa0-6a19-48fd-a5cc-3661ef7c3914 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquired lock "refresh_cache-e1574266-e53b-4169-9db6-c66b84895edb" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2366.856701] env[62824]: DEBUG nova.network.neutron [None req-2d178aa0-6a19-48fd-a5cc-3661ef7c3914 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2366.935142] env[62824]: DEBUG nova.network.neutron [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c70880c9-a69a-49b9-b028-189c04276623] Successfully updated port: f71d9cbc-c160-47f9-9771-af38641d3683 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2366.981155] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146909, 'name': CreateVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2367.116737] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquiring lock "refresh_cache-21a7eed6-c146-4911-9a0a-a1e1644e8e9d" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2367.116737] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquired lock "refresh_cache-21a7eed6-c146-4911-9a0a-a1e1644e8e9d" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2367.117223] env[62824]: DEBUG nova.network.neutron [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2367.208306] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9c8f0adb-9b23-432f-8854-cd75806cadd6 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2367.208632] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9c8f0adb-9b23-432f-8854-cd75806cadd6 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2367.208842] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9c8f0adb-9b23-432f-8854-cd75806cadd6 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2367.234051] env[62824]: INFO nova.scheduler.client.report [None req-9c8f0adb-9b23-432f-8854-cd75806cadd6 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Deleted allocations for instance c480f680-61fd-496b-8cf7-fa50f580b10d [ 2367.347819] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b70c134a-661f-4db0-98ce-919ec5cda15d tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2367.348130] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b70c134a-661f-4db0-98ce-919ec5cda15d tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2367.348373] env[62824]: DEBUG nova.objects.instance [None req-b70c134a-661f-4db0-98ce-919ec5cda15d tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lazy-loading 'resources' on Instance uuid fb9cfb53-285a-4586-acf0-cb08d56b93e7 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2367.380166] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61c85310-e1a5-4526-a359-b912af0161d2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2367.388244] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-b29395ee-c43c-48ba-8452-8c3a08c40767 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Suspending the VM {{(pid=62824) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 2367.388524] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-200cefe3-e6bd-4a3f-9240-2425de2ff1cc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2367.396417] env[62824]: DEBUG oslo_vmware.api [None req-b29395ee-c43c-48ba-8452-8c3a08c40767 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Waiting for the task: (returnval){ [ 2367.396417] env[62824]: value = "task-2146910" [ 2367.396417] env[62824]: _type = "Task" [ 2367.396417] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2367.407417] env[62824]: DEBUG oslo_vmware.api [None req-b29395ee-c43c-48ba-8452-8c3a08c40767 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146910, 'name': SuspendVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2367.439707] env[62824]: DEBUG oslo_concurrency.lockutils [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Acquiring lock "refresh_cache-c70880c9-a69a-49b9-b028-189c04276623" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2367.439852] env[62824]: DEBUG oslo_concurrency.lockutils [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Acquired lock "refresh_cache-c70880c9-a69a-49b9-b028-189c04276623" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2367.440013] env[62824]: DEBUG nova.network.neutron [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c70880c9-a69a-49b9-b028-189c04276623] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2367.484754] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146909, 'name': CreateVM_Task, 'duration_secs': 0.568607} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2367.485018] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2367.485753] env[62824]: DEBUG oslo_concurrency.lockutils [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2367.485930] env[62824]: DEBUG oslo_concurrency.lockutils [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2367.486279] env[62824]: DEBUG oslo_concurrency.lockutils [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2367.486552] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c6ddbb72-89ac-4332-88e2-e40c7170d9a5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2367.492288] env[62824]: DEBUG oslo_vmware.api [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2367.492288] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52120e68-6fb6-239f-7b3a-b36bf9c1bd2b" [ 2367.492288] env[62824]: _type = "Task" [ 2367.492288] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2367.501276] env[62824]: DEBUG oslo_vmware.api [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52120e68-6fb6-239f-7b3a-b36bf9c1bd2b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2367.694638] env[62824]: DEBUG nova.network.neutron [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2367.751249] env[62824]: DEBUG nova.compute.manager [req-ca3c9e7a-f266-42e0-9b7e-86680edaa728 req-1c38bd66-e092-457c-bc53-605feab3ec4d service nova] [instance: fb9cfb53-285a-4586-acf0-cb08d56b93e7] Received event network-vif-deleted-5278435d-72ae-4fec-844c-0f1174154e30 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2367.751952] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9c8f0adb-9b23-432f-8854-cd75806cadd6 tempest-DeleteServersTestJSON-1705475434 tempest-DeleteServersTestJSON-1705475434-project-member] Lock "c480f680-61fd-496b-8cf7-fa50f580b10d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.477s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2367.828539] env[62824]: INFO nova.network.neutron [None req-2d178aa0-6a19-48fd-a5cc-3661ef7c3914 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Port afcd2bdb-2bfc-40d7-a8e5-2a417be0e0e9 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 2367.828934] env[62824]: DEBUG nova.network.neutron [None req-2d178aa0-6a19-48fd-a5cc-3661ef7c3914 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Updating instance_info_cache with network_info: [{"id": "d5c6ba87-e0c7-49c2-bbaf-b7c64f84cdf8", "address": "fa:16:3e:c4:9c:f4", "network": {"id": "aa1c09e5-8d72-43ad-a903-a0f5e711da80", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1643616572-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c987c5f844be42c99cc9d57a9ca84c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9bc2632-36f9-4912-8782-8bbb789f909d", "external-id": "nsx-vlan-transportzone-897", "segmentation_id": 897, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd5c6ba87-e0", "ovs_interfaceid": "d5c6ba87-e0c7-49c2-bbaf-b7c64f84cdf8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2367.906959] env[62824]: DEBUG oslo_vmware.api [None req-b29395ee-c43c-48ba-8452-8c3a08c40767 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146910, 'name': SuspendVM_Task} progress is 62%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2368.006839] env[62824]: DEBUG oslo_vmware.api [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52120e68-6fb6-239f-7b3a-b36bf9c1bd2b, 'name': SearchDatastore_Task, 'duration_secs': 0.046315} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2368.007946] env[62824]: DEBUG nova.network.neutron [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c70880c9-a69a-49b9-b028-189c04276623] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2368.011054] env[62824]: DEBUG oslo_concurrency.lockutils [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2368.011339] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2368.011635] env[62824]: DEBUG oslo_concurrency.lockutils [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2368.011852] env[62824]: DEBUG oslo_concurrency.lockutils [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2368.012039] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2368.012925] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3f57f750-3570-457d-8756-f6154b4a428b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2368.026057] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2368.026294] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2368.028023] env[62824]: DEBUG nova.network.neutron [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Updating instance_info_cache with network_info: [{"id": "edaf5d3c-004d-4ba0-b037-e7eef2acbe44", "address": "fa:16:3e:06:2e:6a", "network": {"id": "12411632-b7f6-4544-b2e4-b04ae7337547", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-917106830-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4479921caf5f405b8fc49baad390a0e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "de7fa486-5f28-44ae-b0cf-72234ff87546", "external-id": "nsx-vlan-transportzone-229", "segmentation_id": 229, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapedaf5d3c-00", "ovs_interfaceid": "edaf5d3c-004d-4ba0-b037-e7eef2acbe44", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2368.029632] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4aeb29f9-3a3c-4cb1-8b5a-4be0741c6e67 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2368.038992] env[62824]: DEBUG oslo_vmware.api [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2368.038992] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5257f0b4-8513-3718-8caa-d8e7acf312ad" [ 2368.038992] env[62824]: _type = "Task" [ 2368.038992] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2368.047776] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f3de61a-9053-499d-ad7d-6cac954ace09 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2368.062119] env[62824]: DEBUG oslo_vmware.api [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5257f0b4-8513-3718-8caa-d8e7acf312ad, 'name': SearchDatastore_Task, 'duration_secs': 0.010378} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2368.063683] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef1d1ffd-eb55-4231-a504-f693160dcfff {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2368.067166] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cc856a29-54c9-462d-9b9a-a9bea4a1b334 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2368.075216] env[62824]: DEBUG oslo_vmware.api [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2368.075216] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52e50c9f-243a-c1da-1730-f42c0fbb4aa6" [ 2368.075216] env[62824]: _type = "Task" [ 2368.075216] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2368.105792] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fea8e697-6876-4af9-be73-97ef2bf63c20 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2368.118354] env[62824]: DEBUG oslo_vmware.api [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52e50c9f-243a-c1da-1730-f42c0fbb4aa6, 'name': SearchDatastore_Task, 'duration_secs': 0.00985} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2368.119776] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efa6cc00-4d3a-4d49-a766-ce47fc511a88 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2368.123681] env[62824]: DEBUG oslo_concurrency.lockutils [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2368.123956] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 1db9ff97-e59c-43f6-984e-6405d8e993a4/1db9ff97-e59c-43f6-984e-6405d8e993a4.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2368.124239] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-54b5f31d-493a-45d5-a67b-de375e861c10 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2368.137212] env[62824]: DEBUG nova.compute.provider_tree [None req-b70c134a-661f-4db0-98ce-919ec5cda15d tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Updating inventory in ProviderTree for provider bf95157b-a274-42de-9ccf-9851128a44a1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 2368.142770] env[62824]: DEBUG oslo_vmware.api [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2368.142770] env[62824]: value = "task-2146911" [ 2368.142770] env[62824]: _type = "Task" [ 2368.142770] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2368.153318] env[62824]: DEBUG oslo_vmware.api [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2146911, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2368.186805] env[62824]: DEBUG nova.virt.hardware [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2368.187128] env[62824]: DEBUG nova.virt.hardware [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2368.187249] env[62824]: DEBUG nova.virt.hardware [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2368.187442] env[62824]: DEBUG nova.virt.hardware [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2368.190023] env[62824]: DEBUG nova.virt.hardware [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2368.190023] env[62824]: DEBUG nova.virt.hardware [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2368.190023] env[62824]: DEBUG nova.virt.hardware [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2368.190023] env[62824]: DEBUG nova.virt.hardware [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2368.190023] env[62824]: DEBUG nova.virt.hardware [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2368.190023] env[62824]: DEBUG nova.virt.hardware [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2368.190023] env[62824]: DEBUG nova.virt.hardware [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2368.190586] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d95611d-966a-40c5-941a-df22ea97b3b8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2368.198813] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf863d5e-8477-4851-8d85-83329cd680c2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2368.218734] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:17:9c:53', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '69f65356-c85e-4b7f-ad28-7c7b5e8cf50c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f1800205-8a92-425e-9f64-1dabd5809221', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2368.226237] env[62824]: DEBUG oslo.service.loopingcall [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2368.226850] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2368.227089] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-013cfa9f-5073-4b74-ab0d-632cd3637317 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2368.248081] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2368.248081] env[62824]: value = "task-2146912" [ 2368.248081] env[62824]: _type = "Task" [ 2368.248081] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2368.256017] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146912, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2368.331668] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2d178aa0-6a19-48fd-a5cc-3661ef7c3914 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Releasing lock "refresh_cache-e1574266-e53b-4169-9db6-c66b84895edb" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2368.409122] env[62824]: DEBUG oslo_vmware.api [None req-b29395ee-c43c-48ba-8452-8c3a08c40767 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146910, 'name': SuspendVM_Task, 'duration_secs': 0.679735} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2368.409373] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-b29395ee-c43c-48ba-8452-8c3a08c40767 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Suspended the VM {{(pid=62824) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 2368.409595] env[62824]: DEBUG nova.compute.manager [None req-b29395ee-c43c-48ba-8452-8c3a08c40767 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2368.414027] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1db1f2c-64a3-405c-b5dc-56c561793d1c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2368.499915] env[62824]: DEBUG nova.network.neutron [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c70880c9-a69a-49b9-b028-189c04276623] Updating instance_info_cache with network_info: [{"id": "f71d9cbc-c160-47f9-9771-af38641d3683", "address": "fa:16:3e:51:89:60", "network": {"id": "f7936d74-eeef-4c41-8ae7-9427c43663fa", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1643883257-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c95aa87980de4ec8a2170c2640423301", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6eaa481-1f92-4851-b98e-09ed0daad7cb", "external-id": "nsx-vlan-transportzone-636", "segmentation_id": 636, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf71d9cbc-c1", "ovs_interfaceid": "f71d9cbc-c160-47f9-9771-af38641d3683", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2368.533524] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Releasing lock "refresh_cache-21a7eed6-c146-4911-9a0a-a1e1644e8e9d" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2368.533861] env[62824]: DEBUG nova.compute.manager [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Instance network_info: |[{"id": "edaf5d3c-004d-4ba0-b037-e7eef2acbe44", "address": "fa:16:3e:06:2e:6a", "network": {"id": "12411632-b7f6-4544-b2e4-b04ae7337547", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-917106830-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4479921caf5f405b8fc49baad390a0e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "de7fa486-5f28-44ae-b0cf-72234ff87546", "external-id": "nsx-vlan-transportzone-229", "segmentation_id": 229, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapedaf5d3c-00", "ovs_interfaceid": "edaf5d3c-004d-4ba0-b037-e7eef2acbe44", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2368.534333] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:06:2e:6a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'de7fa486-5f28-44ae-b0cf-72234ff87546', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'edaf5d3c-004d-4ba0-b037-e7eef2acbe44', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2368.544392] env[62824]: DEBUG oslo.service.loopingcall [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2368.544569] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2368.544838] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3daa4c91-e9d2-46ca-87b1-a6f0612c1178 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2368.569350] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2368.569350] env[62824]: value = "task-2146913" [ 2368.569350] env[62824]: _type = "Task" [ 2368.569350] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2368.580259] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146913, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2368.621564] env[62824]: DEBUG nova.compute.manager [req-e1ac7983-616b-4c48-8b20-9cb7de8fb175 req-e5129aa6-20ea-4a2b-898a-fb517da29a21 service nova] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Received event network-changed-edaf5d3c-004d-4ba0-b037-e7eef2acbe44 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2368.622220] env[62824]: DEBUG nova.compute.manager [req-e1ac7983-616b-4c48-8b20-9cb7de8fb175 req-e5129aa6-20ea-4a2b-898a-fb517da29a21 service nova] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Refreshing instance network info cache due to event network-changed-edaf5d3c-004d-4ba0-b037-e7eef2acbe44. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2368.622220] env[62824]: DEBUG oslo_concurrency.lockutils [req-e1ac7983-616b-4c48-8b20-9cb7de8fb175 req-e5129aa6-20ea-4a2b-898a-fb517da29a21 service nova] Acquiring lock "refresh_cache-21a7eed6-c146-4911-9a0a-a1e1644e8e9d" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2368.623073] env[62824]: DEBUG oslo_concurrency.lockutils [req-e1ac7983-616b-4c48-8b20-9cb7de8fb175 req-e5129aa6-20ea-4a2b-898a-fb517da29a21 service nova] Acquired lock "refresh_cache-21a7eed6-c146-4911-9a0a-a1e1644e8e9d" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2368.623073] env[62824]: DEBUG nova.network.neutron [req-e1ac7983-616b-4c48-8b20-9cb7de8fb175 req-e5129aa6-20ea-4a2b-898a-fb517da29a21 service nova] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Refreshing network info cache for port edaf5d3c-004d-4ba0-b037-e7eef2acbe44 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2368.660824] env[62824]: DEBUG oslo_vmware.api [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2146911, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.503048} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2368.661496] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 1db9ff97-e59c-43f6-984e-6405d8e993a4/1db9ff97-e59c-43f6-984e-6405d8e993a4.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2368.661775] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2368.662016] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ddd55994-4d1d-45c0-9a65-c35098a8bd0a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2368.671637] env[62824]: DEBUG oslo_vmware.api [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2368.671637] env[62824]: value = "task-2146915" [ 2368.671637] env[62824]: _type = "Task" [ 2368.671637] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2368.677774] env[62824]: DEBUG nova.scheduler.client.report [None req-b70c134a-661f-4db0-98ce-919ec5cda15d tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Updated inventory for provider bf95157b-a274-42de-9ccf-9851128a44a1 with generation 140 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 2368.678218] env[62824]: DEBUG nova.compute.provider_tree [None req-b70c134a-661f-4db0-98ce-919ec5cda15d tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Updating resource provider bf95157b-a274-42de-9ccf-9851128a44a1 generation from 140 to 141 during operation: update_inventory {{(pid=62824) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 2368.678495] env[62824]: DEBUG nova.compute.provider_tree [None req-b70c134a-661f-4db0-98ce-919ec5cda15d tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Updating inventory in ProviderTree for provider bf95157b-a274-42de-9ccf-9851128a44a1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 2368.686027] env[62824]: DEBUG oslo_vmware.api [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2146915, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2368.760812] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146912, 'name': CreateVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2368.837842] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2d178aa0-6a19-48fd-a5cc-3661ef7c3914 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lock "interface-e1574266-e53b-4169-9db6-c66b84895edb-afcd2bdb-2bfc-40d7-a8e5-2a417be0e0e9" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 18.412s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2369.004324] env[62824]: DEBUG oslo_concurrency.lockutils [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Releasing lock "refresh_cache-c70880c9-a69a-49b9-b028-189c04276623" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2369.004690] env[62824]: DEBUG nova.compute.manager [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c70880c9-a69a-49b9-b028-189c04276623] Instance network_info: |[{"id": "f71d9cbc-c160-47f9-9771-af38641d3683", "address": "fa:16:3e:51:89:60", "network": {"id": "f7936d74-eeef-4c41-8ae7-9427c43663fa", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1643883257-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c95aa87980de4ec8a2170c2640423301", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6eaa481-1f92-4851-b98e-09ed0daad7cb", "external-id": "nsx-vlan-transportzone-636", "segmentation_id": 636, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf71d9cbc-c1", "ovs_interfaceid": "f71d9cbc-c160-47f9-9771-af38641d3683", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2369.005162] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c70880c9-a69a-49b9-b028-189c04276623] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:51:89:60', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c6eaa481-1f92-4851-b98e-09ed0daad7cb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f71d9cbc-c160-47f9-9771-af38641d3683', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2369.013187] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Creating folder: Project (c95aa87980de4ec8a2170c2640423301). Parent ref: group-v438503. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 2369.013765] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-086f1fa3-2137-4858-a726-0e96a46d606e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2369.026014] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Created folder: Project (c95aa87980de4ec8a2170c2640423301) in parent group-v438503. [ 2369.026224] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Creating folder: Instances. Parent ref: group-v438767. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 2369.026462] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0c775450-0f28-4727-b6d7-f09493afc5dd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2369.036180] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Created folder: Instances in parent group-v438767. [ 2369.036473] env[62824]: DEBUG oslo.service.loopingcall [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2369.036679] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c70880c9-a69a-49b9-b028-189c04276623] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2369.036959] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-492f9dd8-6547-48f6-9973-48baad156069 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2369.057484] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2369.057484] env[62824]: value = "task-2146918" [ 2369.057484] env[62824]: _type = "Task" [ 2369.057484] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2369.065738] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146918, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2369.077548] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146913, 'name': CreateVM_Task, 'duration_secs': 0.429725} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2369.078135] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2369.078875] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2369.079058] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2369.079715] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2369.079715] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d10d13d5-57f0-4bb3-aed3-c957e468b9f3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2369.085440] env[62824]: DEBUG oslo_vmware.api [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2369.085440] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52a40988-13f5-0c7b-e255-fed4f0003c67" [ 2369.085440] env[62824]: _type = "Task" [ 2369.085440] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2369.094878] env[62824]: DEBUG oslo_vmware.api [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52a40988-13f5-0c7b-e255-fed4f0003c67, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2369.181300] env[62824]: DEBUG oslo_vmware.api [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2146915, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.100257} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2369.181620] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2369.182408] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06d25919-c918-4252-9bd8-98d1efc90316 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2369.196568] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b70c134a-661f-4db0-98ce-919ec5cda15d tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.848s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2369.198676] env[62824]: DEBUG oslo_concurrency.lockutils [None req-97300fec-1d35-4a49-954c-b0818a0040aa tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquiring lock "interface-511ef5bb-1dca-4eae-a0f1-19d1d7a17371-afcd2bdb-2bfc-40d7-a8e5-2a417be0e0e9" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2369.198897] env[62824]: DEBUG oslo_concurrency.lockutils [None req-97300fec-1d35-4a49-954c-b0818a0040aa tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lock "interface-511ef5bb-1dca-4eae-a0f1-19d1d7a17371-afcd2bdb-2bfc-40d7-a8e5-2a417be0e0e9" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2369.199286] env[62824]: DEBUG nova.objects.instance [None req-97300fec-1d35-4a49-954c-b0818a0040aa tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lazy-loading 'flavor' on Instance uuid 511ef5bb-1dca-4eae-a0f1-19d1d7a17371 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2369.209121] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Reconfiguring VM instance instance-00000068 to attach disk [datastore2] 1db9ff97-e59c-43f6-984e-6405d8e993a4/1db9ff97-e59c-43f6-984e-6405d8e993a4.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2369.212468] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8b38a8bf-0158-41db-a4ca-f5d00d63bf9d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2369.235706] env[62824]: DEBUG oslo_vmware.api [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2369.235706] env[62824]: value = "task-2146919" [ 2369.235706] env[62824]: _type = "Task" [ 2369.235706] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2369.236639] env[62824]: INFO nova.scheduler.client.report [None req-b70c134a-661f-4db0-98ce-919ec5cda15d tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Deleted allocations for instance fb9cfb53-285a-4586-acf0-cb08d56b93e7 [ 2369.249360] env[62824]: DEBUG oslo_vmware.api [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2146919, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2369.260159] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146912, 'name': CreateVM_Task, 'duration_secs': 0.54646} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2369.260366] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2369.260994] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2369.388526] env[62824]: DEBUG nova.network.neutron [req-e1ac7983-616b-4c48-8b20-9cb7de8fb175 req-e5129aa6-20ea-4a2b-898a-fb517da29a21 service nova] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Updated VIF entry in instance network info cache for port edaf5d3c-004d-4ba0-b037-e7eef2acbe44. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2369.389100] env[62824]: DEBUG nova.network.neutron [req-e1ac7983-616b-4c48-8b20-9cb7de8fb175 req-e5129aa6-20ea-4a2b-898a-fb517da29a21 service nova] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Updating instance_info_cache with network_info: [{"id": "edaf5d3c-004d-4ba0-b037-e7eef2acbe44", "address": "fa:16:3e:06:2e:6a", "network": {"id": "12411632-b7f6-4544-b2e4-b04ae7337547", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-917106830-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4479921caf5f405b8fc49baad390a0e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "de7fa486-5f28-44ae-b0cf-72234ff87546", "external-id": "nsx-vlan-transportzone-229", "segmentation_id": 229, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapedaf5d3c-00", "ovs_interfaceid": "edaf5d3c-004d-4ba0-b037-e7eef2acbe44", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2369.569777] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146918, 'name': CreateVM_Task, 'duration_secs': 0.435266} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2369.569950] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c70880c9-a69a-49b9-b028-189c04276623] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2369.570614] env[62824]: DEBUG oslo_concurrency.lockutils [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2369.596755] env[62824]: DEBUG oslo_vmware.api [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52a40988-13f5-0c7b-e255-fed4f0003c67, 'name': SearchDatastore_Task, 'duration_secs': 0.012364} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2369.596890] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2369.597122] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2369.597352] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2369.597577] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2369.597658] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2369.598241] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2369.598321] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2369.598784] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-56ae3765-82bd-4cde-b34b-7b4d91e69661 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2369.600334] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c40e11c6-d851-47f4-8d47-62229d3d2df8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2369.605932] env[62824]: DEBUG oslo_vmware.api [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2369.605932] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52b2caed-203e-a613-0ae1-03e5d87c8588" [ 2369.605932] env[62824]: _type = "Task" [ 2369.605932] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2369.612120] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2369.612313] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2369.615631] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1e66a483-e024-4455-9851-0a5a3d79afe5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2369.617781] env[62824]: DEBUG oslo_vmware.api [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52b2caed-203e-a613-0ae1-03e5d87c8588, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2369.620880] env[62824]: DEBUG oslo_vmware.api [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2369.620880] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52958e0b-7079-b260-5e36-6689e830540d" [ 2369.620880] env[62824]: _type = "Task" [ 2369.620880] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2369.628547] env[62824]: DEBUG oslo_vmware.api [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52958e0b-7079-b260-5e36-6689e830540d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2369.748284] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b70c134a-661f-4db0-98ce-919ec5cda15d tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "fb9cfb53-285a-4586-acf0-cb08d56b93e7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.356s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2369.752638] env[62824]: DEBUG oslo_vmware.api [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2146919, 'name': ReconfigVM_Task, 'duration_secs': 0.288632} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2369.753137] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Reconfigured VM instance instance-00000068 to attach disk [datastore2] 1db9ff97-e59c-43f6-984e-6405d8e993a4/1db9ff97-e59c-43f6-984e-6405d8e993a4.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2369.753730] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-70ebbdc6-849c-4d88-b36b-706308297027 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2369.761870] env[62824]: DEBUG oslo_vmware.api [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2369.761870] env[62824]: value = "task-2146920" [ 2369.761870] env[62824]: _type = "Task" [ 2369.761870] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2369.772764] env[62824]: DEBUG oslo_vmware.api [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2146920, 'name': Rename_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2369.817531] env[62824]: DEBUG nova.objects.instance [None req-97300fec-1d35-4a49-954c-b0818a0040aa tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lazy-loading 'pci_requests' on Instance uuid 511ef5bb-1dca-4eae-a0f1-19d1d7a17371 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2369.892282] env[62824]: DEBUG oslo_concurrency.lockutils [req-e1ac7983-616b-4c48-8b20-9cb7de8fb175 req-e5129aa6-20ea-4a2b-898a-fb517da29a21 service nova] Releasing lock "refresh_cache-21a7eed6-c146-4911-9a0a-a1e1644e8e9d" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2369.892604] env[62824]: DEBUG nova.compute.manager [req-e1ac7983-616b-4c48-8b20-9cb7de8fb175 req-e5129aa6-20ea-4a2b-898a-fb517da29a21 service nova] [instance: c70880c9-a69a-49b9-b028-189c04276623] Received event network-vif-plugged-f71d9cbc-c160-47f9-9771-af38641d3683 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2369.892843] env[62824]: DEBUG oslo_concurrency.lockutils [req-e1ac7983-616b-4c48-8b20-9cb7de8fb175 req-e5129aa6-20ea-4a2b-898a-fb517da29a21 service nova] Acquiring lock "c70880c9-a69a-49b9-b028-189c04276623-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2369.893094] env[62824]: DEBUG oslo_concurrency.lockutils [req-e1ac7983-616b-4c48-8b20-9cb7de8fb175 req-e5129aa6-20ea-4a2b-898a-fb517da29a21 service nova] Lock "c70880c9-a69a-49b9-b028-189c04276623-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2369.893332] env[62824]: DEBUG oslo_concurrency.lockutils [req-e1ac7983-616b-4c48-8b20-9cb7de8fb175 req-e5129aa6-20ea-4a2b-898a-fb517da29a21 service nova] Lock "c70880c9-a69a-49b9-b028-189c04276623-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2369.893566] env[62824]: DEBUG nova.compute.manager [req-e1ac7983-616b-4c48-8b20-9cb7de8fb175 req-e5129aa6-20ea-4a2b-898a-fb517da29a21 service nova] [instance: c70880c9-a69a-49b9-b028-189c04276623] No waiting events found dispatching network-vif-plugged-f71d9cbc-c160-47f9-9771-af38641d3683 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2369.893783] env[62824]: WARNING nova.compute.manager [req-e1ac7983-616b-4c48-8b20-9cb7de8fb175 req-e5129aa6-20ea-4a2b-898a-fb517da29a21 service nova] [instance: c70880c9-a69a-49b9-b028-189c04276623] Received unexpected event network-vif-plugged-f71d9cbc-c160-47f9-9771-af38641d3683 for instance with vm_state building and task_state spawning. [ 2369.894015] env[62824]: DEBUG nova.compute.manager [req-e1ac7983-616b-4c48-8b20-9cb7de8fb175 req-e5129aa6-20ea-4a2b-898a-fb517da29a21 service nova] [instance: c70880c9-a69a-49b9-b028-189c04276623] Received event network-changed-f71d9cbc-c160-47f9-9771-af38641d3683 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2369.894241] env[62824]: DEBUG nova.compute.manager [req-e1ac7983-616b-4c48-8b20-9cb7de8fb175 req-e5129aa6-20ea-4a2b-898a-fb517da29a21 service nova] [instance: c70880c9-a69a-49b9-b028-189c04276623] Refreshing instance network info cache due to event network-changed-f71d9cbc-c160-47f9-9771-af38641d3683. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2369.894476] env[62824]: DEBUG oslo_concurrency.lockutils [req-e1ac7983-616b-4c48-8b20-9cb7de8fb175 req-e5129aa6-20ea-4a2b-898a-fb517da29a21 service nova] Acquiring lock "refresh_cache-c70880c9-a69a-49b9-b028-189c04276623" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2369.894653] env[62824]: DEBUG oslo_concurrency.lockutils [req-e1ac7983-616b-4c48-8b20-9cb7de8fb175 req-e5129aa6-20ea-4a2b-898a-fb517da29a21 service nova] Acquired lock "refresh_cache-c70880c9-a69a-49b9-b028-189c04276623" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2369.894859] env[62824]: DEBUG nova.network.neutron [req-e1ac7983-616b-4c48-8b20-9cb7de8fb175 req-e5129aa6-20ea-4a2b-898a-fb517da29a21 service nova] [instance: c70880c9-a69a-49b9-b028-189c04276623] Refreshing network info cache for port f71d9cbc-c160-47f9-9771-af38641d3683 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2370.031572] env[62824]: INFO nova.compute.manager [None req-09c485ce-ef98-43f8-a6af-31be637f0d94 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Resuming [ 2370.032199] env[62824]: DEBUG nova.objects.instance [None req-09c485ce-ef98-43f8-a6af-31be637f0d94 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Lazy-loading 'flavor' on Instance uuid 71e9733b-c84b-4501-8faf-a487f089b498 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2370.117137] env[62824]: DEBUG oslo_vmware.api [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52b2caed-203e-a613-0ae1-03e5d87c8588, 'name': SearchDatastore_Task, 'duration_secs': 0.016798} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2370.117414] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2370.117657] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2370.117866] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2370.118104] env[62824]: DEBUG oslo_concurrency.lockutils [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2370.118406] env[62824]: DEBUG oslo_concurrency.lockutils [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2370.118659] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-72e5363c-b04a-4e4b-b042-1d16468207ef {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2370.125944] env[62824]: DEBUG oslo_vmware.api [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Waiting for the task: (returnval){ [ 2370.125944] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52a9eab9-d6d8-e825-c64f-80c3ee5c3b1e" [ 2370.125944] env[62824]: _type = "Task" [ 2370.125944] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2370.132579] env[62824]: DEBUG oslo_vmware.api [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52958e0b-7079-b260-5e36-6689e830540d, 'name': SearchDatastore_Task, 'duration_secs': 0.008979} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2370.133637] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-611e3769-1c03-452f-b02a-169c495d5431 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2370.138588] env[62824]: DEBUG oslo_vmware.api [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52a9eab9-d6d8-e825-c64f-80c3ee5c3b1e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2370.143970] env[62824]: DEBUG oslo_vmware.api [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2370.143970] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52e2cd86-22a4-5bd8-dc45-8da60fff83b8" [ 2370.143970] env[62824]: _type = "Task" [ 2370.143970] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2370.153276] env[62824]: DEBUG oslo_vmware.api [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52e2cd86-22a4-5bd8-dc45-8da60fff83b8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2370.271625] env[62824]: DEBUG oslo_vmware.api [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2146920, 'name': Rename_Task, 'duration_secs': 0.147946} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2370.271952] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2370.272284] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-57cd30c1-014b-41ae-8f58-2aaf3739a1b6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2370.279257] env[62824]: DEBUG oslo_vmware.api [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2370.279257] env[62824]: value = "task-2146921" [ 2370.279257] env[62824]: _type = "Task" [ 2370.279257] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2370.287797] env[62824]: DEBUG oslo_vmware.api [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2146921, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2370.320933] env[62824]: DEBUG nova.objects.base [None req-97300fec-1d35-4a49-954c-b0818a0040aa tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Object Instance<511ef5bb-1dca-4eae-a0f1-19d1d7a17371> lazy-loaded attributes: flavor,pci_requests {{(pid=62824) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 2370.321351] env[62824]: DEBUG nova.network.neutron [None req-97300fec-1d35-4a49-954c-b0818a0040aa tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2370.386266] env[62824]: DEBUG nova.policy [None req-97300fec-1d35-4a49-954c-b0818a0040aa tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b989eb89498b45dab6604116870d2fa1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c987c5f844be42c99cc9d57a9ca84c31', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2370.638878] env[62824]: DEBUG oslo_vmware.api [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52a9eab9-d6d8-e825-c64f-80c3ee5c3b1e, 'name': SearchDatastore_Task, 'duration_secs': 0.015115} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2370.639225] env[62824]: DEBUG oslo_concurrency.lockutils [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2370.639471] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c70880c9-a69a-49b9-b028-189c04276623] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2370.639707] env[62824]: DEBUG oslo_concurrency.lockutils [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2370.645259] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5d6a43b2-91e7-4cee-b3f7-f24637b92565 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquiring lock "a6cd2032-de60-4f78-bf1e-79801d049df0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2370.645529] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5d6a43b2-91e7-4cee-b3f7-f24637b92565 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "a6cd2032-de60-4f78-bf1e-79801d049df0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2370.645729] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5d6a43b2-91e7-4cee-b3f7-f24637b92565 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquiring lock "a6cd2032-de60-4f78-bf1e-79801d049df0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2370.645934] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5d6a43b2-91e7-4cee-b3f7-f24637b92565 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "a6cd2032-de60-4f78-bf1e-79801d049df0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2370.646157] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5d6a43b2-91e7-4cee-b3f7-f24637b92565 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "a6cd2032-de60-4f78-bf1e-79801d049df0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2370.652588] env[62824]: INFO nova.compute.manager [None req-5d6a43b2-91e7-4cee-b3f7-f24637b92565 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: a6cd2032-de60-4f78-bf1e-79801d049df0] Terminating instance [ 2370.661951] env[62824]: DEBUG oslo_vmware.api [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52e2cd86-22a4-5bd8-dc45-8da60fff83b8, 'name': SearchDatastore_Task, 'duration_secs': 0.016989} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2370.662793] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2370.663066] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 21a7eed6-c146-4911-9a0a-a1e1644e8e9d/21a7eed6-c146-4911-9a0a-a1e1644e8e9d.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2370.663578] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2370.663836] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2370.664294] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b4a8fda0-78e2-46a2-af1e-cf2826a74c48 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2370.670419] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9a4263a9-512b-408f-b8bb-aa065a9863ca {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2370.679214] env[62824]: DEBUG oslo_vmware.api [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2370.679214] env[62824]: value = "task-2146922" [ 2370.679214] env[62824]: _type = "Task" [ 2370.679214] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2370.685816] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2370.686033] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2370.687354] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b4523269-673c-4372-84d6-232c930a8643 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2370.693377] env[62824]: DEBUG oslo_vmware.api [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146922, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2370.696890] env[62824]: DEBUG oslo_vmware.api [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2370.696890] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]528b149f-9cc7-5b7a-5029-259da319bd48" [ 2370.696890] env[62824]: _type = "Task" [ 2370.696890] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2370.706629] env[62824]: DEBUG oslo_vmware.api [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]528b149f-9cc7-5b7a-5029-259da319bd48, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2370.738362] env[62824]: DEBUG nova.network.neutron [req-e1ac7983-616b-4c48-8b20-9cb7de8fb175 req-e5129aa6-20ea-4a2b-898a-fb517da29a21 service nova] [instance: c70880c9-a69a-49b9-b028-189c04276623] Updated VIF entry in instance network info cache for port f71d9cbc-c160-47f9-9771-af38641d3683. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2370.738984] env[62824]: DEBUG nova.network.neutron [req-e1ac7983-616b-4c48-8b20-9cb7de8fb175 req-e5129aa6-20ea-4a2b-898a-fb517da29a21 service nova] [instance: c70880c9-a69a-49b9-b028-189c04276623] Updating instance_info_cache with network_info: [{"id": "f71d9cbc-c160-47f9-9771-af38641d3683", "address": "fa:16:3e:51:89:60", "network": {"id": "f7936d74-eeef-4c41-8ae7-9427c43663fa", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1643883257-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c95aa87980de4ec8a2170c2640423301", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6eaa481-1f92-4851-b98e-09ed0daad7cb", "external-id": "nsx-vlan-transportzone-636", "segmentation_id": 636, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf71d9cbc-c1", "ovs_interfaceid": "f71d9cbc-c160-47f9-9771-af38641d3683", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2370.790673] env[62824]: DEBUG oslo_vmware.api [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2146921, 'name': PowerOnVM_Task, 'duration_secs': 0.478991} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2370.791028] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2370.791281] env[62824]: DEBUG nova.compute.manager [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2370.792170] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b3eb00d-7698-4db0-8818-3c4ea5b52772 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2371.157847] env[62824]: DEBUG nova.compute.manager [None req-5d6a43b2-91e7-4cee-b3f7-f24637b92565 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: a6cd2032-de60-4f78-bf1e-79801d049df0] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2371.158113] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5d6a43b2-91e7-4cee-b3f7-f24637b92565 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: a6cd2032-de60-4f78-bf1e-79801d049df0] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2371.159209] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ccdefac-e51c-4cbe-b90d-2e117525474a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2371.167876] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d6a43b2-91e7-4cee-b3f7-f24637b92565 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: a6cd2032-de60-4f78-bf1e-79801d049df0] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2371.168181] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-82aaeada-f0de-43a0-9cfa-5a10f12b39f5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2371.175872] env[62824]: DEBUG oslo_vmware.api [None req-5d6a43b2-91e7-4cee-b3f7-f24637b92565 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2371.175872] env[62824]: value = "task-2146923" [ 2371.175872] env[62824]: _type = "Task" [ 2371.175872] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2371.190649] env[62824]: DEBUG oslo_vmware.api [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146922, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.506855} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2371.193408] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 21a7eed6-c146-4911-9a0a-a1e1644e8e9d/21a7eed6-c146-4911-9a0a-a1e1644e8e9d.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2371.193639] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2371.193901] env[62824]: DEBUG oslo_vmware.api [None req-5d6a43b2-91e7-4cee-b3f7-f24637b92565 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146923, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2371.194156] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0b04afb2-c39b-462d-8fee-17b4a91ee639 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2371.202647] env[62824]: DEBUG oslo_vmware.api [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2371.202647] env[62824]: value = "task-2146924" [ 2371.202647] env[62824]: _type = "Task" [ 2371.202647] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2371.211959] env[62824]: DEBUG oslo_vmware.api [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]528b149f-9cc7-5b7a-5029-259da319bd48, 'name': SearchDatastore_Task, 'duration_secs': 0.012088} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2371.213465] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8a064dd0-7b64-4be5-aaa2-a53ea0049b60 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2371.219205] env[62824]: DEBUG oslo_vmware.api [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146924, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2371.222349] env[62824]: DEBUG oslo_vmware.api [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2371.222349] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]526373b4-5459-f828-ed30-5ea9864822dc" [ 2371.222349] env[62824]: _type = "Task" [ 2371.222349] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2371.229935] env[62824]: DEBUG oslo_vmware.api [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]526373b4-5459-f828-ed30-5ea9864822dc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2371.243162] env[62824]: DEBUG oslo_concurrency.lockutils [req-e1ac7983-616b-4c48-8b20-9cb7de8fb175 req-e5129aa6-20ea-4a2b-898a-fb517da29a21 service nova] Releasing lock "refresh_cache-c70880c9-a69a-49b9-b028-189c04276623" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2371.243162] env[62824]: DEBUG nova.compute.manager [req-e1ac7983-616b-4c48-8b20-9cb7de8fb175 req-e5129aa6-20ea-4a2b-898a-fb517da29a21 service nova] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Received event network-changed-d5c6ba87-e0c7-49c2-bbaf-b7c64f84cdf8 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2371.243162] env[62824]: DEBUG nova.compute.manager [req-e1ac7983-616b-4c48-8b20-9cb7de8fb175 req-e5129aa6-20ea-4a2b-898a-fb517da29a21 service nova] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Refreshing instance network info cache due to event network-changed-d5c6ba87-e0c7-49c2-bbaf-b7c64f84cdf8. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2371.243162] env[62824]: DEBUG oslo_concurrency.lockutils [req-e1ac7983-616b-4c48-8b20-9cb7de8fb175 req-e5129aa6-20ea-4a2b-898a-fb517da29a21 service nova] Acquiring lock "refresh_cache-e1574266-e53b-4169-9db6-c66b84895edb" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2371.243162] env[62824]: DEBUG oslo_concurrency.lockutils [req-e1ac7983-616b-4c48-8b20-9cb7de8fb175 req-e5129aa6-20ea-4a2b-898a-fb517da29a21 service nova] Acquired lock "refresh_cache-e1574266-e53b-4169-9db6-c66b84895edb" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2371.243162] env[62824]: DEBUG nova.network.neutron [req-e1ac7983-616b-4c48-8b20-9cb7de8fb175 req-e5129aa6-20ea-4a2b-898a-fb517da29a21 service nova] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Refreshing network info cache for port d5c6ba87-e0c7-49c2-bbaf-b7c64f84cdf8 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2371.312017] env[62824]: DEBUG oslo_concurrency.lockutils [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2371.312294] env[62824]: DEBUG oslo_concurrency.lockutils [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2371.312473] env[62824]: DEBUG nova.objects.instance [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62824) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 2371.544309] env[62824]: DEBUG oslo_concurrency.lockutils [None req-09c485ce-ef98-43f8-a6af-31be637f0d94 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Acquiring lock "refresh_cache-71e9733b-c84b-4501-8faf-a487f089b498" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2371.545064] env[62824]: DEBUG oslo_concurrency.lockutils [None req-09c485ce-ef98-43f8-a6af-31be637f0d94 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Acquired lock "refresh_cache-71e9733b-c84b-4501-8faf-a487f089b498" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2371.545363] env[62824]: DEBUG nova.network.neutron [None req-09c485ce-ef98-43f8-a6af-31be637f0d94 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2371.686767] env[62824]: DEBUG oslo_vmware.api [None req-5d6a43b2-91e7-4cee-b3f7-f24637b92565 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146923, 'name': PowerOffVM_Task, 'duration_secs': 0.20796} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2371.687081] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d6a43b2-91e7-4cee-b3f7-f24637b92565 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: a6cd2032-de60-4f78-bf1e-79801d049df0] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2371.687218] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5d6a43b2-91e7-4cee-b3f7-f24637b92565 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: a6cd2032-de60-4f78-bf1e-79801d049df0] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2371.687467] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a9d67c24-2232-4adb-bf9d-e419093a7527 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2371.713949] env[62824]: DEBUG oslo_vmware.api [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146924, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06617} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2371.714218] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2371.714992] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95c3502b-5384-4a9f-a05e-12670d413fbf {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2371.737848] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Reconfiguring VM instance instance-0000006d to attach disk [datastore2] 21a7eed6-c146-4911-9a0a-a1e1644e8e9d/21a7eed6-c146-4911-9a0a-a1e1644e8e9d.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2371.741240] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3c97f953-afe9-45e4-bba9-764357999c0c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2371.763429] env[62824]: DEBUG oslo_vmware.api [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]526373b4-5459-f828-ed30-5ea9864822dc, 'name': SearchDatastore_Task, 'duration_secs': 0.009217} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2371.764595] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2371.764781] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 3d294ee9-92f6-4115-95db-9d4b66562b56/3d294ee9-92f6-4115-95db-9d4b66562b56.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2371.765242] env[62824]: DEBUG oslo_vmware.api [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2371.765242] env[62824]: value = "task-2146926" [ 2371.765242] env[62824]: _type = "Task" [ 2371.765242] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2371.765471] env[62824]: DEBUG oslo_concurrency.lockutils [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2371.765545] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2371.765762] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d8ba28ec-e976-4593-bc95-90172889287b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2371.767715] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dc471e65-40aa-4363-8cf8-abd440fdff7e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2371.778382] env[62824]: DEBUG oslo_vmware.api [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146926, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2371.779348] env[62824]: DEBUG oslo_vmware.api [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2371.779348] env[62824]: value = "task-2146927" [ 2371.779348] env[62824]: _type = "Task" [ 2371.779348] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2371.780395] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2371.780566] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2371.784250] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e77333e3-da01-4f3f-805b-e20b31c1dfa6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2371.792729] env[62824]: DEBUG oslo_vmware.api [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Waiting for the task: (returnval){ [ 2371.792729] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5206a798-85f6-1ba7-ed9e-7d6191cfef90" [ 2371.792729] env[62824]: _type = "Task" [ 2371.792729] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2371.795985] env[62824]: DEBUG oslo_vmware.api [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146927, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2371.804446] env[62824]: DEBUG oslo_vmware.api [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5206a798-85f6-1ba7-ed9e-7d6191cfef90, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2371.943725] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5d6a43b2-91e7-4cee-b3f7-f24637b92565 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: a6cd2032-de60-4f78-bf1e-79801d049df0] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2371.944021] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5d6a43b2-91e7-4cee-b3f7-f24637b92565 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: a6cd2032-de60-4f78-bf1e-79801d049df0] Deleting contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2371.944318] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-5d6a43b2-91e7-4cee-b3f7-f24637b92565 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Deleting the datastore file [datastore1] a6cd2032-de60-4f78-bf1e-79801d049df0 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2371.944696] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2f066cd6-a82f-42d7-aaae-a135e590f8f3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2371.955537] env[62824]: DEBUG oslo_vmware.api [None req-5d6a43b2-91e7-4cee-b3f7-f24637b92565 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for the task: (returnval){ [ 2371.955537] env[62824]: value = "task-2146928" [ 2371.955537] env[62824]: _type = "Task" [ 2371.955537] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2371.966013] env[62824]: DEBUG oslo_vmware.api [None req-5d6a43b2-91e7-4cee-b3f7-f24637b92565 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146928, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2372.105653] env[62824]: DEBUG nova.network.neutron [req-e1ac7983-616b-4c48-8b20-9cb7de8fb175 req-e5129aa6-20ea-4a2b-898a-fb517da29a21 service nova] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Updated VIF entry in instance network info cache for port d5c6ba87-e0c7-49c2-bbaf-b7c64f84cdf8. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2372.106038] env[62824]: DEBUG nova.network.neutron [req-e1ac7983-616b-4c48-8b20-9cb7de8fb175 req-e5129aa6-20ea-4a2b-898a-fb517da29a21 service nova] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Updating instance_info_cache with network_info: [{"id": "d5c6ba87-e0c7-49c2-bbaf-b7c64f84cdf8", "address": "fa:16:3e:c4:9c:f4", "network": {"id": "aa1c09e5-8d72-43ad-a903-a0f5e711da80", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1643616572-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c987c5f844be42c99cc9d57a9ca84c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9bc2632-36f9-4912-8782-8bbb789f909d", "external-id": "nsx-vlan-transportzone-897", "segmentation_id": 897, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd5c6ba87-e0", "ovs_interfaceid": "d5c6ba87-e0c7-49c2-bbaf-b7c64f84cdf8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2372.119927] env[62824]: DEBUG nova.compute.manager [req-9092c2f3-fe68-4868-a43c-ec421af56c87 req-73644f52-81ab-43b0-9715-797bd9523f2c service nova] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Received event network-vif-plugged-afcd2bdb-2bfc-40d7-a8e5-2a417be0e0e9 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2372.120173] env[62824]: DEBUG oslo_concurrency.lockutils [req-9092c2f3-fe68-4868-a43c-ec421af56c87 req-73644f52-81ab-43b0-9715-797bd9523f2c service nova] Acquiring lock "511ef5bb-1dca-4eae-a0f1-19d1d7a17371-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2372.120385] env[62824]: DEBUG oslo_concurrency.lockutils [req-9092c2f3-fe68-4868-a43c-ec421af56c87 req-73644f52-81ab-43b0-9715-797bd9523f2c service nova] Lock "511ef5bb-1dca-4eae-a0f1-19d1d7a17371-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2372.120553] env[62824]: DEBUG oslo_concurrency.lockutils [req-9092c2f3-fe68-4868-a43c-ec421af56c87 req-73644f52-81ab-43b0-9715-797bd9523f2c service nova] Lock "511ef5bb-1dca-4eae-a0f1-19d1d7a17371-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2372.120721] env[62824]: DEBUG nova.compute.manager [req-9092c2f3-fe68-4868-a43c-ec421af56c87 req-73644f52-81ab-43b0-9715-797bd9523f2c service nova] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] No waiting events found dispatching network-vif-plugged-afcd2bdb-2bfc-40d7-a8e5-2a417be0e0e9 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2372.120884] env[62824]: WARNING nova.compute.manager [req-9092c2f3-fe68-4868-a43c-ec421af56c87 req-73644f52-81ab-43b0-9715-797bd9523f2c service nova] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Received unexpected event network-vif-plugged-afcd2bdb-2bfc-40d7-a8e5-2a417be0e0e9 for instance with vm_state active and task_state None. [ 2372.284082] env[62824]: DEBUG oslo_vmware.api [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146926, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2372.301529] env[62824]: DEBUG oslo_vmware.api [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146927, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2372.307708] env[62824]: DEBUG nova.network.neutron [None req-97300fec-1d35-4a49-954c-b0818a0040aa tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Successfully updated port: afcd2bdb-2bfc-40d7-a8e5-2a417be0e0e9 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2372.318439] env[62824]: DEBUG oslo_vmware.api [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5206a798-85f6-1ba7-ed9e-7d6191cfef90, 'name': SearchDatastore_Task, 'duration_secs': 0.011822} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2372.320353] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f32f0e7d-4385-4414-9554-b8cba9a7e20b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2372.325735] env[62824]: DEBUG oslo_concurrency.lockutils [None req-cf6bb79a-a913-4007-b0c5-073d6ba3eaab tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.013s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2372.334686] env[62824]: DEBUG oslo_vmware.api [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Waiting for the task: (returnval){ [ 2372.334686] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52b2c3d6-a633-9997-20c8-0b000a9e1988" [ 2372.334686] env[62824]: _type = "Task" [ 2372.334686] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2372.345372] env[62824]: DEBUG oslo_vmware.api [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52b2c3d6-a633-9997-20c8-0b000a9e1988, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2372.387107] env[62824]: DEBUG nova.network.neutron [None req-09c485ce-ef98-43f8-a6af-31be637f0d94 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Updating instance_info_cache with network_info: [{"id": "4214bf6e-29f0-4d92-be44-44fccf7ab149", "address": "fa:16:3e:a6:d8:c4", "network": {"id": "fc5d7128-fe9b-44ef-9d91-839190a97226", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-796405330-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35b7289f4746463e9b1cc47fe914a835", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0954fad3-d24d-496c-83e6-a09d3cb556fc", "external-id": "nsx-vlan-transportzone-216", "segmentation_id": 216, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4214bf6e-29", "ovs_interfaceid": "4214bf6e-29f0-4d92-be44-44fccf7ab149", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2372.465983] env[62824]: DEBUG oslo_vmware.api [None req-5d6a43b2-91e7-4cee-b3f7-f24637b92565 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Task: {'id': task-2146928, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.377456} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2372.466274] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-5d6a43b2-91e7-4cee-b3f7-f24637b92565 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2372.466466] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5d6a43b2-91e7-4cee-b3f7-f24637b92565 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: a6cd2032-de60-4f78-bf1e-79801d049df0] Deleted contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2372.466658] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5d6a43b2-91e7-4cee-b3f7-f24637b92565 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: a6cd2032-de60-4f78-bf1e-79801d049df0] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2372.466832] env[62824]: INFO nova.compute.manager [None req-5d6a43b2-91e7-4cee-b3f7-f24637b92565 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] [instance: a6cd2032-de60-4f78-bf1e-79801d049df0] Took 1.31 seconds to destroy the instance on the hypervisor. [ 2372.467090] env[62824]: DEBUG oslo.service.loopingcall [None req-5d6a43b2-91e7-4cee-b3f7-f24637b92565 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2372.467538] env[62824]: DEBUG nova.compute.manager [-] [instance: a6cd2032-de60-4f78-bf1e-79801d049df0] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2372.467538] env[62824]: DEBUG nova.network.neutron [-] [instance: a6cd2032-de60-4f78-bf1e-79801d049df0] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2372.609569] env[62824]: DEBUG oslo_concurrency.lockutils [req-e1ac7983-616b-4c48-8b20-9cb7de8fb175 req-e5129aa6-20ea-4a2b-898a-fb517da29a21 service nova] Releasing lock "refresh_cache-e1574266-e53b-4169-9db6-c66b84895edb" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2372.609858] env[62824]: DEBUG nova.compute.manager [req-e1ac7983-616b-4c48-8b20-9cb7de8fb175 req-e5129aa6-20ea-4a2b-898a-fb517da29a21 service nova] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Received event network-changed-3c1f3341-20a4-4b77-8dae-12ef5901a16e {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2372.610044] env[62824]: DEBUG nova.compute.manager [req-e1ac7983-616b-4c48-8b20-9cb7de8fb175 req-e5129aa6-20ea-4a2b-898a-fb517da29a21 service nova] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Refreshing instance network info cache due to event network-changed-3c1f3341-20a4-4b77-8dae-12ef5901a16e. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2372.610260] env[62824]: DEBUG oslo_concurrency.lockutils [req-e1ac7983-616b-4c48-8b20-9cb7de8fb175 req-e5129aa6-20ea-4a2b-898a-fb517da29a21 service nova] Acquiring lock "refresh_cache-511ef5bb-1dca-4eae-a0f1-19d1d7a17371" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2372.610405] env[62824]: DEBUG oslo_concurrency.lockutils [req-e1ac7983-616b-4c48-8b20-9cb7de8fb175 req-e5129aa6-20ea-4a2b-898a-fb517da29a21 service nova] Acquired lock "refresh_cache-511ef5bb-1dca-4eae-a0f1-19d1d7a17371" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2372.610564] env[62824]: DEBUG nova.network.neutron [req-e1ac7983-616b-4c48-8b20-9cb7de8fb175 req-e5129aa6-20ea-4a2b-898a-fb517da29a21 service nova] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Refreshing network info cache for port 3c1f3341-20a4-4b77-8dae-12ef5901a16e {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2372.779702] env[62824]: DEBUG oslo_vmware.api [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146926, 'name': ReconfigVM_Task, 'duration_secs': 0.619902} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2372.779702] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Reconfigured VM instance instance-0000006d to attach disk [datastore2] 21a7eed6-c146-4911-9a0a-a1e1644e8e9d/21a7eed6-c146-4911-9a0a-a1e1644e8e9d.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2372.781440] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5f8a74f6-5889-4e9d-9e05-8a6c1513ade5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2372.789777] env[62824]: DEBUG oslo_vmware.api [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2372.789777] env[62824]: value = "task-2146929" [ 2372.789777] env[62824]: _type = "Task" [ 2372.789777] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2372.794729] env[62824]: DEBUG oslo_vmware.api [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146927, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.557253} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2372.801378] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 3d294ee9-92f6-4115-95db-9d4b66562b56/3d294ee9-92f6-4115-95db-9d4b66562b56.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2372.801610] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2372.802241] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d854c12f-eab3-4246-a11c-dd30b5ae66cc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2372.811110] env[62824]: DEBUG oslo_concurrency.lockutils [None req-97300fec-1d35-4a49-954c-b0818a0040aa tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquiring lock "refresh_cache-511ef5bb-1dca-4eae-a0f1-19d1d7a17371" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2372.811110] env[62824]: DEBUG oslo_vmware.api [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146929, 'name': Rename_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2372.812454] env[62824]: DEBUG oslo_vmware.api [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2372.812454] env[62824]: value = "task-2146930" [ 2372.812454] env[62824]: _type = "Task" [ 2372.812454] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2372.822912] env[62824]: DEBUG oslo_vmware.api [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146930, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2372.845928] env[62824]: DEBUG oslo_vmware.api [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52b2c3d6-a633-9997-20c8-0b000a9e1988, 'name': SearchDatastore_Task, 'duration_secs': 0.023715} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2372.846221] env[62824]: DEBUG oslo_concurrency.lockutils [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2372.846457] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] c70880c9-a69a-49b9-b028-189c04276623/c70880c9-a69a-49b9-b028-189c04276623.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2372.846712] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-38b9752f-4f28-48ba-bd6f-f6e39371866d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2372.854578] env[62824]: DEBUG oslo_vmware.api [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Waiting for the task: (returnval){ [ 2372.854578] env[62824]: value = "task-2146931" [ 2372.854578] env[62824]: _type = "Task" [ 2372.854578] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2372.863158] env[62824]: DEBUG oslo_vmware.api [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2146931, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2372.889506] env[62824]: DEBUG oslo_concurrency.lockutils [None req-09c485ce-ef98-43f8-a6af-31be637f0d94 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Releasing lock "refresh_cache-71e9733b-c84b-4501-8faf-a487f089b498" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2372.890503] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71ccb0d4-267e-45cc-9978-ea69e94300b5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2372.897501] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-09c485ce-ef98-43f8-a6af-31be637f0d94 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Resuming the VM {{(pid=62824) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1183}} [ 2372.897743] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-07c33693-f768-4574-acf0-5d50b3c348e9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2372.904691] env[62824]: DEBUG oslo_vmware.api [None req-09c485ce-ef98-43f8-a6af-31be637f0d94 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Waiting for the task: (returnval){ [ 2372.904691] env[62824]: value = "task-2146932" [ 2372.904691] env[62824]: _type = "Task" [ 2372.904691] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2372.917789] env[62824]: DEBUG oslo_vmware.api [None req-09c485ce-ef98-43f8-a6af-31be637f0d94 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146932, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2373.285304] env[62824]: DEBUG nova.network.neutron [-] [instance: a6cd2032-de60-4f78-bf1e-79801d049df0] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2373.310350] env[62824]: DEBUG oslo_vmware.api [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146929, 'name': Rename_Task, 'duration_secs': 0.159067} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2373.311339] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2373.311652] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-06d55ab5-59cf-45d9-ba9e-d5754c1f6cca {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2373.325240] env[62824]: DEBUG oslo_vmware.api [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146930, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073166} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2373.326823] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2373.327280] env[62824]: DEBUG oslo_vmware.api [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2373.327280] env[62824]: value = "task-2146933" [ 2373.327280] env[62824]: _type = "Task" [ 2373.327280] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2373.328144] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c93e1aad-2037-471a-87b3-0e84b1127e37 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2373.349134] env[62824]: DEBUG oslo_vmware.api [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146933, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2373.372136] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Reconfiguring VM instance instance-00000066 to attach disk [datastore2] 3d294ee9-92f6-4115-95db-9d4b66562b56/3d294ee9-92f6-4115-95db-9d4b66562b56.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2373.376695] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4a481f4c-4d42-4096-8e71-7b4f37fc7fe3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2373.399422] env[62824]: DEBUG oslo_vmware.api [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2146931, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2373.401040] env[62824]: DEBUG oslo_vmware.api [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2373.401040] env[62824]: value = "task-2146934" [ 2373.401040] env[62824]: _type = "Task" [ 2373.401040] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2373.416035] env[62824]: DEBUG oslo_vmware.api [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146934, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2373.421551] env[62824]: DEBUG oslo_vmware.api [None req-09c485ce-ef98-43f8-a6af-31be637f0d94 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146932, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2373.534986] env[62824]: DEBUG nova.network.neutron [req-e1ac7983-616b-4c48-8b20-9cb7de8fb175 req-e5129aa6-20ea-4a2b-898a-fb517da29a21 service nova] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Updated VIF entry in instance network info cache for port 3c1f3341-20a4-4b77-8dae-12ef5901a16e. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2373.535407] env[62824]: DEBUG nova.network.neutron [req-e1ac7983-616b-4c48-8b20-9cb7de8fb175 req-e5129aa6-20ea-4a2b-898a-fb517da29a21 service nova] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Updating instance_info_cache with network_info: [{"id": "3c1f3341-20a4-4b77-8dae-12ef5901a16e", "address": "fa:16:3e:1c:3a:8c", "network": {"id": "aa1c09e5-8d72-43ad-a903-a0f5e711da80", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1643616572-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.166", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c987c5f844be42c99cc9d57a9ca84c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9bc2632-36f9-4912-8782-8bbb789f909d", "external-id": "nsx-vlan-transportzone-897", "segmentation_id": 897, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3c1f3341-20", "ovs_interfaceid": "3c1f3341-20a4-4b77-8dae-12ef5901a16e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2373.788031] env[62824]: INFO nova.compute.manager [-] [instance: a6cd2032-de60-4f78-bf1e-79801d049df0] Took 1.32 seconds to deallocate network for instance. [ 2373.842164] env[62824]: DEBUG oslo_vmware.api [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146933, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2373.875582] env[62824]: DEBUG oslo_vmware.api [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2146931, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2373.914629] env[62824]: DEBUG oslo_vmware.api [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146934, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2373.921255] env[62824]: DEBUG oslo_vmware.api [None req-09c485ce-ef98-43f8-a6af-31be637f0d94 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146932, 'name': PowerOnVM_Task, 'duration_secs': 0.615193} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2373.921591] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-09c485ce-ef98-43f8-a6af-31be637f0d94 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Resumed the VM {{(pid=62824) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1188}} [ 2373.921794] env[62824]: DEBUG nova.compute.manager [None req-09c485ce-ef98-43f8-a6af-31be637f0d94 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2373.922668] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08c1cba0-e661-45aa-9417-a00c1baf1187 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2374.038300] env[62824]: DEBUG oslo_concurrency.lockutils [req-e1ac7983-616b-4c48-8b20-9cb7de8fb175 req-e5129aa6-20ea-4a2b-898a-fb517da29a21 service nova] Releasing lock "refresh_cache-511ef5bb-1dca-4eae-a0f1-19d1d7a17371" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2374.039257] env[62824]: DEBUG oslo_concurrency.lockutils [None req-97300fec-1d35-4a49-954c-b0818a0040aa tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquired lock "refresh_cache-511ef5bb-1dca-4eae-a0f1-19d1d7a17371" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2374.039472] env[62824]: DEBUG nova.network.neutron [None req-97300fec-1d35-4a49-954c-b0818a0040aa tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2374.213174] env[62824]: DEBUG nova.compute.manager [req-b32baecb-7d1a-4c62-b075-2fb4a7a9411a req-9ca3fc98-c8db-4d7c-b7ca-5a9323be3280 service nova] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Received event network-changed-afcd2bdb-2bfc-40d7-a8e5-2a417be0e0e9 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2374.213693] env[62824]: DEBUG nova.compute.manager [req-b32baecb-7d1a-4c62-b075-2fb4a7a9411a req-9ca3fc98-c8db-4d7c-b7ca-5a9323be3280 service nova] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Refreshing instance network info cache due to event network-changed-afcd2bdb-2bfc-40d7-a8e5-2a417be0e0e9. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2374.213920] env[62824]: DEBUG oslo_concurrency.lockutils [req-b32baecb-7d1a-4c62-b075-2fb4a7a9411a req-9ca3fc98-c8db-4d7c-b7ca-5a9323be3280 service nova] Acquiring lock "refresh_cache-511ef5bb-1dca-4eae-a0f1-19d1d7a17371" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2374.296409] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5d6a43b2-91e7-4cee-b3f7-f24637b92565 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2374.296794] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5d6a43b2-91e7-4cee-b3f7-f24637b92565 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2374.297125] env[62824]: DEBUG nova.objects.instance [None req-5d6a43b2-91e7-4cee-b3f7-f24637b92565 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lazy-loading 'resources' on Instance uuid a6cd2032-de60-4f78-bf1e-79801d049df0 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2374.344828] env[62824]: DEBUG oslo_vmware.api [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146933, 'name': PowerOnVM_Task} progress is 71%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2374.373414] env[62824]: DEBUG oslo_vmware.api [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2146931, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2374.413396] env[62824]: DEBUG oslo_vmware.api [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146934, 'name': ReconfigVM_Task, 'duration_secs': 0.774774} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2374.413733] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Reconfigured VM instance instance-00000066 to attach disk [datastore2] 3d294ee9-92f6-4115-95db-9d4b66562b56/3d294ee9-92f6-4115-95db-9d4b66562b56.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2374.415773] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'encryption_options': None, 'device_type': 'disk', 'boot_index': 0, 'disk_bus': None, 'encryption_secret_uuid': None, 'size': 0, 'encrypted': False, 'guest_format': None, 'encryption_format': None, 'device_name': '/dev/sda', 'image_id': '9e2a7d30-212d-4ab8-9606-c5c6d52629e8'}], 'ephemerals': [], 'block_device_mapping': [{'device_type': None, 'boot_index': None, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-438761', 'volume_id': '42b59773-27a4-492d-b430-c629b14a2500', 'name': 'volume-42b59773-27a4-492d-b430-c629b14a2500', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '3d294ee9-92f6-4115-95db-9d4b66562b56', 'attached_at': '', 'detached_at': '', 'volume_id': '42b59773-27a4-492d-b430-c629b14a2500', 'serial': '42b59773-27a4-492d-b430-c629b14a2500'}, 'disk_bus': None, 'mount_device': '/dev/sdb', 'delete_on_termination': False, 'attachment_id': '12710dd9-9937-4c84-8856-b560c4a74c35', 'guest_format': None, 'volume_type': None}], 'swap': None} {{(pid=62824) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 2374.415773] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Volume attach. Driver type: vmdk {{(pid=62824) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 2374.415773] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-438761', 'volume_id': '42b59773-27a4-492d-b430-c629b14a2500', 'name': 'volume-42b59773-27a4-492d-b430-c629b14a2500', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '3d294ee9-92f6-4115-95db-9d4b66562b56', 'attached_at': '', 'detached_at': '', 'volume_id': '42b59773-27a4-492d-b430-c629b14a2500', 'serial': '42b59773-27a4-492d-b430-c629b14a2500'} {{(pid=62824) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 2374.416622] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00c8470c-f6ef-4d46-b4ee-a61f3a11a815 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2374.434864] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5b5a3fe-23f4-4136-b00b-1e842eaddd06 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2374.465421] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Reconfiguring VM instance instance-00000066 to attach disk [datastore2] volume-42b59773-27a4-492d-b430-c629b14a2500/volume-42b59773-27a4-492d-b430-c629b14a2500.vmdk or device None with type thin {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2374.467097] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4504c611-1d88-4b70-b527-9018b327745c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2374.492114] env[62824]: DEBUG oslo_vmware.api [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2374.492114] env[62824]: value = "task-2146935" [ 2374.492114] env[62824]: _type = "Task" [ 2374.492114] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2374.502404] env[62824]: DEBUG oslo_vmware.api [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146935, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2374.596917] env[62824]: WARNING nova.network.neutron [None req-97300fec-1d35-4a49-954c-b0818a0040aa tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] aa1c09e5-8d72-43ad-a903-a0f5e711da80 already exists in list: networks containing: ['aa1c09e5-8d72-43ad-a903-a0f5e711da80']. ignoring it [ 2374.843765] env[62824]: DEBUG oslo_vmware.api [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146933, 'name': PowerOnVM_Task, 'duration_secs': 1.332203} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2374.844087] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2374.844323] env[62824]: INFO nova.compute.manager [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Took 17.71 seconds to spawn the instance on the hypervisor. [ 2374.844509] env[62824]: DEBUG nova.compute.manager [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2374.845373] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f908932c-c84c-42b1-b40f-658eb019f8c4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2374.873463] env[62824]: DEBUG oslo_vmware.api [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2146931, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.545319} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2374.876132] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] c70880c9-a69a-49b9-b028-189c04276623/c70880c9-a69a-49b9-b028-189c04276623.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2374.876434] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c70880c9-a69a-49b9-b028-189c04276623] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2374.876751] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3830f921-619c-4747-a2b6-8055341073a7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2374.885786] env[62824]: DEBUG oslo_vmware.api [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Waiting for the task: (returnval){ [ 2374.885786] env[62824]: value = "task-2146936" [ 2374.885786] env[62824]: _type = "Task" [ 2374.885786] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2374.894770] env[62824]: DEBUG oslo_vmware.api [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2146936, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2375.004440] env[62824]: DEBUG oslo_vmware.api [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146935, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2375.123026] env[62824]: DEBUG nova.network.neutron [None req-97300fec-1d35-4a49-954c-b0818a0040aa tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Updating instance_info_cache with network_info: [{"id": "3c1f3341-20a4-4b77-8dae-12ef5901a16e", "address": "fa:16:3e:1c:3a:8c", "network": {"id": "aa1c09e5-8d72-43ad-a903-a0f5e711da80", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1643616572-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.166", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c987c5f844be42c99cc9d57a9ca84c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9bc2632-36f9-4912-8782-8bbb789f909d", "external-id": "nsx-vlan-transportzone-897", "segmentation_id": 897, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3c1f3341-20", "ovs_interfaceid": "3c1f3341-20a4-4b77-8dae-12ef5901a16e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "afcd2bdb-2bfc-40d7-a8e5-2a417be0e0e9", "address": "fa:16:3e:06:1e:c0", "network": {"id": "aa1c09e5-8d72-43ad-a903-a0f5e711da80", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1643616572-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c987c5f844be42c99cc9d57a9ca84c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9bc2632-36f9-4912-8782-8bbb789f909d", "external-id": "nsx-vlan-transportzone-897", "segmentation_id": 897, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapafcd2bdb-2b", "ovs_interfaceid": "afcd2bdb-2bfc-40d7-a8e5-2a417be0e0e9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2375.191766] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d2487c6-5998-4a91-9726-df14ee49039f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2375.200394] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f62544ac-39b2-41d9-9d97-bbffb6d82e69 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2375.231219] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e92aa286-b4f7-4f57-8689-5f0d0442fc8e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2375.239652] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-396d3af3-10b0-4c04-bd0d-5da8688d038f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2375.254400] env[62824]: DEBUG nova.compute.provider_tree [None req-5d6a43b2-91e7-4cee-b3f7-f24637b92565 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Updating inventory in ProviderTree for provider bf95157b-a274-42de-9ccf-9851128a44a1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 2375.362488] env[62824]: INFO nova.compute.manager [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Took 23.12 seconds to build instance. [ 2375.402023] env[62824]: DEBUG oslo_vmware.api [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2146936, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.240677} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2375.402023] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c70880c9-a69a-49b9-b028-189c04276623] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2375.402023] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d14ff10-a708-40b7-9ecf-a2b9d29a5979 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2375.424817] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c70880c9-a69a-49b9-b028-189c04276623] Reconfiguring VM instance instance-0000006e to attach disk [datastore2] c70880c9-a69a-49b9-b028-189c04276623/c70880c9-a69a-49b9-b028-189c04276623.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2375.424817] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-96dfeb3c-a661-4c35-a9b6-5e0321cdb820 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2375.445914] env[62824]: DEBUG oslo_vmware.api [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Waiting for the task: (returnval){ [ 2375.445914] env[62824]: value = "task-2146937" [ 2375.445914] env[62824]: _type = "Task" [ 2375.445914] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2375.455342] env[62824]: DEBUG oslo_vmware.api [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2146937, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2375.502326] env[62824]: DEBUG oslo_vmware.api [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146935, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2375.626129] env[62824]: DEBUG oslo_concurrency.lockutils [None req-97300fec-1d35-4a49-954c-b0818a0040aa tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Releasing lock "refresh_cache-511ef5bb-1dca-4eae-a0f1-19d1d7a17371" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2375.627155] env[62824]: DEBUG oslo_concurrency.lockutils [None req-97300fec-1d35-4a49-954c-b0818a0040aa tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquiring lock "511ef5bb-1dca-4eae-a0f1-19d1d7a17371" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2375.627405] env[62824]: DEBUG oslo_concurrency.lockutils [None req-97300fec-1d35-4a49-954c-b0818a0040aa tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquired lock "511ef5bb-1dca-4eae-a0f1-19d1d7a17371" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2375.627830] env[62824]: DEBUG oslo_concurrency.lockutils [req-b32baecb-7d1a-4c62-b075-2fb4a7a9411a req-9ca3fc98-c8db-4d7c-b7ca-5a9323be3280 service nova] Acquired lock "refresh_cache-511ef5bb-1dca-4eae-a0f1-19d1d7a17371" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2375.628131] env[62824]: DEBUG nova.network.neutron [req-b32baecb-7d1a-4c62-b075-2fb4a7a9411a req-9ca3fc98-c8db-4d7c-b7ca-5a9323be3280 service nova] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Refreshing network info cache for port afcd2bdb-2bfc-40d7-a8e5-2a417be0e0e9 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2375.630230] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bea4e001-388b-40fb-9786-17a7cc0726e7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2375.654939] env[62824]: DEBUG nova.virt.hardware [None req-97300fec-1d35-4a49-954c-b0818a0040aa tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2375.655371] env[62824]: DEBUG nova.virt.hardware [None req-97300fec-1d35-4a49-954c-b0818a0040aa tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2375.655620] env[62824]: DEBUG nova.virt.hardware [None req-97300fec-1d35-4a49-954c-b0818a0040aa tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2375.655880] env[62824]: DEBUG nova.virt.hardware [None req-97300fec-1d35-4a49-954c-b0818a0040aa tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2375.656104] env[62824]: DEBUG nova.virt.hardware [None req-97300fec-1d35-4a49-954c-b0818a0040aa tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2375.656327] env[62824]: DEBUG nova.virt.hardware [None req-97300fec-1d35-4a49-954c-b0818a0040aa tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2375.656580] env[62824]: DEBUG nova.virt.hardware [None req-97300fec-1d35-4a49-954c-b0818a0040aa tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2375.656793] env[62824]: DEBUG nova.virt.hardware [None req-97300fec-1d35-4a49-954c-b0818a0040aa tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2375.657050] env[62824]: DEBUG nova.virt.hardware [None req-97300fec-1d35-4a49-954c-b0818a0040aa tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2375.657248] env[62824]: DEBUG nova.virt.hardware [None req-97300fec-1d35-4a49-954c-b0818a0040aa tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2375.657904] env[62824]: DEBUG nova.virt.hardware [None req-97300fec-1d35-4a49-954c-b0818a0040aa tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2375.667977] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-97300fec-1d35-4a49-954c-b0818a0040aa tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Reconfiguring VM to attach interface {{(pid=62824) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 2375.668974] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cc0b6c6c-b3b6-4078-a6e4-1972ad3a8508 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2375.701084] env[62824]: DEBUG oslo_vmware.api [None req-97300fec-1d35-4a49-954c-b0818a0040aa tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Waiting for the task: (returnval){ [ 2375.701084] env[62824]: value = "task-2146938" [ 2375.701084] env[62824]: _type = "Task" [ 2375.701084] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2375.710524] env[62824]: DEBUG oslo_vmware.api [None req-97300fec-1d35-4a49-954c-b0818a0040aa tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146938, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2375.796378] env[62824]: DEBUG nova.scheduler.client.report [None req-5d6a43b2-91e7-4cee-b3f7-f24637b92565 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Updated inventory for provider bf95157b-a274-42de-9ccf-9851128a44a1 with generation 141 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 2375.796753] env[62824]: DEBUG nova.compute.provider_tree [None req-5d6a43b2-91e7-4cee-b3f7-f24637b92565 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Updating resource provider bf95157b-a274-42de-9ccf-9851128a44a1 generation from 141 to 142 during operation: update_inventory {{(pid=62824) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 2375.796987] env[62824]: DEBUG nova.compute.provider_tree [None req-5d6a43b2-91e7-4cee-b3f7-f24637b92565 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Updating inventory in ProviderTree for provider bf95157b-a274-42de-9ccf-9851128a44a1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 2375.864336] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1b274dd3-2843-49bb-bcb7-5ed2b8c50c6b tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "21a7eed6-c146-4911-9a0a-a1e1644e8e9d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.635s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2375.957786] env[62824]: DEBUG oslo_vmware.api [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2146937, 'name': ReconfigVM_Task, 'duration_secs': 0.405964} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2375.957786] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c70880c9-a69a-49b9-b028-189c04276623] Reconfigured VM instance instance-0000006e to attach disk [datastore2] c70880c9-a69a-49b9-b028-189c04276623/c70880c9-a69a-49b9-b028-189c04276623.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2375.958162] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f5baa4cd-9fe2-4d64-b3b0-a7351e369a40 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2375.965358] env[62824]: DEBUG oslo_vmware.api [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Waiting for the task: (returnval){ [ 2375.965358] env[62824]: value = "task-2146939" [ 2375.965358] env[62824]: _type = "Task" [ 2375.965358] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2375.973893] env[62824]: DEBUG oslo_vmware.api [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2146939, 'name': Rename_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2376.001624] env[62824]: DEBUG oslo_vmware.api [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146935, 'name': ReconfigVM_Task, 'duration_secs': 1.296402} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2376.002024] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Reconfigured VM instance instance-00000066 to attach disk [datastore2] volume-42b59773-27a4-492d-b430-c629b14a2500/volume-42b59773-27a4-492d-b430-c629b14a2500.vmdk or device None with type thin {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2376.006908] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-743c05cd-9bb0-4192-955b-cce058db5d0e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2376.023656] env[62824]: DEBUG oslo_vmware.api [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2376.023656] env[62824]: value = "task-2146940" [ 2376.023656] env[62824]: _type = "Task" [ 2376.023656] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2376.032956] env[62824]: DEBUG oslo_vmware.api [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146940, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2376.216653] env[62824]: DEBUG oslo_vmware.api [None req-97300fec-1d35-4a49-954c-b0818a0040aa tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146938, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2376.306352] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5d6a43b2-91e7-4cee-b3f7-f24637b92565 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.006s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2376.329105] env[62824]: INFO nova.scheduler.client.report [None req-5d6a43b2-91e7-4cee-b3f7-f24637b92565 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Deleted allocations for instance a6cd2032-de60-4f78-bf1e-79801d049df0 [ 2376.386017] env[62824]: DEBUG nova.network.neutron [req-b32baecb-7d1a-4c62-b075-2fb4a7a9411a req-9ca3fc98-c8db-4d7c-b7ca-5a9323be3280 service nova] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Updated VIF entry in instance network info cache for port afcd2bdb-2bfc-40d7-a8e5-2a417be0e0e9. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2376.387019] env[62824]: DEBUG nova.network.neutron [req-b32baecb-7d1a-4c62-b075-2fb4a7a9411a req-9ca3fc98-c8db-4d7c-b7ca-5a9323be3280 service nova] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Updating instance_info_cache with network_info: [{"id": "3c1f3341-20a4-4b77-8dae-12ef5901a16e", "address": "fa:16:3e:1c:3a:8c", "network": {"id": "aa1c09e5-8d72-43ad-a903-a0f5e711da80", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1643616572-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.166", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c987c5f844be42c99cc9d57a9ca84c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9bc2632-36f9-4912-8782-8bbb789f909d", "external-id": "nsx-vlan-transportzone-897", "segmentation_id": 897, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3c1f3341-20", "ovs_interfaceid": "3c1f3341-20a4-4b77-8dae-12ef5901a16e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "afcd2bdb-2bfc-40d7-a8e5-2a417be0e0e9", "address": "fa:16:3e:06:1e:c0", "network": {"id": "aa1c09e5-8d72-43ad-a903-a0f5e711da80", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1643616572-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c987c5f844be42c99cc9d57a9ca84c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9bc2632-36f9-4912-8782-8bbb789f909d", "external-id": "nsx-vlan-transportzone-897", "segmentation_id": 897, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapafcd2bdb-2b", "ovs_interfaceid": "afcd2bdb-2bfc-40d7-a8e5-2a417be0e0e9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2376.479050] env[62824]: DEBUG oslo_vmware.api [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2146939, 'name': Rename_Task, 'duration_secs': 0.164694} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2376.479405] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c70880c9-a69a-49b9-b028-189c04276623] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2376.479700] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d587b833-c87c-4a2a-9882-1f84477166b5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2376.490638] env[62824]: DEBUG oslo_vmware.api [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Waiting for the task: (returnval){ [ 2376.490638] env[62824]: value = "task-2146941" [ 2376.490638] env[62824]: _type = "Task" [ 2376.490638] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2376.505240] env[62824]: DEBUG oslo_vmware.api [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2146941, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2376.536865] env[62824]: DEBUG oslo_vmware.api [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146940, 'name': ReconfigVM_Task, 'duration_secs': 0.169957} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2376.537546] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-438761', 'volume_id': '42b59773-27a4-492d-b430-c629b14a2500', 'name': 'volume-42b59773-27a4-492d-b430-c629b14a2500', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '3d294ee9-92f6-4115-95db-9d4b66562b56', 'attached_at': '', 'detached_at': '', 'volume_id': '42b59773-27a4-492d-b430-c629b14a2500', 'serial': '42b59773-27a4-492d-b430-c629b14a2500'} {{(pid=62824) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 2376.538388] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b5a506ee-e7b8-4903-8149-5d876c9e578c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2376.546700] env[62824]: DEBUG oslo_vmware.api [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2376.546700] env[62824]: value = "task-2146942" [ 2376.546700] env[62824]: _type = "Task" [ 2376.546700] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2376.557830] env[62824]: DEBUG oslo_vmware.api [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146942, 'name': Rename_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2376.720112] env[62824]: DEBUG oslo_vmware.api [None req-97300fec-1d35-4a49-954c-b0818a0040aa tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146938, 'name': ReconfigVM_Task, 'duration_secs': 0.724499} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2376.720663] env[62824]: DEBUG oslo_concurrency.lockutils [None req-97300fec-1d35-4a49-954c-b0818a0040aa tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Releasing lock "511ef5bb-1dca-4eae-a0f1-19d1d7a17371" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2376.720903] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-97300fec-1d35-4a49-954c-b0818a0040aa tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Reconfigured VM to attach interface {{(pid=62824) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 2376.801387] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2376.836975] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5d6a43b2-91e7-4cee-b3f7-f24637b92565 tempest-ServersTestJSON-916807641 tempest-ServersTestJSON-916807641-project-member] Lock "a6cd2032-de60-4f78-bf1e-79801d049df0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.191s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2376.891276] env[62824]: DEBUG oslo_concurrency.lockutils [req-b32baecb-7d1a-4c62-b075-2fb4a7a9411a req-9ca3fc98-c8db-4d7c-b7ca-5a9323be3280 service nova] Releasing lock "refresh_cache-511ef5bb-1dca-4eae-a0f1-19d1d7a17371" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2376.891276] env[62824]: DEBUG nova.compute.manager [req-b32baecb-7d1a-4c62-b075-2fb4a7a9411a req-9ca3fc98-c8db-4d7c-b7ca-5a9323be3280 service nova] [instance: a6cd2032-de60-4f78-bf1e-79801d049df0] Received event network-vif-deleted-018929f6-4f6c-4148-b1eb-9cf11f5cc746 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2377.006594] env[62824]: DEBUG oslo_vmware.api [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2146941, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2377.059314] env[62824]: DEBUG oslo_vmware.api [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146942, 'name': Rename_Task, 'duration_secs': 0.220157} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2377.060682] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2377.060682] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-897e792c-8be1-48bc-ad20-e404ef62faf0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2377.069362] env[62824]: DEBUG oslo_vmware.api [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2377.069362] env[62824]: value = "task-2146943" [ 2377.069362] env[62824]: _type = "Task" [ 2377.069362] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2377.080341] env[62824]: DEBUG oslo_vmware.api [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146943, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2377.226918] env[62824]: DEBUG oslo_concurrency.lockutils [None req-97300fec-1d35-4a49-954c-b0818a0040aa tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lock "interface-511ef5bb-1dca-4eae-a0f1-19d1d7a17371-afcd2bdb-2bfc-40d7-a8e5-2a417be0e0e9" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 8.028s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2377.507579] env[62824]: DEBUG oslo_vmware.api [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2146941, 'name': PowerOnVM_Task, 'duration_secs': 0.716146} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2377.508312] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c70880c9-a69a-49b9-b028-189c04276623] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2377.508533] env[62824]: INFO nova.compute.manager [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c70880c9-a69a-49b9-b028-189c04276623] Took 14.42 seconds to spawn the instance on the hypervisor. [ 2377.508762] env[62824]: DEBUG nova.compute.manager [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c70880c9-a69a-49b9-b028-189c04276623] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2377.509748] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fe26c5d-5b6f-43db-91ee-da52263b058a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2377.584690] env[62824]: DEBUG oslo_vmware.api [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146943, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2377.772913] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2377.774472] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager.update_available_resource {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2378.038027] env[62824]: INFO nova.compute.manager [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c70880c9-a69a-49b9-b028-189c04276623] Took 19.73 seconds to build instance. [ 2378.084703] env[62824]: DEBUG oslo_vmware.api [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146943, 'name': PowerOnVM_Task, 'duration_secs': 0.990524} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2378.085194] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2378.085492] env[62824]: DEBUG nova.compute.manager [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2378.086891] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80a7fdd6-e822-4f10-99cb-20437144e232 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2378.220941] env[62824]: DEBUG nova.compute.manager [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Stashing vm_state: active {{(pid=62824) _prep_resize /opt/stack/nova/nova/compute/manager.py:6136}} [ 2378.276293] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2378.276527] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2378.276706] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2378.276863] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62824) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 2378.278287] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8641ec93-6771-447c-9a00-8982839d5081 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2378.289586] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db7c23ab-3569-4e68-beba-742b49b8142f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2378.308264] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b926ce6-6023-4ab2-8d1d-ed1b55933c6b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2378.320537] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acdc79bc-a7e8-48d8-90cb-ee1652b4aac6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2378.356047] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179663MB free_disk=174GB free_vcpus=48 pci_devices=None {{(pid=62824) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 2378.356229] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2378.356560] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2378.543023] env[62824]: DEBUG oslo_concurrency.lockutils [None req-4e633ad3-b140-40c7-8171-d8c211ac1787 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Lock "c70880c9-a69a-49b9-b028-189c04276623" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.245s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2378.612627] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2378.743563] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2378.993235] env[62824]: DEBUG oslo_concurrency.lockutils [None req-926ba5af-9099-485e-806f-e6c1b328a404 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquiring lock "interface-511ef5bb-1dca-4eae-a0f1-19d1d7a17371-afcd2bdb-2bfc-40d7-a8e5-2a417be0e0e9" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2378.993235] env[62824]: DEBUG oslo_concurrency.lockutils [None req-926ba5af-9099-485e-806f-e6c1b328a404 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lock "interface-511ef5bb-1dca-4eae-a0f1-19d1d7a17371-afcd2bdb-2bfc-40d7-a8e5-2a417be0e0e9" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2379.368259] env[62824]: INFO nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Updating resource usage from migration ea8a1301-70c4-4e5c-bb7f-97e03c659c29 [ 2379.386460] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance ca42fef0-1e90-4ab6-9d60-8ef7e4997884 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2379.386611] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 5b3df31f-15fe-473f-992c-ddb272661c53 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2379.386731] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2379.386850] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance e1574266-e53b-4169-9db6-c66b84895edb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2379.386962] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 3d294ee9-92f6-4115-95db-9d4b66562b56 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2379.387086] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 1db9ff97-e59c-43f6-984e-6405d8e993a4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2379.387200] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 511ef5bb-1dca-4eae-a0f1-19d1d7a17371 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2379.387712] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 71e9733b-c84b-4501-8faf-a487f089b498 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2379.387712] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance c70880c9-a69a-49b9-b028-189c04276623 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2379.387712] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Migration ea8a1301-70c4-4e5c-bb7f-97e03c659c29 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 2379.387712] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 21a7eed6-c146-4911-9a0a-a1e1644e8e9d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2379.496417] env[62824]: DEBUG oslo_concurrency.lockutils [None req-926ba5af-9099-485e-806f-e6c1b328a404 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquiring lock "511ef5bb-1dca-4eae-a0f1-19d1d7a17371" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2379.496639] env[62824]: DEBUG oslo_concurrency.lockutils [None req-926ba5af-9099-485e-806f-e6c1b328a404 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquired lock "511ef5bb-1dca-4eae-a0f1-19d1d7a17371" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2379.497948] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76670f60-14c8-4b10-b354-3052b8b8a605 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2379.530464] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbf41edd-2876-4bbe-ae7a-a4086139097d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2379.565927] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-926ba5af-9099-485e-806f-e6c1b328a404 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Reconfiguring VM to detach interface {{(pid=62824) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 2379.568292] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3f25ed7a-eaa4-486a-bcd9-9329a56bac1c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2379.583567] env[62824]: DEBUG nova.compute.manager [req-c17a263c-394b-498f-ae17-8e0a58f7765d req-eb6d9882-65c2-47ed-a007-d4bba21b940e service nova] [instance: c70880c9-a69a-49b9-b028-189c04276623] Received event network-changed-f71d9cbc-c160-47f9-9771-af38641d3683 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2379.583810] env[62824]: DEBUG nova.compute.manager [req-c17a263c-394b-498f-ae17-8e0a58f7765d req-eb6d9882-65c2-47ed-a007-d4bba21b940e service nova] [instance: c70880c9-a69a-49b9-b028-189c04276623] Refreshing instance network info cache due to event network-changed-f71d9cbc-c160-47f9-9771-af38641d3683. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2379.584085] env[62824]: DEBUG oslo_concurrency.lockutils [req-c17a263c-394b-498f-ae17-8e0a58f7765d req-eb6d9882-65c2-47ed-a007-d4bba21b940e service nova] Acquiring lock "refresh_cache-c70880c9-a69a-49b9-b028-189c04276623" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2379.584306] env[62824]: DEBUG oslo_concurrency.lockutils [req-c17a263c-394b-498f-ae17-8e0a58f7765d req-eb6d9882-65c2-47ed-a007-d4bba21b940e service nova] Acquired lock "refresh_cache-c70880c9-a69a-49b9-b028-189c04276623" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2379.584511] env[62824]: DEBUG nova.network.neutron [req-c17a263c-394b-498f-ae17-8e0a58f7765d req-eb6d9882-65c2-47ed-a007-d4bba21b940e service nova] [instance: c70880c9-a69a-49b9-b028-189c04276623] Refreshing network info cache for port f71d9cbc-c160-47f9-9771-af38641d3683 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2379.589217] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquiring lock "7c47bb92-a58e-4537-a68a-289e2a5fc19a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2379.589397] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lock "7c47bb92-a58e-4537-a68a-289e2a5fc19a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2379.598592] env[62824]: DEBUG oslo_vmware.api [None req-926ba5af-9099-485e-806f-e6c1b328a404 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Waiting for the task: (returnval){ [ 2379.598592] env[62824]: value = "task-2146944" [ 2379.598592] env[62824]: _type = "Task" [ 2379.598592] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2379.612442] env[62824]: DEBUG oslo_vmware.api [None req-926ba5af-9099-485e-806f-e6c1b328a404 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146944, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2379.890690] env[62824]: INFO nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 7c47bb92-a58e-4537-a68a-289e2a5fc19a has allocations against this compute host but is not found in the database. [ 2379.890944] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Total usable vcpus: 48, total allocated vcpus: 10 {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 2379.891266] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2432MB phys_disk=200GB used_disk=10GB total_vcpus=48 used_vcpus=10 pci_stats=[] {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 2380.065731] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ebc5798-bcb8-430d-9f80-9957655f5c1e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2380.076311] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60bf961e-7fec-41fc-8939-7dd45f3c18b9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2380.112475] env[62824]: DEBUG nova.compute.manager [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2380.119757] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92633364-a869-487e-8a3b-b981983eae4b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2380.129765] env[62824]: DEBUG oslo_vmware.api [None req-926ba5af-9099-485e-806f-e6c1b328a404 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146944, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2380.133631] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9c8ef33-15c4-404d-b4e3-6eaff6138f3b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2380.149514] env[62824]: DEBUG nova.compute.provider_tree [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2380.368710] env[62824]: DEBUG nova.network.neutron [req-c17a263c-394b-498f-ae17-8e0a58f7765d req-eb6d9882-65c2-47ed-a007-d4bba21b940e service nova] [instance: c70880c9-a69a-49b9-b028-189c04276623] Updated VIF entry in instance network info cache for port f71d9cbc-c160-47f9-9771-af38641d3683. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2380.369206] env[62824]: DEBUG nova.network.neutron [req-c17a263c-394b-498f-ae17-8e0a58f7765d req-eb6d9882-65c2-47ed-a007-d4bba21b940e service nova] [instance: c70880c9-a69a-49b9-b028-189c04276623] Updating instance_info_cache with network_info: [{"id": "f71d9cbc-c160-47f9-9771-af38641d3683", "address": "fa:16:3e:51:89:60", "network": {"id": "f7936d74-eeef-4c41-8ae7-9427c43663fa", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1643883257-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.154", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c95aa87980de4ec8a2170c2640423301", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6eaa481-1f92-4851-b98e-09ed0daad7cb", "external-id": "nsx-vlan-transportzone-636", "segmentation_id": 636, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf71d9cbc-c1", "ovs_interfaceid": "f71d9cbc-c160-47f9-9771-af38641d3683", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2380.632016] env[62824]: DEBUG oslo_vmware.api [None req-926ba5af-9099-485e-806f-e6c1b328a404 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146944, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2380.639987] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2380.653470] env[62824]: DEBUG nova.scheduler.client.report [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2380.872244] env[62824]: DEBUG oslo_concurrency.lockutils [req-c17a263c-394b-498f-ae17-8e0a58f7765d req-eb6d9882-65c2-47ed-a007-d4bba21b940e service nova] Releasing lock "refresh_cache-c70880c9-a69a-49b9-b028-189c04276623" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2381.001664] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fa321c8f-c411-4769-8cf0-625400daaa15 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Acquiring lock "71e9733b-c84b-4501-8faf-a487f089b498" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2381.002056] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fa321c8f-c411-4769-8cf0-625400daaa15 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Lock "71e9733b-c84b-4501-8faf-a487f089b498" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2381.002217] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fa321c8f-c411-4769-8cf0-625400daaa15 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Acquiring lock "71e9733b-c84b-4501-8faf-a487f089b498-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2381.002414] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fa321c8f-c411-4769-8cf0-625400daaa15 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Lock "71e9733b-c84b-4501-8faf-a487f089b498-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2381.002589] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fa321c8f-c411-4769-8cf0-625400daaa15 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Lock "71e9733b-c84b-4501-8faf-a487f089b498-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2381.004835] env[62824]: INFO nova.compute.manager [None req-fa321c8f-c411-4769-8cf0-625400daaa15 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Terminating instance [ 2381.126210] env[62824]: DEBUG oslo_vmware.api [None req-926ba5af-9099-485e-806f-e6c1b328a404 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146944, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2381.158765] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62824) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 2381.159017] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.803s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2381.159420] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 2.547s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2381.159612] env[62824]: DEBUG nova.objects.instance [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62824) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 2381.508131] env[62824]: DEBUG nova.compute.manager [None req-fa321c8f-c411-4769-8cf0-625400daaa15 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2381.509813] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-fa321c8f-c411-4769-8cf0-625400daaa15 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2381.509813] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08d2bc26-247a-4660-ad72-3fb2ccb32439 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2381.517937] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa321c8f-c411-4769-8cf0-625400daaa15 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2381.517937] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b43c18b7-fc7e-430a-8ffb-2150af841353 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2381.526252] env[62824]: DEBUG oslo_vmware.api [None req-fa321c8f-c411-4769-8cf0-625400daaa15 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Waiting for the task: (returnval){ [ 2381.526252] env[62824]: value = "task-2146945" [ 2381.526252] env[62824]: _type = "Task" [ 2381.526252] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2381.535438] env[62824]: DEBUG oslo_vmware.api [None req-fa321c8f-c411-4769-8cf0-625400daaa15 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146945, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2381.628300] env[62824]: DEBUG oslo_vmware.api [None req-926ba5af-9099-485e-806f-e6c1b328a404 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146944, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2382.036610] env[62824]: DEBUG oslo_vmware.api [None req-fa321c8f-c411-4769-8cf0-625400daaa15 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146945, 'name': PowerOffVM_Task, 'duration_secs': 0.283558} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2382.036910] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa321c8f-c411-4769-8cf0-625400daaa15 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2382.037070] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-fa321c8f-c411-4769-8cf0-625400daaa15 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2382.037343] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3b74734d-0c60-4b9f-acc6-65e8dce1f19c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2382.129134] env[62824]: DEBUG oslo_vmware.api [None req-926ba5af-9099-485e-806f-e6c1b328a404 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146944, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2382.161804] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2382.161965] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Starting heal instance info cache {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10445}} [ 2382.169242] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b537e8ad-56a5-4379-8980-b1bedccab634 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.010s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2382.170294] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 3.427s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2382.629468] env[62824]: DEBUG oslo_vmware.api [None req-926ba5af-9099-485e-806f-e6c1b328a404 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146944, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2382.674833] env[62824]: INFO nova.compute.claims [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2383.130586] env[62824]: DEBUG oslo_vmware.api [None req-926ba5af-9099-485e-806f-e6c1b328a404 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146944, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2383.181608] env[62824]: INFO nova.compute.resource_tracker [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Updating resource usage from migration ea8a1301-70c4-4e5c-bb7f-97e03c659c29 [ 2383.327776] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2be95676-b53b-4aa7-adb4-ca12e9850776 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2383.336200] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d98a743-7541-4f9a-86e5-76a54f029581 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2383.367031] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-059556e5-5d19-4910-8d7d-d0d41ca1a85f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2383.375152] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d4d7bb2-1ccc-4049-a6d8-822d3111b9bd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2383.388145] env[62824]: DEBUG nova.compute.provider_tree [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2383.629360] env[62824]: DEBUG oslo_vmware.api [None req-926ba5af-9099-485e-806f-e6c1b328a404 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146944, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2383.713932] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "refresh_cache-ca42fef0-1e90-4ab6-9d60-8ef7e4997884" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2383.714107] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquired lock "refresh_cache-ca42fef0-1e90-4ab6-9d60-8ef7e4997884" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2383.714269] env[62824]: DEBUG nova.network.neutron [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] Forcefully refreshing network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 2383.891165] env[62824]: DEBUG nova.scheduler.client.report [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2384.130479] env[62824]: DEBUG oslo_vmware.api [None req-926ba5af-9099-485e-806f-e6c1b328a404 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146944, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2384.396114] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.226s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2384.396520] env[62824]: INFO nova.compute.manager [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Migrating [ 2384.403519] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.764s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2384.405145] env[62824]: INFO nova.compute.claims [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2384.630715] env[62824]: DEBUG oslo_vmware.api [None req-926ba5af-9099-485e-806f-e6c1b328a404 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146944, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2384.918044] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquiring lock "refresh_cache-21a7eed6-c146-4911-9a0a-a1e1644e8e9d" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2384.918044] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquired lock "refresh_cache-21a7eed6-c146-4911-9a0a-a1e1644e8e9d" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2384.918238] env[62824]: DEBUG nova.network.neutron [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2384.935875] env[62824]: DEBUG nova.network.neutron [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] Updating instance_info_cache with network_info: [{"id": "3eb965bf-7ceb-4a51-bc6f-00803bc6b99e", "address": "fa:16:3e:3f:e9:09", "network": {"id": "b4ed285c-26a7-461b-8a71-d18e07582ab8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-997150974-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.185", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "815ecf9340934733a83d4dcb26612fd4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69f65356-c85e-4b7f-ad28-7c7b5e8cf50c", "external-id": "nsx-vlan-transportzone-281", "segmentation_id": 281, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3eb965bf-7c", "ovs_interfaceid": "3eb965bf-7ceb-4a51-bc6f-00803bc6b99e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2385.135094] env[62824]: DEBUG oslo_vmware.api [None req-926ba5af-9099-485e-806f-e6c1b328a404 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146944, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2385.438335] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Releasing lock "refresh_cache-ca42fef0-1e90-4ab6-9d60-8ef7e4997884" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2385.438666] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] Updated the network info_cache for instance {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10516}} [ 2385.438759] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2385.438976] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2385.439110] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2385.439220] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2385.439368] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2385.439499] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62824) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11064}} [ 2385.574234] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca988cac-4eaf-48a3-9b7a-a71526dd3ed0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2385.582733] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-234d8cd7-c113-4fd5-9040-902a64c23c89 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2385.616821] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-036f084d-ea46-42f4-93ef-7cbcba284cfd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2385.627767] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4c9e941-f550-421b-9174-bc7facf0b417 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2385.637764] env[62824]: DEBUG oslo_vmware.api [None req-926ba5af-9099-485e-806f-e6c1b328a404 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146944, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2385.645913] env[62824]: DEBUG nova.compute.provider_tree [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2385.652948] env[62824]: DEBUG nova.network.neutron [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Updating instance_info_cache with network_info: [{"id": "edaf5d3c-004d-4ba0-b037-e7eef2acbe44", "address": "fa:16:3e:06:2e:6a", "network": {"id": "12411632-b7f6-4544-b2e4-b04ae7337547", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-917106830-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4479921caf5f405b8fc49baad390a0e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "de7fa486-5f28-44ae-b0cf-72234ff87546", "external-id": "nsx-vlan-transportzone-229", "segmentation_id": 229, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapedaf5d3c-00", "ovs_interfaceid": "edaf5d3c-004d-4ba0-b037-e7eef2acbe44", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2386.135238] env[62824]: DEBUG oslo_vmware.api [None req-926ba5af-9099-485e-806f-e6c1b328a404 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146944, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2386.149466] env[62824]: DEBUG nova.scheduler.client.report [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2386.155550] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Releasing lock "refresh_cache-21a7eed6-c146-4911-9a0a-a1e1644e8e9d" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2386.635321] env[62824]: DEBUG oslo_vmware.api [None req-926ba5af-9099-485e-806f-e6c1b328a404 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146944, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2386.654202] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.251s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2386.654678] env[62824]: DEBUG nova.compute.manager [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2387.136406] env[62824]: DEBUG oslo_vmware.api [None req-926ba5af-9099-485e-806f-e6c1b328a404 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146944, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2387.159958] env[62824]: DEBUG nova.compute.utils [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2387.161841] env[62824]: DEBUG nova.compute.manager [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2387.161959] env[62824]: DEBUG nova.network.neutron [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2387.215153] env[62824]: DEBUG nova.policy [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e6ecc84f6587474f89419598f5c44dbb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0eb41836305847fca773afebdafa01db', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2387.487148] env[62824]: DEBUG nova.network.neutron [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Successfully created port: 821db814-549f-4b95-befb-cdb81b9bcb70 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2387.636835] env[62824]: DEBUG oslo_vmware.api [None req-926ba5af-9099-485e-806f-e6c1b328a404 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146944, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2387.664636] env[62824]: DEBUG nova.compute.manager [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2387.671933] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74b4a0ae-846e-493f-91d6-012c71cb0314 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2387.695023] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Updating instance '21a7eed6-c146-4911-9a0a-a1e1644e8e9d' progress to 0 {{(pid=62824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 2388.137426] env[62824]: DEBUG oslo_vmware.api [None req-926ba5af-9099-485e-806f-e6c1b328a404 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146944, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2388.201664] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2388.201793] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-34c8e383-d79c-48bd-8c6a-1cd6c7b34a6b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2388.209900] env[62824]: DEBUG oslo_vmware.api [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2388.209900] env[62824]: value = "task-2146947" [ 2388.209900] env[62824]: _type = "Task" [ 2388.209900] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2388.221200] env[62824]: DEBUG oslo_vmware.api [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146947, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2388.244401] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-fa321c8f-c411-4769-8cf0-625400daaa15 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2388.244670] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-fa321c8f-c411-4769-8cf0-625400daaa15 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Deleting contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2388.244835] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-fa321c8f-c411-4769-8cf0-625400daaa15 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Deleting the datastore file [datastore2] 71e9733b-c84b-4501-8faf-a487f089b498 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2388.245122] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a9685ac1-f5b4-4306-a6ec-8765b3025877 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2388.252550] env[62824]: DEBUG oslo_vmware.api [None req-fa321c8f-c411-4769-8cf0-625400daaa15 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Waiting for the task: (returnval){ [ 2388.252550] env[62824]: value = "task-2146948" [ 2388.252550] env[62824]: _type = "Task" [ 2388.252550] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2388.263214] env[62824]: DEBUG oslo_vmware.api [None req-fa321c8f-c411-4769-8cf0-625400daaa15 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146948, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2388.637020] env[62824]: DEBUG oslo_vmware.api [None req-926ba5af-9099-485e-806f-e6c1b328a404 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146944, 'name': ReconfigVM_Task, 'duration_secs': 8.854888} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2388.637294] env[62824]: DEBUG oslo_concurrency.lockutils [None req-926ba5af-9099-485e-806f-e6c1b328a404 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Releasing lock "511ef5bb-1dca-4eae-a0f1-19d1d7a17371" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2388.637510] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-926ba5af-9099-485e-806f-e6c1b328a404 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Reconfigured VM to detach interface {{(pid=62824) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 2388.676517] env[62824]: DEBUG nova.compute.manager [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2388.702616] env[62824]: DEBUG nova.virt.hardware [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2388.703050] env[62824]: DEBUG nova.virt.hardware [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2388.703397] env[62824]: DEBUG nova.virt.hardware [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2388.703613] env[62824]: DEBUG nova.virt.hardware [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2388.704103] env[62824]: DEBUG nova.virt.hardware [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2388.704103] env[62824]: DEBUG nova.virt.hardware [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2388.704217] env[62824]: DEBUG nova.virt.hardware [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2388.704290] env[62824]: DEBUG nova.virt.hardware [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2388.704458] env[62824]: DEBUG nova.virt.hardware [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2388.704624] env[62824]: DEBUG nova.virt.hardware [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2388.704796] env[62824]: DEBUG nova.virt.hardware [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2388.705675] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d53abf6f-42ef-4df5-afbd-ed91d8107555 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2388.716741] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e7a61dd-5e43-412c-93b4-41304e101b91 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2388.735836] env[62824]: DEBUG oslo_vmware.api [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146947, 'name': PowerOffVM_Task, 'duration_secs': 0.28691} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2388.735998] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2388.736113] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Updating instance '21a7eed6-c146-4911-9a0a-a1e1644e8e9d' progress to 17 {{(pid=62824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 2388.762883] env[62824]: DEBUG oslo_vmware.api [None req-fa321c8f-c411-4769-8cf0-625400daaa15 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Task: {'id': task-2146948, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.159609} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2388.763156] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-fa321c8f-c411-4769-8cf0-625400daaa15 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2388.763342] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-fa321c8f-c411-4769-8cf0-625400daaa15 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Deleted contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2388.763586] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-fa321c8f-c411-4769-8cf0-625400daaa15 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2388.763781] env[62824]: INFO nova.compute.manager [None req-fa321c8f-c411-4769-8cf0-625400daaa15 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Took 7.26 seconds to destroy the instance on the hypervisor. [ 2388.764036] env[62824]: DEBUG oslo.service.loopingcall [None req-fa321c8f-c411-4769-8cf0-625400daaa15 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2388.764232] env[62824]: DEBUG nova.compute.manager [-] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2388.764327] env[62824]: DEBUG nova.network.neutron [-] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2389.148651] env[62824]: DEBUG nova.compute.manager [req-f713f53d-c5d1-40ee-a08e-7aa892712813 req-577f41b9-cb1f-4709-8f4b-ef3478f32c82 service nova] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Received event network-vif-deleted-4214bf6e-29f0-4d92-be44-44fccf7ab149 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2389.148860] env[62824]: INFO nova.compute.manager [req-f713f53d-c5d1-40ee-a08e-7aa892712813 req-577f41b9-cb1f-4709-8f4b-ef3478f32c82 service nova] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Neutron deleted interface 4214bf6e-29f0-4d92-be44-44fccf7ab149; detaching it from the instance and deleting it from the info cache [ 2389.149049] env[62824]: DEBUG nova.network.neutron [req-f713f53d-c5d1-40ee-a08e-7aa892712813 req-577f41b9-cb1f-4709-8f4b-ef3478f32c82 service nova] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2389.242806] env[62824]: DEBUG nova.virt.hardware [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:24Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2389.243082] env[62824]: DEBUG nova.virt.hardware [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2389.243248] env[62824]: DEBUG nova.virt.hardware [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2389.243433] env[62824]: DEBUG nova.virt.hardware [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2389.243779] env[62824]: DEBUG nova.virt.hardware [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2389.243958] env[62824]: DEBUG nova.virt.hardware [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2389.244209] env[62824]: DEBUG nova.virt.hardware [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2389.244559] env[62824]: DEBUG nova.virt.hardware [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2389.244621] env[62824]: DEBUG nova.virt.hardware [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2389.244756] env[62824]: DEBUG nova.virt.hardware [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2389.244930] env[62824]: DEBUG nova.virt.hardware [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2389.250467] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-555345ea-2e02-4ff8-8d8b-5602eb559fe4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2389.268496] env[62824]: DEBUG oslo_vmware.api [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2389.268496] env[62824]: value = "task-2146949" [ 2389.268496] env[62824]: _type = "Task" [ 2389.268496] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2389.276993] env[62824]: DEBUG oslo_vmware.api [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146949, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2389.466205] env[62824]: DEBUG nova.network.neutron [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Successfully updated port: 821db814-549f-4b95-befb-cdb81b9bcb70 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2389.623621] env[62824]: DEBUG nova.network.neutron [-] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2389.651266] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-12833471-6c08-4ee1-a7db-85900d2679db {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2389.660512] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc0e531f-f579-464a-a5a6-16502d9276e8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2389.690251] env[62824]: DEBUG nova.compute.manager [req-f713f53d-c5d1-40ee-a08e-7aa892712813 req-577f41b9-cb1f-4709-8f4b-ef3478f32c82 service nova] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Detach interface failed, port_id=4214bf6e-29f0-4d92-be44-44fccf7ab149, reason: Instance 71e9733b-c84b-4501-8faf-a487f089b498 could not be found. {{(pid=62824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11482}} [ 2389.778730] env[62824]: DEBUG oslo_vmware.api [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146949, 'name': ReconfigVM_Task, 'duration_secs': 0.181668} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2389.779028] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Updating instance '21a7eed6-c146-4911-9a0a-a1e1644e8e9d' progress to 33 {{(pid=62824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 2389.969930] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquiring lock "refresh_cache-7c47bb92-a58e-4537-a68a-289e2a5fc19a" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2389.969930] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquired lock "refresh_cache-7c47bb92-a58e-4537-a68a-289e2a5fc19a" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2389.969930] env[62824]: DEBUG nova.network.neutron [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2390.072479] env[62824]: DEBUG oslo_concurrency.lockutils [None req-926ba5af-9099-485e-806f-e6c1b328a404 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquiring lock "refresh_cache-511ef5bb-1dca-4eae-a0f1-19d1d7a17371" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2390.072479] env[62824]: DEBUG oslo_concurrency.lockutils [None req-926ba5af-9099-485e-806f-e6c1b328a404 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquired lock "refresh_cache-511ef5bb-1dca-4eae-a0f1-19d1d7a17371" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2390.072675] env[62824]: DEBUG nova.network.neutron [None req-926ba5af-9099-485e-806f-e6c1b328a404 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2390.125764] env[62824]: INFO nova.compute.manager [-] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Took 1.36 seconds to deallocate network for instance. [ 2390.286065] env[62824]: DEBUG nova.virt.hardware [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2390.286065] env[62824]: DEBUG nova.virt.hardware [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2390.286065] env[62824]: DEBUG nova.virt.hardware [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2390.286065] env[62824]: DEBUG nova.virt.hardware [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2390.286482] env[62824]: DEBUG nova.virt.hardware [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2390.286482] env[62824]: DEBUG nova.virt.hardware [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2390.286595] env[62824]: DEBUG nova.virt.hardware [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2390.286705] env[62824]: DEBUG nova.virt.hardware [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2390.286870] env[62824]: DEBUG nova.virt.hardware [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2390.287047] env[62824]: DEBUG nova.virt.hardware [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2390.287225] env[62824]: DEBUG nova.virt.hardware [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2390.292522] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Reconfiguring VM instance instance-0000006d to detach disk 2000 {{(pid=62824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 2390.293327] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b77ccb97-f33d-48ed-8f73-e3ec72b8e777 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2390.311760] env[62824]: DEBUG oslo_vmware.api [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2390.311760] env[62824]: value = "task-2146950" [ 2390.311760] env[62824]: _type = "Task" [ 2390.311760] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2390.319233] env[62824]: DEBUG oslo_vmware.api [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146950, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2390.501108] env[62824]: DEBUG nova.network.neutron [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2390.508760] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._sync_power_states {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2390.528087] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1304009c-16b0-48c1-b409-6a50c077ff73 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquiring lock "511ef5bb-1dca-4eae-a0f1-19d1d7a17371" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2390.528341] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1304009c-16b0-48c1-b409-6a50c077ff73 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lock "511ef5bb-1dca-4eae-a0f1-19d1d7a17371" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2390.528546] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1304009c-16b0-48c1-b409-6a50c077ff73 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquiring lock "511ef5bb-1dca-4eae-a0f1-19d1d7a17371-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2390.528726] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1304009c-16b0-48c1-b409-6a50c077ff73 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lock "511ef5bb-1dca-4eae-a0f1-19d1d7a17371-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2390.528890] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1304009c-16b0-48c1-b409-6a50c077ff73 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lock "511ef5bb-1dca-4eae-a0f1-19d1d7a17371-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2390.531896] env[62824]: INFO nova.compute.manager [None req-1304009c-16b0-48c1-b409-6a50c077ff73 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Terminating instance [ 2390.632639] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fa321c8f-c411-4769-8cf0-625400daaa15 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2390.632639] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fa321c8f-c411-4769-8cf0-625400daaa15 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2390.632639] env[62824]: DEBUG nova.objects.instance [None req-fa321c8f-c411-4769-8cf0-625400daaa15 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Lazy-loading 'resources' on Instance uuid 71e9733b-c84b-4501-8faf-a487f089b498 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2390.684436] env[62824]: DEBUG nova.network.neutron [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Updating instance_info_cache with network_info: [{"id": "821db814-549f-4b95-befb-cdb81b9bcb70", "address": "fa:16:3e:29:1d:69", "network": {"id": "5233a794-3514-4912-a857-7e76a66b1e8f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-910685188-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0eb41836305847fca773afebdafa01db", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd98a25d-a7a9-4fb5-8fef-e8df4dbbbf11", "external-id": "nsx-vlan-transportzone-707", "segmentation_id": 707, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap821db814-54", "ovs_interfaceid": "821db814-549f-4b95-befb-cdb81b9bcb70", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2390.821268] env[62824]: DEBUG oslo_vmware.api [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146950, 'name': ReconfigVM_Task, 'duration_secs': 0.152449} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2390.821559] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Reconfigured VM instance instance-0000006d to detach disk 2000 {{(pid=62824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 2390.822387] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57978ab9-d83c-45f4-a5df-131244098232 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2390.844389] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Reconfiguring VM instance instance-0000006d to attach disk [datastore2] 21a7eed6-c146-4911-9a0a-a1e1644e8e9d/21a7eed6-c146-4911-9a0a-a1e1644e8e9d.vmdk or device None with type thin {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2390.845198] env[62824]: INFO nova.network.neutron [None req-926ba5af-9099-485e-806f-e6c1b328a404 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Port afcd2bdb-2bfc-40d7-a8e5-2a417be0e0e9 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 2390.845522] env[62824]: DEBUG nova.network.neutron [None req-926ba5af-9099-485e-806f-e6c1b328a404 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Updating instance_info_cache with network_info: [{"id": "3c1f3341-20a4-4b77-8dae-12ef5901a16e", "address": "fa:16:3e:1c:3a:8c", "network": {"id": "aa1c09e5-8d72-43ad-a903-a0f5e711da80", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1643616572-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.166", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c987c5f844be42c99cc9d57a9ca84c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9bc2632-36f9-4912-8782-8bbb789f909d", "external-id": "nsx-vlan-transportzone-897", "segmentation_id": 897, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3c1f3341-20", "ovs_interfaceid": "3c1f3341-20a4-4b77-8dae-12ef5901a16e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2390.847167] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eec61259-7716-4119-abe9-3a58bf2af70e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2390.866680] env[62824]: DEBUG oslo_vmware.api [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2390.866680] env[62824]: value = "task-2146951" [ 2390.866680] env[62824]: _type = "Task" [ 2390.866680] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2390.874540] env[62824]: DEBUG oslo_vmware.api [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146951, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2391.016042] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-41517485-76de-4e8f-a181-a012faec848d None None] Getting list of instances from cluster (obj){ [ 2391.016042] env[62824]: value = "domain-c8" [ 2391.016042] env[62824]: _type = "ClusterComputeResource" [ 2391.016042] env[62824]: } {{(pid=62824) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 2391.017105] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1953cde4-ba09-4dfe-996d-c2e9ae2f7b55 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2391.036190] env[62824]: DEBUG nova.compute.manager [None req-1304009c-16b0-48c1-b409-6a50c077ff73 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2391.036417] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-1304009c-16b0-48c1-b409-6a50c077ff73 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2391.036733] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-41517485-76de-4e8f-a181-a012faec848d None None] Got total of 9 instances {{(pid=62824) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 2391.036882] env[62824]: WARNING nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] While synchronizing instance power states, found 11 instances in the database and 9 instances on the hypervisor. [ 2391.037034] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Triggering sync for uuid ca42fef0-1e90-4ab6-9d60-8ef7e4997884 {{(pid=62824) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10855}} [ 2391.037241] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Triggering sync for uuid 5b3df31f-15fe-473f-992c-ddb272661c53 {{(pid=62824) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10855}} [ 2391.037415] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Triggering sync for uuid c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412 {{(pid=62824) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10855}} [ 2391.037578] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Triggering sync for uuid 71e9733b-c84b-4501-8faf-a487f089b498 {{(pid=62824) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10855}} [ 2391.037741] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Triggering sync for uuid e1574266-e53b-4169-9db6-c66b84895edb {{(pid=62824) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10855}} [ 2391.037908] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Triggering sync for uuid 3d294ee9-92f6-4115-95db-9d4b66562b56 {{(pid=62824) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10855}} [ 2391.038074] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Triggering sync for uuid 1db9ff97-e59c-43f6-984e-6405d8e993a4 {{(pid=62824) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10855}} [ 2391.038238] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Triggering sync for uuid 511ef5bb-1dca-4eae-a0f1-19d1d7a17371 {{(pid=62824) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10855}} [ 2391.038395] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Triggering sync for uuid 21a7eed6-c146-4911-9a0a-a1e1644e8e9d {{(pid=62824) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10855}} [ 2391.038551] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Triggering sync for uuid c70880c9-a69a-49b9-b028-189c04276623 {{(pid=62824) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10855}} [ 2391.038698] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Triggering sync for uuid 7c47bb92-a58e-4537-a68a-289e2a5fc19a {{(pid=62824) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10855}} [ 2391.039527] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-136c7dda-2592-4d64-884f-8019e87ccc38 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2391.042863] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "ca42fef0-1e90-4ab6-9d60-8ef7e4997884" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2391.043110] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "ca42fef0-1e90-4ab6-9d60-8ef7e4997884" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2391.043399] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "5b3df31f-15fe-473f-992c-ddb272661c53" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2391.043662] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "5b3df31f-15fe-473f-992c-ddb272661c53" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2391.043925] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2391.044145] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2391.044403] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "71e9733b-c84b-4501-8faf-a487f089b498" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2391.044639] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "e1574266-e53b-4169-9db6-c66b84895edb" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2391.044820] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "e1574266-e53b-4169-9db6-c66b84895edb" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2391.045085] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "3d294ee9-92f6-4115-95db-9d4b66562b56" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2391.047185] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "3d294ee9-92f6-4115-95db-9d4b66562b56" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2391.047447] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "1db9ff97-e59c-43f6-984e-6405d8e993a4" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2391.047666] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "1db9ff97-e59c-43f6-984e-6405d8e993a4" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2391.047915] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "511ef5bb-1dca-4eae-a0f1-19d1d7a17371" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2391.048156] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "21a7eed6-c146-4911-9a0a-a1e1644e8e9d" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2391.048352] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "21a7eed6-c146-4911-9a0a-a1e1644e8e9d" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2391.048524] env[62824]: INFO nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] During sync_power_state the instance has a pending task (resize_migrating). Skip. [ 2391.048700] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "21a7eed6-c146-4911-9a0a-a1e1644e8e9d" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2391.048907] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "c70880c9-a69a-49b9-b028-189c04276623" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2391.049601] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "c70880c9-a69a-49b9-b028-189c04276623" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2391.049601] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "7c47bb92-a58e-4537-a68a-289e2a5fc19a" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2391.050186] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e86ecccc-0f98-4abf-87d7-60a78e413ac0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2391.053200] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70bc0b9e-b179-40c7-9247-be92da4754b8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2391.056137] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ed2995f-9ff6-4719-8d8f-0e20d71e0628 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2391.058766] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3062b2af-765e-48ac-8658-c64a72067db5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2391.061512] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20c9b020-c229-4c13-b708-06272ddec5ff {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2391.064603] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32900dd2-df98-4830-bb54-db826496503f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2391.068033] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acda0402-046b-4e44-8e1e-a970499b3157 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2391.075649] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-1304009c-16b0-48c1-b409-6a50c077ff73 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2391.084343] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ba396fce-a4bf-4115-a74a-e56beaa6851f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2391.101939] env[62824]: DEBUG oslo_vmware.api [None req-1304009c-16b0-48c1-b409-6a50c077ff73 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Waiting for the task: (returnval){ [ 2391.101939] env[62824]: value = "task-2146952" [ 2391.101939] env[62824]: _type = "Task" [ 2391.101939] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2391.110656] env[62824]: DEBUG oslo_vmware.api [None req-1304009c-16b0-48c1-b409-6a50c077ff73 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146952, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2391.177028] env[62824]: DEBUG nova.compute.manager [req-ee407100-a4c0-4599-afa3-071273229217 req-a9c66509-f926-46fb-bf73-bb7d5b42ccd8 service nova] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Received event network-vif-plugged-821db814-549f-4b95-befb-cdb81b9bcb70 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2391.177028] env[62824]: DEBUG oslo_concurrency.lockutils [req-ee407100-a4c0-4599-afa3-071273229217 req-a9c66509-f926-46fb-bf73-bb7d5b42ccd8 service nova] Acquiring lock "7c47bb92-a58e-4537-a68a-289e2a5fc19a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2391.177202] env[62824]: DEBUG oslo_concurrency.lockutils [req-ee407100-a4c0-4599-afa3-071273229217 req-a9c66509-f926-46fb-bf73-bb7d5b42ccd8 service nova] Lock "7c47bb92-a58e-4537-a68a-289e2a5fc19a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2391.177360] env[62824]: DEBUG oslo_concurrency.lockutils [req-ee407100-a4c0-4599-afa3-071273229217 req-a9c66509-f926-46fb-bf73-bb7d5b42ccd8 service nova] Lock "7c47bb92-a58e-4537-a68a-289e2a5fc19a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2391.177526] env[62824]: DEBUG nova.compute.manager [req-ee407100-a4c0-4599-afa3-071273229217 req-a9c66509-f926-46fb-bf73-bb7d5b42ccd8 service nova] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] No waiting events found dispatching network-vif-plugged-821db814-549f-4b95-befb-cdb81b9bcb70 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2391.177716] env[62824]: WARNING nova.compute.manager [req-ee407100-a4c0-4599-afa3-071273229217 req-a9c66509-f926-46fb-bf73-bb7d5b42ccd8 service nova] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Received unexpected event network-vif-plugged-821db814-549f-4b95-befb-cdb81b9bcb70 for instance with vm_state building and task_state spawning. [ 2391.177885] env[62824]: DEBUG nova.compute.manager [req-ee407100-a4c0-4599-afa3-071273229217 req-a9c66509-f926-46fb-bf73-bb7d5b42ccd8 service nova] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Received event network-changed-821db814-549f-4b95-befb-cdb81b9bcb70 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2391.178052] env[62824]: DEBUG nova.compute.manager [req-ee407100-a4c0-4599-afa3-071273229217 req-a9c66509-f926-46fb-bf73-bb7d5b42ccd8 service nova] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Refreshing instance network info cache due to event network-changed-821db814-549f-4b95-befb-cdb81b9bcb70. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2391.178223] env[62824]: DEBUG oslo_concurrency.lockutils [req-ee407100-a4c0-4599-afa3-071273229217 req-a9c66509-f926-46fb-bf73-bb7d5b42ccd8 service nova] Acquiring lock "refresh_cache-7c47bb92-a58e-4537-a68a-289e2a5fc19a" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2391.186339] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Releasing lock "refresh_cache-7c47bb92-a58e-4537-a68a-289e2a5fc19a" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2391.186471] env[62824]: DEBUG nova.compute.manager [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Instance network_info: |[{"id": "821db814-549f-4b95-befb-cdb81b9bcb70", "address": "fa:16:3e:29:1d:69", "network": {"id": "5233a794-3514-4912-a857-7e76a66b1e8f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-910685188-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0eb41836305847fca773afebdafa01db", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd98a25d-a7a9-4fb5-8fef-e8df4dbbbf11", "external-id": "nsx-vlan-transportzone-707", "segmentation_id": 707, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap821db814-54", "ovs_interfaceid": "821db814-549f-4b95-befb-cdb81b9bcb70", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2391.187352] env[62824]: DEBUG oslo_concurrency.lockutils [req-ee407100-a4c0-4599-afa3-071273229217 req-a9c66509-f926-46fb-bf73-bb7d5b42ccd8 service nova] Acquired lock "refresh_cache-7c47bb92-a58e-4537-a68a-289e2a5fc19a" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2391.187352] env[62824]: DEBUG nova.network.neutron [req-ee407100-a4c0-4599-afa3-071273229217 req-a9c66509-f926-46fb-bf73-bb7d5b42ccd8 service nova] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Refreshing network info cache for port 821db814-549f-4b95-befb-cdb81b9bcb70 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2391.188038] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:29:1d:69', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dd98a25d-a7a9-4fb5-8fef-e8df4dbbbf11', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '821db814-549f-4b95-befb-cdb81b9bcb70', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2391.195563] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Creating folder: Project (0eb41836305847fca773afebdafa01db). Parent ref: group-v438503. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 2391.200650] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8c96be03-ad2e-499a-8a0d-81faadbabffb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2391.216391] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Created folder: Project (0eb41836305847fca773afebdafa01db) in parent group-v438503. [ 2391.216596] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Creating folder: Instances. Parent ref: group-v438770. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 2391.217209] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0df29ce4-217d-4145-a50d-c7cd3f567b72 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2391.227926] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Created folder: Instances in parent group-v438770. [ 2391.228204] env[62824]: DEBUG oslo.service.loopingcall [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2391.231503] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2391.231503] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-31a128fe-5ba2-4414-bb2b-b2e405a361f7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2391.260058] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2391.260058] env[62824]: value = "task-2146955" [ 2391.260058] env[62824]: _type = "Task" [ 2391.260058] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2391.270816] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146955, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2391.320837] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bee558c3-eaa6-4d3e-b7c7-e6579ae2faae {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2391.329507] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca3e84de-404c-4ddf-9e51-c5a5e13db95d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2391.360472] env[62824]: DEBUG oslo_concurrency.lockutils [None req-926ba5af-9099-485e-806f-e6c1b328a404 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Releasing lock "refresh_cache-511ef5bb-1dca-4eae-a0f1-19d1d7a17371" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2391.365972] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecb51824-849b-4c98-bc27-f538ea74afed {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2391.379826] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55865646-aabc-4ec1-bda5-894c8a3860a1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2391.383759] env[62824]: DEBUG oslo_vmware.api [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146951, 'name': ReconfigVM_Task, 'duration_secs': 0.239421} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2391.384249] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Reconfigured VM instance instance-0000006d to attach disk [datastore2] 21a7eed6-c146-4911-9a0a-a1e1644e8e9d/21a7eed6-c146-4911-9a0a-a1e1644e8e9d.vmdk or device None with type thin {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2391.384527] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Updating instance '21a7eed6-c146-4911-9a0a-a1e1644e8e9d' progress to 50 {{(pid=62824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 2391.397223] env[62824]: DEBUG nova.compute.provider_tree [None req-fa321c8f-c411-4769-8cf0-625400daaa15 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2391.434894] env[62824]: DEBUG nova.network.neutron [req-ee407100-a4c0-4599-afa3-071273229217 req-a9c66509-f926-46fb-bf73-bb7d5b42ccd8 service nova] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Updated VIF entry in instance network info cache for port 821db814-549f-4b95-befb-cdb81b9bcb70. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2391.435283] env[62824]: DEBUG nova.network.neutron [req-ee407100-a4c0-4599-afa3-071273229217 req-a9c66509-f926-46fb-bf73-bb7d5b42ccd8 service nova] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Updating instance_info_cache with network_info: [{"id": "821db814-549f-4b95-befb-cdb81b9bcb70", "address": "fa:16:3e:29:1d:69", "network": {"id": "5233a794-3514-4912-a857-7e76a66b1e8f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-910685188-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0eb41836305847fca773afebdafa01db", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd98a25d-a7a9-4fb5-8fef-e8df4dbbbf11", "external-id": "nsx-vlan-transportzone-707", "segmentation_id": 707, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap821db814-54", "ovs_interfaceid": "821db814-549f-4b95-befb-cdb81b9bcb70", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2391.588756] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "1db9ff97-e59c-43f6-984e-6405d8e993a4" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.541s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2391.590332] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "5b3df31f-15fe-473f-992c-ddb272661c53" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.547s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2391.594801] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "3d294ee9-92f6-4115-95db-9d4b66562b56" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.549s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2391.596550] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "ca42fef0-1e90-4ab6-9d60-8ef7e4997884" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.553s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2391.596854] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.553s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2391.597158] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "e1574266-e53b-4169-9db6-c66b84895edb" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.552s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2391.604046] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "c70880c9-a69a-49b9-b028-189c04276623" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.554s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2391.611785] env[62824]: DEBUG oslo_vmware.api [None req-1304009c-16b0-48c1-b409-6a50c077ff73 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146952, 'name': PowerOffVM_Task, 'duration_secs': 0.170123} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2391.612075] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-1304009c-16b0-48c1-b409-6a50c077ff73 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2391.612296] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-1304009c-16b0-48c1-b409-6a50c077ff73 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2391.612890] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-02a0d280-cc08-4927-8559-d2c5a789fc33 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2391.729627] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-1304009c-16b0-48c1-b409-6a50c077ff73 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2391.730028] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-1304009c-16b0-48c1-b409-6a50c077ff73 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Deleting contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2391.730128] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-1304009c-16b0-48c1-b409-6a50c077ff73 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Deleting the datastore file [datastore1] 511ef5bb-1dca-4eae-a0f1-19d1d7a17371 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2391.730358] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f2bfe649-329e-4d08-ab3a-b1565587298e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2391.737837] env[62824]: DEBUG oslo_vmware.api [None req-1304009c-16b0-48c1-b409-6a50c077ff73 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Waiting for the task: (returnval){ [ 2391.737837] env[62824]: value = "task-2146957" [ 2391.737837] env[62824]: _type = "Task" [ 2391.737837] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2391.745489] env[62824]: DEBUG oslo_vmware.api [None req-1304009c-16b0-48c1-b409-6a50c077ff73 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146957, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2391.769618] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146955, 'name': CreateVM_Task, 'duration_secs': 0.427792} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2391.769785] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2391.770494] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2391.770666] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2391.771017] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2391.771305] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-99b06224-fb0c-47af-8632-acbb9cc80a92 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2391.775555] env[62824]: DEBUG oslo_vmware.api [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2391.775555] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5222b6c4-9de3-daae-b928-10c9dbaef1cf" [ 2391.775555] env[62824]: _type = "Task" [ 2391.775555] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2391.783367] env[62824]: DEBUG oslo_vmware.api [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5222b6c4-9de3-daae-b928-10c9dbaef1cf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2391.866861] env[62824]: DEBUG oslo_concurrency.lockutils [None req-926ba5af-9099-485e-806f-e6c1b328a404 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lock "interface-511ef5bb-1dca-4eae-a0f1-19d1d7a17371-afcd2bdb-2bfc-40d7-a8e5-2a417be0e0e9" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 12.875s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2391.890613] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc3be408-27d3-4244-95a9-6d3e5e9f84fb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2391.911218] env[62824]: DEBUG nova.scheduler.client.report [None req-fa321c8f-c411-4769-8cf0-625400daaa15 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2391.914845] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acd31a46-8bba-4390-8ec2-f2eece22323f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2391.933393] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Updating instance '21a7eed6-c146-4911-9a0a-a1e1644e8e9d' progress to 67 {{(pid=62824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 2391.938181] env[62824]: DEBUG oslo_concurrency.lockutils [req-ee407100-a4c0-4599-afa3-071273229217 req-a9c66509-f926-46fb-bf73-bb7d5b42ccd8 service nova] Releasing lock "refresh_cache-7c47bb92-a58e-4537-a68a-289e2a5fc19a" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2392.248340] env[62824]: DEBUG oslo_vmware.api [None req-1304009c-16b0-48c1-b409-6a50c077ff73 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146957, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.165213} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2392.248595] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-1304009c-16b0-48c1-b409-6a50c077ff73 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2392.248776] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-1304009c-16b0-48c1-b409-6a50c077ff73 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Deleted contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2392.249753] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-1304009c-16b0-48c1-b409-6a50c077ff73 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2392.249753] env[62824]: INFO nova.compute.manager [None req-1304009c-16b0-48c1-b409-6a50c077ff73 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Took 1.21 seconds to destroy the instance on the hypervisor. [ 2392.249753] env[62824]: DEBUG oslo.service.loopingcall [None req-1304009c-16b0-48c1-b409-6a50c077ff73 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2392.249753] env[62824]: DEBUG nova.compute.manager [-] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2392.249753] env[62824]: DEBUG nova.network.neutron [-] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2392.286350] env[62824]: DEBUG oslo_vmware.api [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5222b6c4-9de3-daae-b928-10c9dbaef1cf, 'name': SearchDatastore_Task, 'duration_secs': 0.008721} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2392.289352] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2392.289651] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2392.289892] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2392.290058] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2392.290241] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2392.290545] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-17c7f218-5c9b-427f-8841-08d43e352b32 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2392.299589] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2392.299702] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2392.300409] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-537e3221-9b41-4444-8d4d-02710effb9bc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2392.306041] env[62824]: DEBUG oslo_vmware.api [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2392.306041] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52c42786-a543-49c9-8224-144ce5ded0c6" [ 2392.306041] env[62824]: _type = "Task" [ 2392.306041] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2392.314509] env[62824]: DEBUG oslo_vmware.api [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52c42786-a543-49c9-8224-144ce5ded0c6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2392.419329] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fa321c8f-c411-4769-8cf0-625400daaa15 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.787s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2392.437986] env[62824]: INFO nova.scheduler.client.report [None req-fa321c8f-c411-4769-8cf0-625400daaa15 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Deleted allocations for instance 71e9733b-c84b-4501-8faf-a487f089b498 [ 2392.502204] env[62824]: DEBUG nova.network.neutron [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Port edaf5d3c-004d-4ba0-b037-e7eef2acbe44 binding to destination host cpu-1 is already ACTIVE {{(pid=62824) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 2392.816522] env[62824]: DEBUG oslo_vmware.api [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52c42786-a543-49c9-8224-144ce5ded0c6, 'name': SearchDatastore_Task, 'duration_secs': 0.009015} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2392.817340] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5b40c36a-7bf8-4517-b91a-65d8d2249ad6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2392.822788] env[62824]: DEBUG oslo_vmware.api [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2392.822788] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52129512-31d3-e057-d5b2-fb06a0701483" [ 2392.822788] env[62824]: _type = "Task" [ 2392.822788] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2392.830691] env[62824]: DEBUG oslo_vmware.api [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52129512-31d3-e057-d5b2-fb06a0701483, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2392.952023] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fa321c8f-c411-4769-8cf0-625400daaa15 tempest-ServersNegativeTestJSON-1727296179 tempest-ServersNegativeTestJSON-1727296179-project-member] Lock "71e9733b-c84b-4501-8faf-a487f089b498" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.947s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2392.952023] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "71e9733b-c84b-4501-8faf-a487f089b498" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 1.906s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2392.952023] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9be09d06-d245-4d7b-a262-3c931ff2d220 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2392.967028] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5508bf01-7763-43ae-b854-5bf4934da50b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2393.207262] env[62824]: DEBUG nova.compute.manager [req-d0b81f8c-ca8a-482b-8a2e-8c8b616333f8 req-15f5eeb9-50bd-4653-9de6-5e63b5aa6d3d service nova] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Received event network-vif-deleted-3c1f3341-20a4-4b77-8dae-12ef5901a16e {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2393.207262] env[62824]: INFO nova.compute.manager [req-d0b81f8c-ca8a-482b-8a2e-8c8b616333f8 req-15f5eeb9-50bd-4653-9de6-5e63b5aa6d3d service nova] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Neutron deleted interface 3c1f3341-20a4-4b77-8dae-12ef5901a16e; detaching it from the instance and deleting it from the info cache [ 2393.207262] env[62824]: DEBUG nova.network.neutron [req-d0b81f8c-ca8a-482b-8a2e-8c8b616333f8 req-15f5eeb9-50bd-4653-9de6-5e63b5aa6d3d service nova] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2393.333958] env[62824]: DEBUG oslo_vmware.api [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52129512-31d3-e057-d5b2-fb06a0701483, 'name': SearchDatastore_Task, 'duration_secs': 0.009264} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2393.334847] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2393.335186] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 7c47bb92-a58e-4537-a68a-289e2a5fc19a/7c47bb92-a58e-4537-a68a-289e2a5fc19a.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2393.335497] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2d494636-4fba-429c-b97d-d13d09ebad88 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2393.342520] env[62824]: DEBUG oslo_vmware.api [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2393.342520] env[62824]: value = "task-2146958" [ 2393.342520] env[62824]: _type = "Task" [ 2393.342520] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2393.349781] env[62824]: DEBUG oslo_vmware.api [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2146958, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2393.404784] env[62824]: DEBUG nova.network.neutron [-] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2393.516105] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "71e9733b-c84b-4501-8faf-a487f089b498" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.565s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2393.535039] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquiring lock "21a7eed6-c146-4911-9a0a-a1e1644e8e9d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2393.535295] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "21a7eed6-c146-4911-9a0a-a1e1644e8e9d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2393.535471] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "21a7eed6-c146-4911-9a0a-a1e1644e8e9d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2393.713024] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a25d7ccf-1534-4200-8ec0-b3c2a64d8329 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2393.724422] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-176bda50-6c3b-40b0-bcf9-95b07539f439 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2393.762454] env[62824]: DEBUG nova.compute.manager [req-d0b81f8c-ca8a-482b-8a2e-8c8b616333f8 req-15f5eeb9-50bd-4653-9de6-5e63b5aa6d3d service nova] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Detach interface failed, port_id=3c1f3341-20a4-4b77-8dae-12ef5901a16e, reason: Instance 511ef5bb-1dca-4eae-a0f1-19d1d7a17371 could not be found. {{(pid=62824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11482}} [ 2393.852480] env[62824]: DEBUG oslo_vmware.api [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2146958, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.457133} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2393.853713] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 7c47bb92-a58e-4537-a68a-289e2a5fc19a/7c47bb92-a58e-4537-a68a-289e2a5fc19a.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2393.853971] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2393.854233] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-47cdf7c2-2a5c-4569-9e22-367daa2598e8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2393.860905] env[62824]: DEBUG oslo_vmware.api [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2393.860905] env[62824]: value = "task-2146959" [ 2393.860905] env[62824]: _type = "Task" [ 2393.860905] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2393.868370] env[62824]: DEBUG oslo_vmware.api [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2146959, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2393.908050] env[62824]: INFO nova.compute.manager [-] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Took 1.66 seconds to deallocate network for instance. [ 2394.371796] env[62824]: DEBUG oslo_vmware.api [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2146959, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060477} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2394.372283] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2394.373878] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc29c8cc-09a4-4737-bcd1-095da82086bf {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2394.396781] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Reconfiguring VM instance instance-0000006f to attach disk [datastore2] 7c47bb92-a58e-4537-a68a-289e2a5fc19a/7c47bb92-a58e-4537-a68a-289e2a5fc19a.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2394.397120] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-aa20989c-5c7e-4ab4-924b-723df897f7b8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2394.414584] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1304009c-16b0-48c1-b409-6a50c077ff73 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2394.414904] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1304009c-16b0-48c1-b409-6a50c077ff73 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2394.415195] env[62824]: DEBUG nova.objects.instance [None req-1304009c-16b0-48c1-b409-6a50c077ff73 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lazy-loading 'resources' on Instance uuid 511ef5bb-1dca-4eae-a0f1-19d1d7a17371 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2394.422051] env[62824]: DEBUG oslo_vmware.api [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2394.422051] env[62824]: value = "task-2146960" [ 2394.422051] env[62824]: _type = "Task" [ 2394.422051] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2394.431679] env[62824]: DEBUG oslo_vmware.api [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2146960, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2394.572651] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquiring lock "refresh_cache-21a7eed6-c146-4911-9a0a-a1e1644e8e9d" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2394.572839] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquired lock "refresh_cache-21a7eed6-c146-4911-9a0a-a1e1644e8e9d" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2394.573022] env[62824]: DEBUG nova.network.neutron [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2394.933422] env[62824]: DEBUG oslo_vmware.api [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2146960, 'name': ReconfigVM_Task, 'duration_secs': 0.262271} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2394.933798] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Reconfigured VM instance instance-0000006f to attach disk [datastore2] 7c47bb92-a58e-4537-a68a-289e2a5fc19a/7c47bb92-a58e-4537-a68a-289e2a5fc19a.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2394.934334] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fb1191a4-0977-41ff-8234-c074d9f1b3ad {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2394.942453] env[62824]: DEBUG oslo_vmware.api [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2394.942453] env[62824]: value = "task-2146961" [ 2394.942453] env[62824]: _type = "Task" [ 2394.942453] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2394.951152] env[62824]: DEBUG oslo_vmware.api [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2146961, 'name': Rename_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2395.058044] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59366086-2367-4dea-bb7d-1d94c3610598 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2395.065721] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db91b596-88b5-4bda-924d-183e200706ca {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2395.099608] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0abbce2-ba94-40f7-9270-91b1acde418a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2395.107229] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d75c809d-c58f-463c-bb91-b7a210eff548 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2395.120769] env[62824]: DEBUG nova.compute.provider_tree [None req-1304009c-16b0-48c1-b409-6a50c077ff73 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2395.297526] env[62824]: DEBUG nova.network.neutron [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Updating instance_info_cache with network_info: [{"id": "edaf5d3c-004d-4ba0-b037-e7eef2acbe44", "address": "fa:16:3e:06:2e:6a", "network": {"id": "12411632-b7f6-4544-b2e4-b04ae7337547", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-917106830-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4479921caf5f405b8fc49baad390a0e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "de7fa486-5f28-44ae-b0cf-72234ff87546", "external-id": "nsx-vlan-transportzone-229", "segmentation_id": 229, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapedaf5d3c-00", "ovs_interfaceid": "edaf5d3c-004d-4ba0-b037-e7eef2acbe44", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2395.452831] env[62824]: DEBUG oslo_vmware.api [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2146961, 'name': Rename_Task, 'duration_secs': 0.12988} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2395.452964] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2395.453117] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7fa9330f-4f67-4c82-949d-ec7c3c310635 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2395.460882] env[62824]: DEBUG oslo_vmware.api [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2395.460882] env[62824]: value = "task-2146962" [ 2395.460882] env[62824]: _type = "Task" [ 2395.460882] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2395.467958] env[62824]: DEBUG oslo_vmware.api [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2146962, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2395.625323] env[62824]: DEBUG nova.scheduler.client.report [None req-1304009c-16b0-48c1-b409-6a50c077ff73 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2395.800550] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Releasing lock "refresh_cache-21a7eed6-c146-4911-9a0a-a1e1644e8e9d" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2395.971709] env[62824]: DEBUG oslo_vmware.api [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2146962, 'name': PowerOnVM_Task, 'duration_secs': 0.507185} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2395.972086] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2395.972204] env[62824]: INFO nova.compute.manager [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Took 7.30 seconds to spawn the instance on the hypervisor. [ 2395.972484] env[62824]: DEBUG nova.compute.manager [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2395.973307] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb5e501e-f0cd-4d7a-8cbb-5a604514fbbc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2396.131160] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1304009c-16b0-48c1-b409-6a50c077ff73 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.716s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2396.157501] env[62824]: INFO nova.scheduler.client.report [None req-1304009c-16b0-48c1-b409-6a50c077ff73 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Deleted allocations for instance 511ef5bb-1dca-4eae-a0f1-19d1d7a17371 [ 2396.323451] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ef0a966-c9a3-4c76-9caa-dc71bc978d40 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2396.343399] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab2c0690-674c-4c80-a1c0-e2fec1ddc625 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2396.351776] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Updating instance '21a7eed6-c146-4911-9a0a-a1e1644e8e9d' progress to 83 {{(pid=62824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 2396.491437] env[62824]: INFO nova.compute.manager [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Took 15.87 seconds to build instance. [ 2396.665220] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1304009c-16b0-48c1-b409-6a50c077ff73 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lock "511ef5bb-1dca-4eae-a0f1-19d1d7a17371" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.137s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2396.666124] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "511ef5bb-1dca-4eae-a0f1-19d1d7a17371" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 5.618s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2396.666321] env[62824]: INFO nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] During sync_power_state the instance has a pending task (deleting). Skip. [ 2396.666496] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "511ef5bb-1dca-4eae-a0f1-19d1d7a17371" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2396.857420] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2396.857724] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7acf2431-0457-4ef3-8a0e-27579dc985d1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2396.865572] env[62824]: DEBUG oslo_vmware.api [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2396.865572] env[62824]: value = "task-2146963" [ 2396.865572] env[62824]: _type = "Task" [ 2396.865572] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2396.875941] env[62824]: DEBUG oslo_vmware.api [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146963, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2396.993792] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f99726f4-8829-4de9-bedb-3fb10488ddca tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lock "7c47bb92-a58e-4537-a68a-289e2a5fc19a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.404s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2396.994388] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "7c47bb92-a58e-4537-a68a-289e2a5fc19a" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 5.945s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2396.994388] env[62824]: INFO nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] During sync_power_state the instance has a pending task (spawning). Skip. [ 2396.994478] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "7c47bb92-a58e-4537-a68a-289e2a5fc19a" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2397.153347] env[62824]: DEBUG nova.compute.manager [req-e80dd19f-0952-4b1b-a45f-3f721e693c1f req-ef35a6d1-73ed-412e-8bfb-9b8ccba1a6ed service nova] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Received event network-changed-821db814-549f-4b95-befb-cdb81b9bcb70 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2397.154200] env[62824]: DEBUG nova.compute.manager [req-e80dd19f-0952-4b1b-a45f-3f721e693c1f req-ef35a6d1-73ed-412e-8bfb-9b8ccba1a6ed service nova] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Refreshing instance network info cache due to event network-changed-821db814-549f-4b95-befb-cdb81b9bcb70. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2397.154200] env[62824]: DEBUG oslo_concurrency.lockutils [req-e80dd19f-0952-4b1b-a45f-3f721e693c1f req-ef35a6d1-73ed-412e-8bfb-9b8ccba1a6ed service nova] Acquiring lock "refresh_cache-7c47bb92-a58e-4537-a68a-289e2a5fc19a" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2397.154200] env[62824]: DEBUG oslo_concurrency.lockutils [req-e80dd19f-0952-4b1b-a45f-3f721e693c1f req-ef35a6d1-73ed-412e-8bfb-9b8ccba1a6ed service nova] Acquired lock "refresh_cache-7c47bb92-a58e-4537-a68a-289e2a5fc19a" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2397.154200] env[62824]: DEBUG nova.network.neutron [req-e80dd19f-0952-4b1b-a45f-3f721e693c1f req-ef35a6d1-73ed-412e-8bfb-9b8ccba1a6ed service nova] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Refreshing network info cache for port 821db814-549f-4b95-befb-cdb81b9bcb70 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2397.173786] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5b6bd879-ac77-4165-abf9-0d78d4ef9fb7 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquiring lock "e1574266-e53b-4169-9db6-c66b84895edb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2397.174470] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5b6bd879-ac77-4165-abf9-0d78d4ef9fb7 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lock "e1574266-e53b-4169-9db6-c66b84895edb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2397.174691] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5b6bd879-ac77-4165-abf9-0d78d4ef9fb7 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquiring lock "e1574266-e53b-4169-9db6-c66b84895edb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2397.174877] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5b6bd879-ac77-4165-abf9-0d78d4ef9fb7 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lock "e1574266-e53b-4169-9db6-c66b84895edb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2397.175057] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5b6bd879-ac77-4165-abf9-0d78d4ef9fb7 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lock "e1574266-e53b-4169-9db6-c66b84895edb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2397.177031] env[62824]: INFO nova.compute.manager [None req-5b6bd879-ac77-4165-abf9-0d78d4ef9fb7 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Terminating instance [ 2397.378352] env[62824]: DEBUG oslo_vmware.api [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146963, 'name': PowerOnVM_Task, 'duration_secs': 0.396794} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2397.378681] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2397.378973] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-9b36fb8a-cb71-434b-b75b-4de7baefa09e tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Updating instance '21a7eed6-c146-4911-9a0a-a1e1644e8e9d' progress to 100 {{(pid=62824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 2397.681300] env[62824]: DEBUG nova.compute.manager [None req-5b6bd879-ac77-4165-abf9-0d78d4ef9fb7 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2397.681680] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5b6bd879-ac77-4165-abf9-0d78d4ef9fb7 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2397.683496] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c5d2143-898d-4963-a635-5596597880dc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2397.697159] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b6bd879-ac77-4165-abf9-0d78d4ef9fb7 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2397.698022] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0d705073-08b4-4401-bf83-0f39face8284 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2397.709172] env[62824]: DEBUG oslo_vmware.api [None req-5b6bd879-ac77-4165-abf9-0d78d4ef9fb7 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Waiting for the task: (returnval){ [ 2397.709172] env[62824]: value = "task-2146964" [ 2397.709172] env[62824]: _type = "Task" [ 2397.709172] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2397.721380] env[62824]: DEBUG oslo_vmware.api [None req-5b6bd879-ac77-4165-abf9-0d78d4ef9fb7 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146964, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2397.917985] env[62824]: DEBUG nova.network.neutron [req-e80dd19f-0952-4b1b-a45f-3f721e693c1f req-ef35a6d1-73ed-412e-8bfb-9b8ccba1a6ed service nova] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Updated VIF entry in instance network info cache for port 821db814-549f-4b95-befb-cdb81b9bcb70. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2397.918649] env[62824]: DEBUG nova.network.neutron [req-e80dd19f-0952-4b1b-a45f-3f721e693c1f req-ef35a6d1-73ed-412e-8bfb-9b8ccba1a6ed service nova] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Updating instance_info_cache with network_info: [{"id": "821db814-549f-4b95-befb-cdb81b9bcb70", "address": "fa:16:3e:29:1d:69", "network": {"id": "5233a794-3514-4912-a857-7e76a66b1e8f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-910685188-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.129", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0eb41836305847fca773afebdafa01db", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd98a25d-a7a9-4fb5-8fef-e8df4dbbbf11", "external-id": "nsx-vlan-transportzone-707", "segmentation_id": 707, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap821db814-54", "ovs_interfaceid": "821db814-549f-4b95-befb-cdb81b9bcb70", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2398.220771] env[62824]: DEBUG oslo_vmware.api [None req-5b6bd879-ac77-4165-abf9-0d78d4ef9fb7 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146964, 'name': PowerOffVM_Task, 'duration_secs': 0.210349} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2398.221211] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b6bd879-ac77-4165-abf9-0d78d4ef9fb7 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2398.221280] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5b6bd879-ac77-4165-abf9-0d78d4ef9fb7 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2398.221575] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4cdc95fa-5bb1-4e29-a707-2d8c911d565d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2398.421413] env[62824]: DEBUG oslo_concurrency.lockutils [req-e80dd19f-0952-4b1b-a45f-3f721e693c1f req-ef35a6d1-73ed-412e-8bfb-9b8ccba1a6ed service nova] Releasing lock "refresh_cache-7c47bb92-a58e-4537-a68a-289e2a5fc19a" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2398.655726] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5b6bd879-ac77-4165-abf9-0d78d4ef9fb7 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2398.655886] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5b6bd879-ac77-4165-abf9-0d78d4ef9fb7 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Deleting contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2398.656090] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b6bd879-ac77-4165-abf9-0d78d4ef9fb7 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Deleting the datastore file [datastore2] e1574266-e53b-4169-9db6-c66b84895edb {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2398.656687] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6381b8d0-4bb6-4c53-bb78-d2256d5aac97 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2398.664492] env[62824]: DEBUG oslo_vmware.api [None req-5b6bd879-ac77-4165-abf9-0d78d4ef9fb7 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Waiting for the task: (returnval){ [ 2398.664492] env[62824]: value = "task-2146966" [ 2398.664492] env[62824]: _type = "Task" [ 2398.664492] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2398.673225] env[62824]: DEBUG oslo_vmware.api [None req-5b6bd879-ac77-4165-abf9-0d78d4ef9fb7 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146966, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2399.126017] env[62824]: DEBUG oslo_concurrency.lockutils [None req-301c6e6e-abfa-4a4a-aabc-d5afb9e12fb6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquiring lock "21a7eed6-c146-4911-9a0a-a1e1644e8e9d" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2399.126292] env[62824]: DEBUG oslo_concurrency.lockutils [None req-301c6e6e-abfa-4a4a-aabc-d5afb9e12fb6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "21a7eed6-c146-4911-9a0a-a1e1644e8e9d" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2399.126487] env[62824]: DEBUG nova.compute.manager [None req-301c6e6e-abfa-4a4a-aabc-d5afb9e12fb6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Going to confirm migration 3 {{(pid=62824) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5250}} [ 2399.174383] env[62824]: DEBUG oslo_vmware.api [None req-5b6bd879-ac77-4165-abf9-0d78d4ef9fb7 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Task: {'id': task-2146966, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.193357} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2399.174653] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b6bd879-ac77-4165-abf9-0d78d4ef9fb7 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2399.174834] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5b6bd879-ac77-4165-abf9-0d78d4ef9fb7 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Deleted contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2399.175034] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5b6bd879-ac77-4165-abf9-0d78d4ef9fb7 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2399.175214] env[62824]: INFO nova.compute.manager [None req-5b6bd879-ac77-4165-abf9-0d78d4ef9fb7 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Took 1.49 seconds to destroy the instance on the hypervisor. [ 2399.175458] env[62824]: DEBUG oslo.service.loopingcall [None req-5b6bd879-ac77-4165-abf9-0d78d4ef9fb7 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2399.175642] env[62824]: DEBUG nova.compute.manager [-] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2399.175738] env[62824]: DEBUG nova.network.neutron [-] [instance: e1574266-e53b-4169-9db6-c66b84895edb] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2399.456624] env[62824]: DEBUG nova.compute.manager [req-4ed3cd9a-fcd2-4494-a373-10d7eeb74337 req-4fb60bc5-5e89-4234-9998-507b56f0a65d service nova] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Received event network-vif-deleted-d5c6ba87-e0c7-49c2-bbaf-b7c64f84cdf8 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2399.456870] env[62824]: INFO nova.compute.manager [req-4ed3cd9a-fcd2-4494-a373-10d7eeb74337 req-4fb60bc5-5e89-4234-9998-507b56f0a65d service nova] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Neutron deleted interface d5c6ba87-e0c7-49c2-bbaf-b7c64f84cdf8; detaching it from the instance and deleting it from the info cache [ 2399.457235] env[62824]: DEBUG nova.network.neutron [req-4ed3cd9a-fcd2-4494-a373-10d7eeb74337 req-4fb60bc5-5e89-4234-9998-507b56f0a65d service nova] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2399.691984] env[62824]: DEBUG oslo_concurrency.lockutils [None req-301c6e6e-abfa-4a4a-aabc-d5afb9e12fb6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquiring lock "refresh_cache-21a7eed6-c146-4911-9a0a-a1e1644e8e9d" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2399.691984] env[62824]: DEBUG oslo_concurrency.lockutils [None req-301c6e6e-abfa-4a4a-aabc-d5afb9e12fb6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquired lock "refresh_cache-21a7eed6-c146-4911-9a0a-a1e1644e8e9d" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2399.692134] env[62824]: DEBUG nova.network.neutron [None req-301c6e6e-abfa-4a4a-aabc-d5afb9e12fb6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2399.692265] env[62824]: DEBUG nova.objects.instance [None req-301c6e6e-abfa-4a4a-aabc-d5afb9e12fb6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lazy-loading 'info_cache' on Instance uuid 21a7eed6-c146-4911-9a0a-a1e1644e8e9d {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2399.935512] env[62824]: DEBUG nova.network.neutron [-] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2399.961165] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-04504afe-835f-40e5-aafe-45d9956b3519 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2399.971971] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab74d42f-ab5c-4b67-96d3-1ac824ee0c77 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2400.000358] env[62824]: DEBUG nova.compute.manager [req-4ed3cd9a-fcd2-4494-a373-10d7eeb74337 req-4fb60bc5-5e89-4234-9998-507b56f0a65d service nova] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Detach interface failed, port_id=d5c6ba87-e0c7-49c2-bbaf-b7c64f84cdf8, reason: Instance e1574266-e53b-4169-9db6-c66b84895edb could not be found. {{(pid=62824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11482}} [ 2400.438194] env[62824]: INFO nova.compute.manager [-] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Took 1.26 seconds to deallocate network for instance. [ 2400.895224] env[62824]: DEBUG nova.network.neutron [None req-301c6e6e-abfa-4a4a-aabc-d5afb9e12fb6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Updating instance_info_cache with network_info: [{"id": "edaf5d3c-004d-4ba0-b037-e7eef2acbe44", "address": "fa:16:3e:06:2e:6a", "network": {"id": "12411632-b7f6-4544-b2e4-b04ae7337547", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-917106830-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4479921caf5f405b8fc49baad390a0e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "de7fa486-5f28-44ae-b0cf-72234ff87546", "external-id": "nsx-vlan-transportzone-229", "segmentation_id": 229, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapedaf5d3c-00", "ovs_interfaceid": "edaf5d3c-004d-4ba0-b037-e7eef2acbe44", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2400.945553] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5b6bd879-ac77-4165-abf9-0d78d4ef9fb7 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2400.945813] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5b6bd879-ac77-4165-abf9-0d78d4ef9fb7 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2400.946041] env[62824]: DEBUG nova.objects.instance [None req-5b6bd879-ac77-4165-abf9-0d78d4ef9fb7 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lazy-loading 'resources' on Instance uuid e1574266-e53b-4169-9db6-c66b84895edb {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2401.398396] env[62824]: DEBUG oslo_concurrency.lockutils [None req-301c6e6e-abfa-4a4a-aabc-d5afb9e12fb6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Releasing lock "refresh_cache-21a7eed6-c146-4911-9a0a-a1e1644e8e9d" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2401.398585] env[62824]: DEBUG nova.objects.instance [None req-301c6e6e-abfa-4a4a-aabc-d5afb9e12fb6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lazy-loading 'migration_context' on Instance uuid 21a7eed6-c146-4911-9a0a-a1e1644e8e9d {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2401.574631] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecdf7c74-d3a3-477f-b12e-4c510c75a6b7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2401.582374] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3003d9f-3487-4d46-bb52-5ef557b38dd5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2401.613877] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87608098-2d70-4c7a-8867-b78a0f44f957 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2401.620962] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-069c8393-babf-4911-84b1-77f0df8e279f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2401.633950] env[62824]: DEBUG nova.compute.provider_tree [None req-5b6bd879-ac77-4165-abf9-0d78d4ef9fb7 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2401.901361] env[62824]: DEBUG nova.objects.base [None req-301c6e6e-abfa-4a4a-aabc-d5afb9e12fb6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Object Instance<21a7eed6-c146-4911-9a0a-a1e1644e8e9d> lazy-loaded attributes: info_cache,migration_context {{(pid=62824) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 2401.902350] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6da46dc-d247-4372-9749-94f7e5a02b4c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2401.921301] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2f837ca8-a61b-4695-8e7b-2668adb87af7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2401.926078] env[62824]: DEBUG oslo_vmware.api [None req-301c6e6e-abfa-4a4a-aabc-d5afb9e12fb6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2401.926078] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52033b5c-4872-9911-a2c0-b8711e5a3230" [ 2401.926078] env[62824]: _type = "Task" [ 2401.926078] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2401.933157] env[62824]: DEBUG oslo_vmware.api [None req-301c6e6e-abfa-4a4a-aabc-d5afb9e12fb6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52033b5c-4872-9911-a2c0-b8711e5a3230, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2402.138017] env[62824]: DEBUG nova.scheduler.client.report [None req-5b6bd879-ac77-4165-abf9-0d78d4ef9fb7 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2402.436781] env[62824]: DEBUG oslo_vmware.api [None req-301c6e6e-abfa-4a4a-aabc-d5afb9e12fb6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52033b5c-4872-9911-a2c0-b8711e5a3230, 'name': SearchDatastore_Task, 'duration_secs': 0.007156} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2402.437128] env[62824]: DEBUG oslo_concurrency.lockutils [None req-301c6e6e-abfa-4a4a-aabc-d5afb9e12fb6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2402.642750] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5b6bd879-ac77-4165-abf9-0d78d4ef9fb7 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.697s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2402.645161] env[62824]: DEBUG oslo_concurrency.lockutils [None req-301c6e6e-abfa-4a4a-aabc-d5afb9e12fb6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.208s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2402.661689] env[62824]: INFO nova.scheduler.client.report [None req-5b6bd879-ac77-4165-abf9-0d78d4ef9fb7 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Deleted allocations for instance e1574266-e53b-4169-9db6-c66b84895edb [ 2403.169163] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5b6bd879-ac77-4165-abf9-0d78d4ef9fb7 tempest-AttachInterfacesTestJSON-174452018 tempest-AttachInterfacesTestJSON-174452018-project-member] Lock "e1574266-e53b-4169-9db6-c66b84895edb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.995s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2403.263445] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d36b19a-f474-4bbc-800c-74826bd362b3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2403.271445] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-417dc641-179c-4775-91cb-ac0b987ba153 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2403.302594] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f022ef2-393a-42a8-b2f9-8abdddb1dbf1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2403.309748] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30ca4428-a09c-4ba7-9022-baa749714b1c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2403.323413] env[62824]: DEBUG nova.compute.provider_tree [None req-301c6e6e-abfa-4a4a-aabc-d5afb9e12fb6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2403.826792] env[62824]: DEBUG nova.scheduler.client.report [None req-301c6e6e-abfa-4a4a-aabc-d5afb9e12fb6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2404.838726] env[62824]: DEBUG oslo_concurrency.lockutils [None req-301c6e6e-abfa-4a4a-aabc-d5afb9e12fb6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.193s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2405.403789] env[62824]: INFO nova.scheduler.client.report [None req-301c6e6e-abfa-4a4a-aabc-d5afb9e12fb6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Deleted allocation for migration ea8a1301-70c4-4e5c-bb7f-97e03c659c29 [ 2405.909562] env[62824]: DEBUG oslo_concurrency.lockutils [None req-301c6e6e-abfa-4a4a-aabc-d5afb9e12fb6 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "21a7eed6-c146-4911-9a0a-a1e1644e8e9d" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.783s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2406.924819] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a8698d94-6446-4cd2-b280-1af18ee5168a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquiring lock "21a7eed6-c146-4911-9a0a-a1e1644e8e9d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2406.924819] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a8698d94-6446-4cd2-b280-1af18ee5168a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "21a7eed6-c146-4911-9a0a-a1e1644e8e9d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2406.924819] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a8698d94-6446-4cd2-b280-1af18ee5168a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquiring lock "21a7eed6-c146-4911-9a0a-a1e1644e8e9d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2406.925278] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a8698d94-6446-4cd2-b280-1af18ee5168a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "21a7eed6-c146-4911-9a0a-a1e1644e8e9d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2406.925278] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a8698d94-6446-4cd2-b280-1af18ee5168a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "21a7eed6-c146-4911-9a0a-a1e1644e8e9d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2406.927438] env[62824]: INFO nova.compute.manager [None req-a8698d94-6446-4cd2-b280-1af18ee5168a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Terminating instance [ 2407.367107] env[62824]: DEBUG oslo_concurrency.lockutils [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquiring lock "df621c9c-8a68-4082-8f30-9bf3dda26fb6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2407.367323] env[62824]: DEBUG oslo_concurrency.lockutils [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "df621c9c-8a68-4082-8f30-9bf3dda26fb6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2407.431717] env[62824]: DEBUG nova.compute.manager [None req-a8698d94-6446-4cd2-b280-1af18ee5168a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2407.431933] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-a8698d94-6446-4cd2-b280-1af18ee5168a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2407.433159] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4edf9da3-b484-4613-a167-d8ea425119e5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2407.441228] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8698d94-6446-4cd2-b280-1af18ee5168a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2407.441490] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f0319ee1-855b-43dc-9d8c-f8cb42537b86 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2407.447709] env[62824]: DEBUG oslo_vmware.api [None req-a8698d94-6446-4cd2-b280-1af18ee5168a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2407.447709] env[62824]: value = "task-2146967" [ 2407.447709] env[62824]: _type = "Task" [ 2407.447709] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2407.455737] env[62824]: DEBUG oslo_vmware.api [None req-a8698d94-6446-4cd2-b280-1af18ee5168a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146967, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2407.576608] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c534ded7-6324-4e31-8da4-2d9d57472080 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquiring lock "1db9ff97-e59c-43f6-984e-6405d8e993a4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2407.576882] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c534ded7-6324-4e31-8da4-2d9d57472080 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lock "1db9ff97-e59c-43f6-984e-6405d8e993a4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2407.577113] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c534ded7-6324-4e31-8da4-2d9d57472080 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquiring lock "1db9ff97-e59c-43f6-984e-6405d8e993a4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2407.577301] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c534ded7-6324-4e31-8da4-2d9d57472080 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lock "1db9ff97-e59c-43f6-984e-6405d8e993a4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2407.577472] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c534ded7-6324-4e31-8da4-2d9d57472080 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lock "1db9ff97-e59c-43f6-984e-6405d8e993a4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2407.580747] env[62824]: INFO nova.compute.manager [None req-c534ded7-6324-4e31-8da4-2d9d57472080 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Terminating instance [ 2407.870076] env[62824]: DEBUG nova.compute.manager [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2407.957358] env[62824]: DEBUG oslo_vmware.api [None req-a8698d94-6446-4cd2-b280-1af18ee5168a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146967, 'name': PowerOffVM_Task, 'duration_secs': 0.200248} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2407.957687] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8698d94-6446-4cd2-b280-1af18ee5168a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2407.957779] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-a8698d94-6446-4cd2-b280-1af18ee5168a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2407.958025] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2fa35a89-9e1b-4c5f-adc8-48077206f0a7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2408.084834] env[62824]: DEBUG nova.compute.manager [None req-c534ded7-6324-4e31-8da4-2d9d57472080 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2408.085024] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c534ded7-6324-4e31-8da4-2d9d57472080 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2408.085982] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eed85d42-23a6-4c78-adf1-8f6931492872 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2408.093558] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c534ded7-6324-4e31-8da4-2d9d57472080 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2408.093808] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-261c7351-6d3a-4351-b1c5-dbeb3cf173a7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2408.100393] env[62824]: DEBUG oslo_vmware.api [None req-c534ded7-6324-4e31-8da4-2d9d57472080 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2408.100393] env[62824]: value = "task-2146969" [ 2408.100393] env[62824]: _type = "Task" [ 2408.100393] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2408.107921] env[62824]: DEBUG oslo_vmware.api [None req-c534ded7-6324-4e31-8da4-2d9d57472080 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2146969, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2408.391772] env[62824]: DEBUG oslo_concurrency.lockutils [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2408.392090] env[62824]: DEBUG oslo_concurrency.lockutils [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2408.393548] env[62824]: INFO nova.compute.claims [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2408.530268] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-a8698d94-6446-4cd2-b280-1af18ee5168a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2408.530490] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-a8698d94-6446-4cd2-b280-1af18ee5168a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Deleting contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2408.530696] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8698d94-6446-4cd2-b280-1af18ee5168a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Deleting the datastore file [datastore2] 21a7eed6-c146-4911-9a0a-a1e1644e8e9d {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2408.530968] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4e2253e1-7b1c-49c7-a3e8-9339db75387f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2408.537436] env[62824]: DEBUG oslo_vmware.api [None req-a8698d94-6446-4cd2-b280-1af18ee5168a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2408.537436] env[62824]: value = "task-2146970" [ 2408.537436] env[62824]: _type = "Task" [ 2408.537436] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2408.545606] env[62824]: DEBUG oslo_vmware.api [None req-a8698d94-6446-4cd2-b280-1af18ee5168a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146970, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2408.614133] env[62824]: DEBUG oslo_vmware.api [None req-c534ded7-6324-4e31-8da4-2d9d57472080 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2146969, 'name': PowerOffVM_Task, 'duration_secs': 0.237018} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2408.614476] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c534ded7-6324-4e31-8da4-2d9d57472080 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2408.614746] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c534ded7-6324-4e31-8da4-2d9d57472080 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2408.615102] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-63fe319f-92ad-4964-b4df-965911bbc8db {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2408.848789] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c534ded7-6324-4e31-8da4-2d9d57472080 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2408.849055] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c534ded7-6324-4e31-8da4-2d9d57472080 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Deleting contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2408.849358] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-c534ded7-6324-4e31-8da4-2d9d57472080 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Deleting the datastore file [datastore2] 1db9ff97-e59c-43f6-984e-6405d8e993a4 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2408.849626] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d3629226-6a2e-4b8c-aefa-6beb8890d21a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2408.855645] env[62824]: DEBUG oslo_vmware.api [None req-c534ded7-6324-4e31-8da4-2d9d57472080 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2408.855645] env[62824]: value = "task-2146972" [ 2408.855645] env[62824]: _type = "Task" [ 2408.855645] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2408.863494] env[62824]: DEBUG oslo_vmware.api [None req-c534ded7-6324-4e31-8da4-2d9d57472080 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2146972, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2409.047062] env[62824]: DEBUG oslo_vmware.api [None req-a8698d94-6446-4cd2-b280-1af18ee5168a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146970, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.13607} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2409.047431] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8698d94-6446-4cd2-b280-1af18ee5168a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2409.047493] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-a8698d94-6446-4cd2-b280-1af18ee5168a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Deleted contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2409.047667] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-a8698d94-6446-4cd2-b280-1af18ee5168a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2409.047841] env[62824]: INFO nova.compute.manager [None req-a8698d94-6446-4cd2-b280-1af18ee5168a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Took 1.62 seconds to destroy the instance on the hypervisor. [ 2409.048090] env[62824]: DEBUG oslo.service.loopingcall [None req-a8698d94-6446-4cd2-b280-1af18ee5168a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2409.048301] env[62824]: DEBUG nova.compute.manager [-] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2409.048464] env[62824]: DEBUG nova.network.neutron [-] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2409.302634] env[62824]: DEBUG nova.compute.manager [req-eab68531-5a1a-4b81-9b25-ee749decb89d req-decdba5e-23f9-4937-8e62-e80fd9f49280 service nova] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Received event network-vif-deleted-edaf5d3c-004d-4ba0-b037-e7eef2acbe44 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2409.302908] env[62824]: INFO nova.compute.manager [req-eab68531-5a1a-4b81-9b25-ee749decb89d req-decdba5e-23f9-4937-8e62-e80fd9f49280 service nova] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Neutron deleted interface edaf5d3c-004d-4ba0-b037-e7eef2acbe44; detaching it from the instance and deleting it from the info cache [ 2409.303045] env[62824]: DEBUG nova.network.neutron [req-eab68531-5a1a-4b81-9b25-ee749decb89d req-decdba5e-23f9-4937-8e62-e80fd9f49280 service nova] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2409.365904] env[62824]: DEBUG oslo_vmware.api [None req-c534ded7-6324-4e31-8da4-2d9d57472080 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2146972, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.142983} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2409.366215] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-c534ded7-6324-4e31-8da4-2d9d57472080 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2409.366508] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c534ded7-6324-4e31-8da4-2d9d57472080 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Deleted contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2409.366745] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c534ded7-6324-4e31-8da4-2d9d57472080 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2409.366950] env[62824]: INFO nova.compute.manager [None req-c534ded7-6324-4e31-8da4-2d9d57472080 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Took 1.28 seconds to destroy the instance on the hypervisor. [ 2409.367357] env[62824]: DEBUG oslo.service.loopingcall [None req-c534ded7-6324-4e31-8da4-2d9d57472080 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2409.367605] env[62824]: DEBUG nova.compute.manager [-] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2409.367720] env[62824]: DEBUG nova.network.neutron [-] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2409.512471] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab68baa2-ce3b-477d-b161-d16e66908577 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2409.520068] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72f5d0fc-c030-4b00-b5d5-2a89ed5cdcd0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2409.548721] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d5406e2-cddc-4736-9e00-e5788a2e1891 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2409.557709] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f626761a-8f32-41ed-8686-9a1427f50ac4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2409.570614] env[62824]: DEBUG nova.compute.provider_tree [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Updating inventory in ProviderTree for provider bf95157b-a274-42de-9ccf-9851128a44a1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 2409.781894] env[62824]: DEBUG nova.network.neutron [-] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2409.805574] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-24bd7c51-20bf-4918-adb9-73f2ac9ba523 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2409.816758] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac5d2cd5-5d80-402f-9143-9daf643a5224 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2409.844680] env[62824]: DEBUG nova.compute.manager [req-eab68531-5a1a-4b81-9b25-ee749decb89d req-decdba5e-23f9-4937-8e62-e80fd9f49280 service nova] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Detach interface failed, port_id=edaf5d3c-004d-4ba0-b037-e7eef2acbe44, reason: Instance 21a7eed6-c146-4911-9a0a-a1e1644e8e9d could not be found. {{(pid=62824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11482}} [ 2410.097884] env[62824]: ERROR nova.scheduler.client.report [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [req-8ca7c4fb-90b2-4419-9753-3f444449e5a4] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID bf95157b-a274-42de-9ccf-9851128a44a1. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-8ca7c4fb-90b2-4419-9753-3f444449e5a4"}]} [ 2410.115746] env[62824]: DEBUG nova.scheduler.client.report [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Refreshing inventories for resource provider bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 2410.130064] env[62824]: DEBUG nova.scheduler.client.report [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Updating ProviderTree inventory for provider bf95157b-a274-42de-9ccf-9851128a44a1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 2410.130310] env[62824]: DEBUG nova.compute.provider_tree [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Updating inventory in ProviderTree for provider bf95157b-a274-42de-9ccf-9851128a44a1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 2410.142592] env[62824]: DEBUG nova.scheduler.client.report [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Refreshing aggregate associations for resource provider bf95157b-a274-42de-9ccf-9851128a44a1, aggregates: None {{(pid=62824) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 2410.161396] env[62824]: DEBUG nova.scheduler.client.report [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Refreshing trait associations for resource provider bf95157b-a274-42de-9ccf-9851128a44a1, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_ISO {{(pid=62824) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 2410.278044] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8079e26c-d763-4dbc-861d-87108b5018de {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2410.284271] env[62824]: INFO nova.compute.manager [-] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Took 1.24 seconds to deallocate network for instance. [ 2410.286786] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5ca191a-52cd-4b03-9c98-cc28d0e61413 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2410.321422] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f94432e-4fbd-4026-afd4-f4514c2326fd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2410.330031] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f363fde2-6a74-4872-91c4-2c0358227c07 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2410.335954] env[62824]: DEBUG nova.network.neutron [-] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2410.347691] env[62824]: DEBUG nova.compute.provider_tree [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Updating inventory in ProviderTree for provider bf95157b-a274-42de-9ccf-9851128a44a1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 2410.795384] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a8698d94-6446-4cd2-b280-1af18ee5168a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2410.839238] env[62824]: INFO nova.compute.manager [-] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Took 1.47 seconds to deallocate network for instance. [ 2410.877042] env[62824]: DEBUG nova.scheduler.client.report [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Updated inventory for provider bf95157b-a274-42de-9ccf-9851128a44a1 with generation 147 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 2410.877313] env[62824]: DEBUG nova.compute.provider_tree [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Updating resource provider bf95157b-a274-42de-9ccf-9851128a44a1 generation from 147 to 148 during operation: update_inventory {{(pid=62824) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 2410.877493] env[62824]: DEBUG nova.compute.provider_tree [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Updating inventory in ProviderTree for provider bf95157b-a274-42de-9ccf-9851128a44a1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 2411.328604] env[62824]: DEBUG nova.compute.manager [req-d4434653-59f3-4384-baf9-363950ed9bd6 req-373e4e91-a9e4-4a57-97e9-a7dd2d2cc57e service nova] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Received event network-vif-deleted-aeeadfdd-d869-4555-bda2-391c435de1bb {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2411.345159] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c534ded7-6324-4e31-8da4-2d9d57472080 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2411.382585] env[62824]: DEBUG oslo_concurrency.lockutils [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.990s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2411.383060] env[62824]: DEBUG nova.compute.manager [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2411.385621] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a8698d94-6446-4cd2-b280-1af18ee5168a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.590s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2411.385829] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a8698d94-6446-4cd2-b280-1af18ee5168a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2411.387802] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c534ded7-6324-4e31-8da4-2d9d57472080 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.043s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2411.388016] env[62824]: DEBUG nova.objects.instance [None req-c534ded7-6324-4e31-8da4-2d9d57472080 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lazy-loading 'resources' on Instance uuid 1db9ff97-e59c-43f6-984e-6405d8e993a4 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2411.405067] env[62824]: INFO nova.scheduler.client.report [None req-a8698d94-6446-4cd2-b280-1af18ee5168a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Deleted allocations for instance 21a7eed6-c146-4911-9a0a-a1e1644e8e9d [ 2411.891552] env[62824]: DEBUG nova.compute.utils [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2411.895666] env[62824]: DEBUG nova.compute.manager [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2411.895810] env[62824]: DEBUG nova.network.neutron [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2411.914493] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a8698d94-6446-4cd2-b280-1af18ee5168a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "21a7eed6-c146-4911-9a0a-a1e1644e8e9d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.990s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2411.946554] env[62824]: DEBUG nova.policy [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e8c10fcbfe69448fa71cfad4a7c8e179', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4479921caf5f405b8fc49baad390a0e9', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2412.009763] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-daa32794-a092-4b20-960c-48a89a705404 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2412.018177] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1437e12f-f120-4881-8411-06a12c1f09d9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2412.049656] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c0f16f6-382e-4b17-ac87-4b57b55a1804 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2412.057286] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29096d85-2129-402a-801e-7b39400cbf37 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2412.072279] env[62824]: DEBUG nova.compute.provider_tree [None req-c534ded7-6324-4e31-8da4-2d9d57472080 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2412.214168] env[62824]: DEBUG nova.network.neutron [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Successfully created port: 28514853-dbb7-4a89-999c-517a28061b9b {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2412.396481] env[62824]: DEBUG nova.compute.manager [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2412.576139] env[62824]: DEBUG nova.scheduler.client.report [None req-c534ded7-6324-4e31-8da4-2d9d57472080 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2413.081591] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c534ded7-6324-4e31-8da4-2d9d57472080 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.694s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2413.103609] env[62824]: INFO nova.scheduler.client.report [None req-c534ded7-6324-4e31-8da4-2d9d57472080 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Deleted allocations for instance 1db9ff97-e59c-43f6-984e-6405d8e993a4 [ 2413.405816] env[62824]: DEBUG nova.compute.manager [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2413.502146] env[62824]: DEBUG nova.virt.hardware [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2413.502389] env[62824]: DEBUG nova.virt.hardware [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2413.502599] env[62824]: DEBUG nova.virt.hardware [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2413.502789] env[62824]: DEBUG nova.virt.hardware [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2413.502936] env[62824]: DEBUG nova.virt.hardware [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2413.503096] env[62824]: DEBUG nova.virt.hardware [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2413.503321] env[62824]: DEBUG nova.virt.hardware [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2413.503481] env[62824]: DEBUG nova.virt.hardware [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2413.503651] env[62824]: DEBUG nova.virt.hardware [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2413.503824] env[62824]: DEBUG nova.virt.hardware [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2413.504093] env[62824]: DEBUG nova.virt.hardware [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2413.505278] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8c83471-3a68-465d-bf5b-57a7cb09e67f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2413.513916] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2103211-99db-4dce-90be-b77eff3eba9f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2413.610822] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c534ded7-6324-4e31-8da4-2d9d57472080 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lock "1db9ff97-e59c-43f6-984e-6405d8e993a4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.034s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2414.121943] env[62824]: DEBUG nova.compute.manager [req-412f8be8-c79f-4475-a542-ed88e2e9aee3 req-67cd0a78-a62b-4b39-a0c2-f782ebf3f063 service nova] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Received event network-vif-plugged-28514853-dbb7-4a89-999c-517a28061b9b {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2414.122175] env[62824]: DEBUG oslo_concurrency.lockutils [req-412f8be8-c79f-4475-a542-ed88e2e9aee3 req-67cd0a78-a62b-4b39-a0c2-f782ebf3f063 service nova] Acquiring lock "df621c9c-8a68-4082-8f30-9bf3dda26fb6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2414.122426] env[62824]: DEBUG oslo_concurrency.lockutils [req-412f8be8-c79f-4475-a542-ed88e2e9aee3 req-67cd0a78-a62b-4b39-a0c2-f782ebf3f063 service nova] Lock "df621c9c-8a68-4082-8f30-9bf3dda26fb6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2414.122664] env[62824]: DEBUG oslo_concurrency.lockutils [req-412f8be8-c79f-4475-a542-ed88e2e9aee3 req-67cd0a78-a62b-4b39-a0c2-f782ebf3f063 service nova] Lock "df621c9c-8a68-4082-8f30-9bf3dda26fb6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2414.122899] env[62824]: DEBUG nova.compute.manager [req-412f8be8-c79f-4475-a542-ed88e2e9aee3 req-67cd0a78-a62b-4b39-a0c2-f782ebf3f063 service nova] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] No waiting events found dispatching network-vif-plugged-28514853-dbb7-4a89-999c-517a28061b9b {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2414.123172] env[62824]: WARNING nova.compute.manager [req-412f8be8-c79f-4475-a542-ed88e2e9aee3 req-67cd0a78-a62b-4b39-a0c2-f782ebf3f063 service nova] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Received unexpected event network-vif-plugged-28514853-dbb7-4a89-999c-517a28061b9b for instance with vm_state building and task_state spawning. [ 2414.221513] env[62824]: DEBUG nova.network.neutron [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Successfully updated port: 28514853-dbb7-4a89-999c-517a28061b9b {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2414.725636] env[62824]: DEBUG oslo_concurrency.lockutils [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquiring lock "refresh_cache-df621c9c-8a68-4082-8f30-9bf3dda26fb6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2414.725636] env[62824]: DEBUG oslo_concurrency.lockutils [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquired lock "refresh_cache-df621c9c-8a68-4082-8f30-9bf3dda26fb6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2414.725636] env[62824]: DEBUG nova.network.neutron [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2415.281381] env[62824]: DEBUG nova.network.neutron [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2415.419278] env[62824]: DEBUG nova.network.neutron [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Updating instance_info_cache with network_info: [{"id": "28514853-dbb7-4a89-999c-517a28061b9b", "address": "fa:16:3e:f1:72:8d", "network": {"id": "12411632-b7f6-4544-b2e4-b04ae7337547", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-917106830-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4479921caf5f405b8fc49baad390a0e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "de7fa486-5f28-44ae-b0cf-72234ff87546", "external-id": "nsx-vlan-transportzone-229", "segmentation_id": 229, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap28514853-db", "ovs_interfaceid": "28514853-dbb7-4a89-999c-517a28061b9b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2415.466260] env[62824]: DEBUG nova.compute.manager [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Stashing vm_state: active {{(pid=62824) _prep_resize /opt/stack/nova/nova/compute/manager.py:6136}} [ 2415.791364] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2bb6e12b-a776-49f0-b53c-5a6cb693ac70 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquiring lock "3d294ee9-92f6-4115-95db-9d4b66562b56" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2415.791886] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2bb6e12b-a776-49f0-b53c-5a6cb693ac70 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "3d294ee9-92f6-4115-95db-9d4b66562b56" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2415.922024] env[62824]: DEBUG oslo_concurrency.lockutils [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Releasing lock "refresh_cache-df621c9c-8a68-4082-8f30-9bf3dda26fb6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2415.922355] env[62824]: DEBUG nova.compute.manager [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Instance network_info: |[{"id": "28514853-dbb7-4a89-999c-517a28061b9b", "address": "fa:16:3e:f1:72:8d", "network": {"id": "12411632-b7f6-4544-b2e4-b04ae7337547", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-917106830-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4479921caf5f405b8fc49baad390a0e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "de7fa486-5f28-44ae-b0cf-72234ff87546", "external-id": "nsx-vlan-transportzone-229", "segmentation_id": 229, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap28514853-db", "ovs_interfaceid": "28514853-dbb7-4a89-999c-517a28061b9b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2415.922826] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f1:72:8d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'de7fa486-5f28-44ae-b0cf-72234ff87546', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '28514853-dbb7-4a89-999c-517a28061b9b', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2415.930486] env[62824]: DEBUG oslo.service.loopingcall [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2415.931062] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2415.931305] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-96d3db2e-945c-4f46-b497-ff99ada9c765 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2415.951707] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2415.951707] env[62824]: value = "task-2146973" [ 2415.951707] env[62824]: _type = "Task" [ 2415.951707] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2415.959267] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146973, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2415.985147] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2415.985446] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2416.149251] env[62824]: DEBUG nova.compute.manager [req-8a71a59b-3b9d-416f-9014-f0ceefd3dc1e req-2542f4f8-6626-4915-91c4-8ca1d8711829 service nova] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Received event network-changed-28514853-dbb7-4a89-999c-517a28061b9b {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2416.149304] env[62824]: DEBUG nova.compute.manager [req-8a71a59b-3b9d-416f-9014-f0ceefd3dc1e req-2542f4f8-6626-4915-91c4-8ca1d8711829 service nova] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Refreshing instance network info cache due to event network-changed-28514853-dbb7-4a89-999c-517a28061b9b. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2416.149546] env[62824]: DEBUG oslo_concurrency.lockutils [req-8a71a59b-3b9d-416f-9014-f0ceefd3dc1e req-2542f4f8-6626-4915-91c4-8ca1d8711829 service nova] Acquiring lock "refresh_cache-df621c9c-8a68-4082-8f30-9bf3dda26fb6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2416.149673] env[62824]: DEBUG oslo_concurrency.lockutils [req-8a71a59b-3b9d-416f-9014-f0ceefd3dc1e req-2542f4f8-6626-4915-91c4-8ca1d8711829 service nova] Acquired lock "refresh_cache-df621c9c-8a68-4082-8f30-9bf3dda26fb6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2416.149824] env[62824]: DEBUG nova.network.neutron [req-8a71a59b-3b9d-416f-9014-f0ceefd3dc1e req-2542f4f8-6626-4915-91c4-8ca1d8711829 service nova] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Refreshing network info cache for port 28514853-dbb7-4a89-999c-517a28061b9b {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2416.295034] env[62824]: INFO nova.compute.manager [None req-2bb6e12b-a776-49f0-b53c-5a6cb693ac70 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Detaching volume 42b59773-27a4-492d-b430-c629b14a2500 [ 2416.329821] env[62824]: INFO nova.virt.block_device [None req-2bb6e12b-a776-49f0-b53c-5a6cb693ac70 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Attempting to driver detach volume 42b59773-27a4-492d-b430-c629b14a2500 from mountpoint /dev/sdb [ 2416.330156] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-2bb6e12b-a776-49f0-b53c-5a6cb693ac70 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Volume detach. Driver type: vmdk {{(pid=62824) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 2416.330294] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-2bb6e12b-a776-49f0-b53c-5a6cb693ac70 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-438761', 'volume_id': '42b59773-27a4-492d-b430-c629b14a2500', 'name': 'volume-42b59773-27a4-492d-b430-c629b14a2500', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '3d294ee9-92f6-4115-95db-9d4b66562b56', 'attached_at': '', 'detached_at': '', 'volume_id': '42b59773-27a4-492d-b430-c629b14a2500', 'serial': '42b59773-27a4-492d-b430-c629b14a2500'} {{(pid=62824) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 2416.331217] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47332416-41c4-4187-a724-79f90450bf35 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2416.352280] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7aede82f-fc35-4133-9d49-8d33334d1e59 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2416.359330] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b689d38-8079-452a-812f-695242d4898a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2416.379504] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ddbad53-15f6-43eb-a736-9464dfa31b6b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2416.393702] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-2bb6e12b-a776-49f0-b53c-5a6cb693ac70 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] The volume has not been displaced from its original location: [datastore2] volume-42b59773-27a4-492d-b430-c629b14a2500/volume-42b59773-27a4-492d-b430-c629b14a2500.vmdk. No consolidation needed. {{(pid=62824) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 2416.399025] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-2bb6e12b-a776-49f0-b53c-5a6cb693ac70 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Reconfiguring VM instance instance-00000066 to detach disk 2001 {{(pid=62824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 2416.399117] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ef5f094d-ace4-4f4f-a38b-9d47e6c8fcbd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2416.415983] env[62824]: DEBUG oslo_vmware.api [None req-2bb6e12b-a776-49f0-b53c-5a6cb693ac70 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2416.415983] env[62824]: value = "task-2146974" [ 2416.415983] env[62824]: _type = "Task" [ 2416.415983] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2416.423964] env[62824]: DEBUG oslo_vmware.api [None req-2bb6e12b-a776-49f0-b53c-5a6cb693ac70 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146974, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2416.460772] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2146973, 'name': CreateVM_Task, 'duration_secs': 0.434944} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2416.460928] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2416.461638] env[62824]: DEBUG oslo_concurrency.lockutils [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2416.461818] env[62824]: DEBUG oslo_concurrency.lockutils [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2416.462186] env[62824]: DEBUG oslo_concurrency.lockutils [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2416.462469] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-225ae54a-05f6-480c-ac8e-52fc39b508f2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2416.466657] env[62824]: DEBUG oslo_vmware.api [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2416.466657] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]529f6788-0d5c-737f-bc91-b24d2e81bdcb" [ 2416.466657] env[62824]: _type = "Task" [ 2416.466657] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2416.473894] env[62824]: DEBUG oslo_vmware.api [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]529f6788-0d5c-737f-bc91-b24d2e81bdcb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2416.490889] env[62824]: INFO nova.compute.claims [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2416.853754] env[62824]: DEBUG nova.network.neutron [req-8a71a59b-3b9d-416f-9014-f0ceefd3dc1e req-2542f4f8-6626-4915-91c4-8ca1d8711829 service nova] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Updated VIF entry in instance network info cache for port 28514853-dbb7-4a89-999c-517a28061b9b. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2416.854163] env[62824]: DEBUG nova.network.neutron [req-8a71a59b-3b9d-416f-9014-f0ceefd3dc1e req-2542f4f8-6626-4915-91c4-8ca1d8711829 service nova] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Updating instance_info_cache with network_info: [{"id": "28514853-dbb7-4a89-999c-517a28061b9b", "address": "fa:16:3e:f1:72:8d", "network": {"id": "12411632-b7f6-4544-b2e4-b04ae7337547", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-917106830-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4479921caf5f405b8fc49baad390a0e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "de7fa486-5f28-44ae-b0cf-72234ff87546", "external-id": "nsx-vlan-transportzone-229", "segmentation_id": 229, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap28514853-db", "ovs_interfaceid": "28514853-dbb7-4a89-999c-517a28061b9b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2416.926750] env[62824]: DEBUG oslo_vmware.api [None req-2bb6e12b-a776-49f0-b53c-5a6cb693ac70 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146974, 'name': ReconfigVM_Task, 'duration_secs': 0.213502} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2416.927074] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-2bb6e12b-a776-49f0-b53c-5a6cb693ac70 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Reconfigured VM instance instance-00000066 to detach disk 2001 {{(pid=62824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 2416.931860] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0061c2e9-c5ad-4044-96d5-94e7f75da3fc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2416.945810] env[62824]: DEBUG oslo_vmware.api [None req-2bb6e12b-a776-49f0-b53c-5a6cb693ac70 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2416.945810] env[62824]: value = "task-2146975" [ 2416.945810] env[62824]: _type = "Task" [ 2416.945810] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2416.953101] env[62824]: DEBUG oslo_vmware.api [None req-2bb6e12b-a776-49f0-b53c-5a6cb693ac70 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146975, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2416.975411] env[62824]: DEBUG oslo_vmware.api [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]529f6788-0d5c-737f-bc91-b24d2e81bdcb, 'name': SearchDatastore_Task, 'duration_secs': 0.010518} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2416.975696] env[62824]: DEBUG oslo_concurrency.lockutils [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2416.975940] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2416.976187] env[62824]: DEBUG oslo_concurrency.lockutils [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2416.976334] env[62824]: DEBUG oslo_concurrency.lockutils [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2416.976514] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2416.976747] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f37bff1f-2987-4059-881c-a83fa1be493d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2416.984493] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2416.984666] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2416.985353] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-45e9c374-f372-4191-802c-c3dcff1586ff {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2416.990443] env[62824]: DEBUG oslo_vmware.api [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2416.990443] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5241d17d-2daa-0ec6-b095-7ddd080fe47a" [ 2416.990443] env[62824]: _type = "Task" [ 2416.990443] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2416.999945] env[62824]: INFO nova.compute.resource_tracker [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Updating resource usage from migration a4f6b1ba-8f40-4ece-af3a-39164086942a [ 2417.002738] env[62824]: DEBUG oslo_vmware.api [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5241d17d-2daa-0ec6-b095-7ddd080fe47a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2417.113078] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e2e04b5-ab41-49d6-b3b7-76127ed20430 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2417.120844] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16214ec2-4109-49fe-ba49-10f006d64185 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2417.126253] env[62824]: DEBUG oslo_concurrency.lockutils [None req-01bd391a-ace2-4614-b82d-cfb73c68f7c9 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Acquiring lock "c70880c9-a69a-49b9-b028-189c04276623" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2417.126475] env[62824]: DEBUG oslo_concurrency.lockutils [None req-01bd391a-ace2-4614-b82d-cfb73c68f7c9 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Lock "c70880c9-a69a-49b9-b028-189c04276623" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2417.153390] env[62824]: DEBUG nova.compute.utils [None req-01bd391a-ace2-4614-b82d-cfb73c68f7c9 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2417.154983] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be0af233-4fe0-442d-b90e-4c2b505514a0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2417.164116] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7052436-3a70-4f05-a46f-06214f0b0a9d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2417.179267] env[62824]: DEBUG nova.compute.provider_tree [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2417.357258] env[62824]: DEBUG oslo_concurrency.lockutils [req-8a71a59b-3b9d-416f-9014-f0ceefd3dc1e req-2542f4f8-6626-4915-91c4-8ca1d8711829 service nova] Releasing lock "refresh_cache-df621c9c-8a68-4082-8f30-9bf3dda26fb6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2417.455315] env[62824]: DEBUG oslo_vmware.api [None req-2bb6e12b-a776-49f0-b53c-5a6cb693ac70 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146975, 'name': ReconfigVM_Task, 'duration_secs': 0.145452} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2417.455566] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-2bb6e12b-a776-49f0-b53c-5a6cb693ac70 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-438761', 'volume_id': '42b59773-27a4-492d-b430-c629b14a2500', 'name': 'volume-42b59773-27a4-492d-b430-c629b14a2500', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '3d294ee9-92f6-4115-95db-9d4b66562b56', 'attached_at': '', 'detached_at': '', 'volume_id': '42b59773-27a4-492d-b430-c629b14a2500', 'serial': '42b59773-27a4-492d-b430-c629b14a2500'} {{(pid=62824) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 2417.499577] env[62824]: DEBUG oslo_vmware.api [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5241d17d-2daa-0ec6-b095-7ddd080fe47a, 'name': SearchDatastore_Task, 'duration_secs': 0.008905} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2417.500335] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0c6caa4e-8365-4b45-a17e-a95e94b5b938 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2417.505334] env[62824]: DEBUG oslo_vmware.api [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2417.505334] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52748dd2-de54-8a18-6195-ae694b369b9b" [ 2417.505334] env[62824]: _type = "Task" [ 2417.505334] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2417.512954] env[62824]: DEBUG oslo_vmware.api [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52748dd2-de54-8a18-6195-ae694b369b9b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2417.659084] env[62824]: DEBUG oslo_concurrency.lockutils [None req-01bd391a-ace2-4614-b82d-cfb73c68f7c9 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Lock "c70880c9-a69a-49b9-b028-189c04276623" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 0.532s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2417.682225] env[62824]: DEBUG nova.scheduler.client.report [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2417.995796] env[62824]: DEBUG nova.objects.instance [None req-2bb6e12b-a776-49f0-b53c-5a6cb693ac70 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lazy-loading 'flavor' on Instance uuid 3d294ee9-92f6-4115-95db-9d4b66562b56 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2418.015581] env[62824]: DEBUG oslo_vmware.api [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52748dd2-de54-8a18-6195-ae694b369b9b, 'name': SearchDatastore_Task, 'duration_secs': 0.009557} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2418.016363] env[62824]: DEBUG oslo_concurrency.lockutils [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2418.016634] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] df621c9c-8a68-4082-8f30-9bf3dda26fb6/df621c9c-8a68-4082-8f30-9bf3dda26fb6.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2418.016893] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0f39079d-2438-4eb7-9373-b87c9c4ab4f7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2418.024954] env[62824]: DEBUG oslo_vmware.api [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2418.024954] env[62824]: value = "task-2146976" [ 2418.024954] env[62824]: _type = "Task" [ 2418.024954] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2418.033055] env[62824]: DEBUG oslo_vmware.api [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146976, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2418.186994] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.201s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2418.187266] env[62824]: INFO nova.compute.manager [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Migrating [ 2418.534852] env[62824]: DEBUG oslo_vmware.api [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146976, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.428002} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2418.535118] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] df621c9c-8a68-4082-8f30-9bf3dda26fb6/df621c9c-8a68-4082-8f30-9bf3dda26fb6.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2418.535333] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2418.535567] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-50e93d72-fda8-4f7d-b53b-f5e28b5c311c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2418.543079] env[62824]: DEBUG oslo_vmware.api [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2418.543079] env[62824]: value = "task-2146977" [ 2418.543079] env[62824]: _type = "Task" [ 2418.543079] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2418.551706] env[62824]: DEBUG oslo_vmware.api [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146977, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2418.704083] env[62824]: DEBUG oslo_concurrency.lockutils [None req-01bd391a-ace2-4614-b82d-cfb73c68f7c9 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Acquiring lock "c70880c9-a69a-49b9-b028-189c04276623" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2418.704374] env[62824]: DEBUG oslo_concurrency.lockutils [None req-01bd391a-ace2-4614-b82d-cfb73c68f7c9 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Lock "c70880c9-a69a-49b9-b028-189c04276623" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2418.704686] env[62824]: INFO nova.compute.manager [None req-01bd391a-ace2-4614-b82d-cfb73c68f7c9 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c70880c9-a69a-49b9-b028-189c04276623] Attaching volume 42f33a8d-0a1f-4132-8b0d-90a607343a57 to /dev/sdb [ 2418.706447] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquiring lock "refresh_cache-5b3df31f-15fe-473f-992c-ddb272661c53" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2418.706598] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquired lock "refresh_cache-5b3df31f-15fe-473f-992c-ddb272661c53" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2418.706767] env[62824]: DEBUG nova.network.neutron [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2418.737398] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38152218-aee2-46da-982a-412d6ed1af82 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2418.744445] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a9ea617-2bee-48d8-998d-ad319b922dc3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2418.757944] env[62824]: DEBUG nova.virt.block_device [None req-01bd391a-ace2-4614-b82d-cfb73c68f7c9 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c70880c9-a69a-49b9-b028-189c04276623] Updating existing volume attachment record: b5f96c48-86af-46e3-a633-0e3628c7ca24 {{(pid=62824) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 2419.003492] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2bb6e12b-a776-49f0-b53c-5a6cb693ac70 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "3d294ee9-92f6-4115-95db-9d4b66562b56" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.212s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2419.053078] env[62824]: DEBUG oslo_vmware.api [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146977, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.057648} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2419.053349] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2419.054116] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d389e0d-4e28-47cf-b98b-0fe68d89a3c1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2419.076354] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Reconfiguring VM instance instance-00000070 to attach disk [datastore2] df621c9c-8a68-4082-8f30-9bf3dda26fb6/df621c9c-8a68-4082-8f30-9bf3dda26fb6.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2419.076628] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-59ed47c1-13bd-4970-aa9c-1cfc651e79a2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2419.097567] env[62824]: DEBUG oslo_vmware.api [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2419.097567] env[62824]: value = "task-2146981" [ 2419.097567] env[62824]: _type = "Task" [ 2419.097567] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2419.107861] env[62824]: DEBUG oslo_vmware.api [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146981, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2419.419945] env[62824]: DEBUG nova.network.neutron [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Updating instance_info_cache with network_info: [{"id": "7be95c5e-6e98-4fec-8550-1c22a787b95b", "address": "fa:16:3e:e8:69:3a", "network": {"id": "bba9b98c-c7eb-4890-8de5-eafacfdebd7d", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1928394728-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.160", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7fcdbfc29df64a1ba8d982bdcc667b64", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b49c5024-2ced-42ca-90cc-6066766d43e6", "external-id": "nsx-vlan-transportzone-239", "segmentation_id": 239, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7be95c5e-6e", "ovs_interfaceid": "7be95c5e-6e98-4fec-8550-1c22a787b95b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2419.607074] env[62824]: DEBUG oslo_vmware.api [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146981, 'name': ReconfigVM_Task, 'duration_secs': 0.317312} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2419.607220] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Reconfigured VM instance instance-00000070 to attach disk [datastore2] df621c9c-8a68-4082-8f30-9bf3dda26fb6/df621c9c-8a68-4082-8f30-9bf3dda26fb6.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2419.607735] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-363930ad-69ed-4719-9c90-e2ce2855c905 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2419.614900] env[62824]: DEBUG oslo_vmware.api [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2419.614900] env[62824]: value = "task-2146982" [ 2419.614900] env[62824]: _type = "Task" [ 2419.614900] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2419.623235] env[62824]: DEBUG oslo_vmware.api [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146982, 'name': Rename_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2419.692111] env[62824]: INFO nova.compute.manager [None req-047f4adc-be01-4800-b20c-0eebdf663a56 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412] Get console output [ 2419.692403] env[62824]: WARNING nova.virt.vmwareapi.driver [None req-047f4adc-be01-4800-b20c-0eebdf663a56 tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412] The console log is missing. Check your VSPC configuration [ 2419.923324] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Releasing lock "refresh_cache-5b3df31f-15fe-473f-992c-ddb272661c53" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2420.044141] env[62824]: DEBUG oslo_concurrency.lockutils [None req-60e3c005-4f57-43b7-8102-c73553b29f5e tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquiring lock "3d294ee9-92f6-4115-95db-9d4b66562b56" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2420.044541] env[62824]: DEBUG oslo_concurrency.lockutils [None req-60e3c005-4f57-43b7-8102-c73553b29f5e tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "3d294ee9-92f6-4115-95db-9d4b66562b56" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2420.044790] env[62824]: DEBUG oslo_concurrency.lockutils [None req-60e3c005-4f57-43b7-8102-c73553b29f5e tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquiring lock "3d294ee9-92f6-4115-95db-9d4b66562b56-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2420.045054] env[62824]: DEBUG oslo_concurrency.lockutils [None req-60e3c005-4f57-43b7-8102-c73553b29f5e tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "3d294ee9-92f6-4115-95db-9d4b66562b56-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2420.045296] env[62824]: DEBUG oslo_concurrency.lockutils [None req-60e3c005-4f57-43b7-8102-c73553b29f5e tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "3d294ee9-92f6-4115-95db-9d4b66562b56-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2420.047323] env[62824]: INFO nova.compute.manager [None req-60e3c005-4f57-43b7-8102-c73553b29f5e tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Terminating instance [ 2420.124438] env[62824]: DEBUG oslo_vmware.api [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146982, 'name': Rename_Task, 'duration_secs': 0.13143} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2420.124673] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2420.124899] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-91dd58a6-7e7c-4ff5-9926-da889556b3aa {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2420.130781] env[62824]: DEBUG oslo_vmware.api [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2420.130781] env[62824]: value = "task-2146983" [ 2420.130781] env[62824]: _type = "Task" [ 2420.130781] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2420.137694] env[62824]: DEBUG oslo_vmware.api [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146983, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2420.551065] env[62824]: DEBUG nova.compute.manager [None req-60e3c005-4f57-43b7-8102-c73553b29f5e tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2420.551261] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-60e3c005-4f57-43b7-8102-c73553b29f5e tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2420.552311] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17e5ce89-eb3b-4543-ab12-53de22720154 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2420.560179] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-60e3c005-4f57-43b7-8102-c73553b29f5e tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2420.560460] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-48f8b6f9-2be8-41f4-8353-c5808c140ba8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2420.566087] env[62824]: DEBUG oslo_vmware.api [None req-60e3c005-4f57-43b7-8102-c73553b29f5e tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2420.566087] env[62824]: value = "task-2146984" [ 2420.566087] env[62824]: _type = "Task" [ 2420.566087] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2420.573364] env[62824]: DEBUG oslo_vmware.api [None req-60e3c005-4f57-43b7-8102-c73553b29f5e tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146984, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2420.640853] env[62824]: DEBUG oslo_vmware.api [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146983, 'name': PowerOnVM_Task, 'duration_secs': 0.442384} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2420.641200] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2420.641435] env[62824]: INFO nova.compute.manager [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Took 7.24 seconds to spawn the instance on the hypervisor. [ 2420.641643] env[62824]: DEBUG nova.compute.manager [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2420.642497] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa41b642-cad5-4499-8fd5-d064acccb208 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2420.773015] env[62824]: DEBUG oslo_concurrency.lockutils [None req-128cbbf1-0973-4bf7-8fb2-3090e28cc3aa tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Acquiring lock "c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2420.773376] env[62824]: DEBUG oslo_concurrency.lockutils [None req-128cbbf1-0973-4bf7-8fb2-3090e28cc3aa tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Lock "c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2420.773651] env[62824]: DEBUG oslo_concurrency.lockutils [None req-128cbbf1-0973-4bf7-8fb2-3090e28cc3aa tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Acquiring lock "c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2420.773874] env[62824]: DEBUG oslo_concurrency.lockutils [None req-128cbbf1-0973-4bf7-8fb2-3090e28cc3aa tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Lock "c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2420.774121] env[62824]: DEBUG oslo_concurrency.lockutils [None req-128cbbf1-0973-4bf7-8fb2-3090e28cc3aa tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Lock "c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2420.776385] env[62824]: INFO nova.compute.manager [None req-128cbbf1-0973-4bf7-8fb2-3090e28cc3aa tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412] Terminating instance [ 2421.077051] env[62824]: DEBUG oslo_vmware.api [None req-60e3c005-4f57-43b7-8102-c73553b29f5e tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146984, 'name': PowerOffVM_Task, 'duration_secs': 0.20843} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2421.077385] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-60e3c005-4f57-43b7-8102-c73553b29f5e tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2421.077530] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-60e3c005-4f57-43b7-8102-c73553b29f5e tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2421.077795] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-667fdcba-9976-416e-a5c8-09c6fe15b0f2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2421.161914] env[62824]: INFO nova.compute.manager [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Took 12.79 seconds to build instance. [ 2421.280592] env[62824]: DEBUG nova.compute.manager [None req-128cbbf1-0973-4bf7-8fb2-3090e28cc3aa tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2421.280881] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-128cbbf1-0973-4bf7-8fb2-3090e28cc3aa tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2421.281859] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6ee99ca-e170-4099-9e11-beded1247b4b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2421.289409] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-128cbbf1-0973-4bf7-8fb2-3090e28cc3aa tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2421.289655] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f3a0396e-4132-4869-bee0-afeb36cfbc4b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2421.295564] env[62824]: DEBUG oslo_vmware.api [None req-128cbbf1-0973-4bf7-8fb2-3090e28cc3aa tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Waiting for the task: (returnval){ [ 2421.295564] env[62824]: value = "task-2146987" [ 2421.295564] env[62824]: _type = "Task" [ 2421.295564] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2421.304771] env[62824]: DEBUG oslo_vmware.api [None req-128cbbf1-0973-4bf7-8fb2-3090e28cc3aa tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Task: {'id': task-2146987, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2421.440112] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6f7c8c3-cfa5-4528-9fd7-f8373599d777 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2421.468728] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Updating instance '5b3df31f-15fe-473f-992c-ddb272661c53' progress to 0 {{(pid=62824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 2421.663980] env[62824]: DEBUG oslo_concurrency.lockutils [None req-15091081-2189-4e7e-a7b7-141eef39f962 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "df621c9c-8a68-4082-8f30-9bf3dda26fb6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.296s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2421.806629] env[62824]: DEBUG oslo_vmware.api [None req-128cbbf1-0973-4bf7-8fb2-3090e28cc3aa tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Task: {'id': task-2146987, 'name': PowerOffVM_Task, 'duration_secs': 0.180456} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2421.806963] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-128cbbf1-0973-4bf7-8fb2-3090e28cc3aa tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2421.807211] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-128cbbf1-0973-4bf7-8fb2-3090e28cc3aa tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2421.807545] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6717182c-db70-4dd6-84be-948ddc8819c0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2421.980363] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2421.980745] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5a597d9c-e1f9-42ae-a54d-783cce310a30 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2421.991455] env[62824]: DEBUG oslo_vmware.api [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2421.991455] env[62824]: value = "task-2146989" [ 2421.991455] env[62824]: _type = "Task" [ 2421.991455] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2422.000167] env[62824]: DEBUG oslo_vmware.api [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2146989, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2422.499073] env[62824]: DEBUG nova.compute.manager [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Stashing vm_state: active {{(pid=62824) _prep_resize /opt/stack/nova/nova/compute/manager.py:6136}} [ 2422.507862] env[62824]: DEBUG oslo_vmware.api [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2146989, 'name': PowerOffVM_Task, 'duration_secs': 0.200003} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2422.508132] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2422.508315] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Updating instance '5b3df31f-15fe-473f-992c-ddb272661c53' progress to 17 {{(pid=62824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 2423.015180] env[62824]: DEBUG nova.virt.hardware [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:24Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2423.015529] env[62824]: DEBUG nova.virt.hardware [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2423.015752] env[62824]: DEBUG nova.virt.hardware [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2423.016058] env[62824]: DEBUG nova.virt.hardware [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2423.016373] env[62824]: DEBUG nova.virt.hardware [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2423.016593] env[62824]: DEBUG nova.virt.hardware [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2423.017232] env[62824]: DEBUG nova.virt.hardware [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2423.017232] env[62824]: DEBUG nova.virt.hardware [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2423.017232] env[62824]: DEBUG nova.virt.hardware [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2423.017413] env[62824]: DEBUG nova.virt.hardware [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2423.017577] env[62824]: DEBUG nova.virt.hardware [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2423.023409] env[62824]: DEBUG oslo_concurrency.lockutils [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2423.023641] env[62824]: DEBUG oslo_concurrency.lockutils [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2423.025008] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cd75c89b-4422-44a1-9dd0-4d07110cc460 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2423.041924] env[62824]: DEBUG oslo_vmware.api [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2423.041924] env[62824]: value = "task-2146990" [ 2423.041924] env[62824]: _type = "Task" [ 2423.041924] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2423.050262] env[62824]: DEBUG oslo_vmware.api [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2146990, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2423.300504] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-01bd391a-ace2-4614-b82d-cfb73c68f7c9 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c70880c9-a69a-49b9-b028-189c04276623] Volume attach. Driver type: vmdk {{(pid=62824) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 2423.300752] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-01bd391a-ace2-4614-b82d-cfb73c68f7c9 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c70880c9-a69a-49b9-b028-189c04276623] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-438775', 'volume_id': '42f33a8d-0a1f-4132-8b0d-90a607343a57', 'name': 'volume-42f33a8d-0a1f-4132-8b0d-90a607343a57', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'c70880c9-a69a-49b9-b028-189c04276623', 'attached_at': '', 'detached_at': '', 'volume_id': '42f33a8d-0a1f-4132-8b0d-90a607343a57', 'serial': '42f33a8d-0a1f-4132-8b0d-90a607343a57'} {{(pid=62824) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 2423.301670] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9f9123a-2a25-40fa-abbe-1f4384f3d433 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2423.319373] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-325e2b99-284c-411b-bcc9-87ac86d06434 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2423.343901] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-01bd391a-ace2-4614-b82d-cfb73c68f7c9 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c70880c9-a69a-49b9-b028-189c04276623] Reconfiguring VM instance instance-0000006e to attach disk [datastore1] volume-42f33a8d-0a1f-4132-8b0d-90a607343a57/volume-42f33a8d-0a1f-4132-8b0d-90a607343a57.vmdk or device None with type thin {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2423.344221] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1b84a9ad-f848-4449-afef-bf1fc407aa36 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2423.363778] env[62824]: DEBUG oslo_vmware.api [None req-01bd391a-ace2-4614-b82d-cfb73c68f7c9 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Waiting for the task: (returnval){ [ 2423.363778] env[62824]: value = "task-2146991" [ 2423.363778] env[62824]: _type = "Task" [ 2423.363778] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2423.371739] env[62824]: DEBUG oslo_vmware.api [None req-01bd391a-ace2-4614-b82d-cfb73c68f7c9 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2146991, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2423.539137] env[62824]: INFO nova.compute.claims [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2423.553714] env[62824]: DEBUG oslo_vmware.api [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2146990, 'name': ReconfigVM_Task, 'duration_secs': 0.149178} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2423.553988] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Updating instance '5b3df31f-15fe-473f-992c-ddb272661c53' progress to 33 {{(pid=62824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 2423.873805] env[62824]: DEBUG oslo_vmware.api [None req-01bd391a-ace2-4614-b82d-cfb73c68f7c9 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2146991, 'name': ReconfigVM_Task, 'duration_secs': 0.355605} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2423.874242] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-01bd391a-ace2-4614-b82d-cfb73c68f7c9 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c70880c9-a69a-49b9-b028-189c04276623] Reconfigured VM instance instance-0000006e to attach disk [datastore1] volume-42f33a8d-0a1f-4132-8b0d-90a607343a57/volume-42f33a8d-0a1f-4132-8b0d-90a607343a57.vmdk or device None with type thin {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2423.878998] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5908089e-5f40-4f39-91c7-2141b5a6f38a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2423.893168] env[62824]: DEBUG oslo_vmware.api [None req-01bd391a-ace2-4614-b82d-cfb73c68f7c9 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Waiting for the task: (returnval){ [ 2423.893168] env[62824]: value = "task-2146992" [ 2423.893168] env[62824]: _type = "Task" [ 2423.893168] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2423.900572] env[62824]: DEBUG oslo_vmware.api [None req-01bd391a-ace2-4614-b82d-cfb73c68f7c9 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2146992, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2424.050675] env[62824]: INFO nova.compute.resource_tracker [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Updating resource usage from migration a132a1b4-4c08-4cae-bc8c-af84308301aa [ 2424.060267] env[62824]: DEBUG nova.virt.hardware [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2424.060530] env[62824]: DEBUG nova.virt.hardware [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2424.060690] env[62824]: DEBUG nova.virt.hardware [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2424.060875] env[62824]: DEBUG nova.virt.hardware [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2424.061033] env[62824]: DEBUG nova.virt.hardware [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2424.061184] env[62824]: DEBUG nova.virt.hardware [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2424.061387] env[62824]: DEBUG nova.virt.hardware [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2424.061541] env[62824]: DEBUG nova.virt.hardware [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2424.061729] env[62824]: DEBUG nova.virt.hardware [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2424.061904] env[62824]: DEBUG nova.virt.hardware [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2424.062091] env[62824]: DEBUG nova.virt.hardware [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2424.067294] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Reconfiguring VM instance instance-00000043 to detach disk 2000 {{(pid=62824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 2424.070214] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d91e4b0a-5f2f-4f8c-87a4-0ed6803d3079 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2424.088708] env[62824]: DEBUG oslo_vmware.api [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2424.088708] env[62824]: value = "task-2146993" [ 2424.088708] env[62824]: _type = "Task" [ 2424.088708] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2424.099999] env[62824]: DEBUG oslo_vmware.api [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2146993, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2424.173010] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a11ed96-d6ab-4ba7-bcbb-e0850e06619f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2424.180192] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fc25cf6-18f4-402c-b167-0da785571fde {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2424.210369] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c176f336-a902-4e1f-838d-d6ec4a2a561d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2424.217427] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94e5fbe3-c167-4984-ba46-7e96e6916260 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2424.230403] env[62824]: DEBUG nova.compute.provider_tree [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2424.403054] env[62824]: DEBUG oslo_vmware.api [None req-01bd391a-ace2-4614-b82d-cfb73c68f7c9 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2146992, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2424.599274] env[62824]: DEBUG oslo_vmware.api [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2146993, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2424.733714] env[62824]: DEBUG nova.scheduler.client.report [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2424.903599] env[62824]: DEBUG oslo_vmware.api [None req-01bd391a-ace2-4614-b82d-cfb73c68f7c9 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2146992, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2425.099989] env[62824]: DEBUG oslo_vmware.api [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2146993, 'name': ReconfigVM_Task, 'duration_secs': 0.544555} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2425.103051] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Reconfigured VM instance instance-00000043 to detach disk 2000 {{(pid=62824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 2425.103051] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11c09b23-b3af-47dd-86e7-e027766ba65c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2425.123200] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Reconfiguring VM instance instance-00000043 to attach disk [datastore1] 5b3df31f-15fe-473f-992c-ddb272661c53/5b3df31f-15fe-473f-992c-ddb272661c53.vmdk or device None with type thin {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2425.123379] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4b950eeb-aa57-4df3-88aa-866b9eecf661 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2425.140924] env[62824]: DEBUG oslo_vmware.api [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2425.140924] env[62824]: value = "task-2146994" [ 2425.140924] env[62824]: _type = "Task" [ 2425.140924] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2425.148587] env[62824]: DEBUG oslo_vmware.api [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2146994, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2425.238714] env[62824]: DEBUG oslo_concurrency.lockutils [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.215s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2425.238934] env[62824]: INFO nova.compute.manager [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Migrating [ 2425.404442] env[62824]: DEBUG oslo_vmware.api [None req-01bd391a-ace2-4614-b82d-cfb73c68f7c9 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2146992, 'name': ReconfigVM_Task, 'duration_secs': 1.1362} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2425.404710] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-01bd391a-ace2-4614-b82d-cfb73c68f7c9 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c70880c9-a69a-49b9-b028-189c04276623] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-438775', 'volume_id': '42f33a8d-0a1f-4132-8b0d-90a607343a57', 'name': 'volume-42f33a8d-0a1f-4132-8b0d-90a607343a57', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'c70880c9-a69a-49b9-b028-189c04276623', 'attached_at': '', 'detached_at': '', 'volume_id': '42f33a8d-0a1f-4132-8b0d-90a607343a57', 'serial': '42f33a8d-0a1f-4132-8b0d-90a607343a57'} {{(pid=62824) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 2425.651253] env[62824]: DEBUG oslo_vmware.api [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2146994, 'name': ReconfigVM_Task, 'duration_secs': 0.252892} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2425.651603] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Reconfigured VM instance instance-00000043 to attach disk [datastore1] 5b3df31f-15fe-473f-992c-ddb272661c53/5b3df31f-15fe-473f-992c-ddb272661c53.vmdk or device None with type thin {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2425.651808] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Updating instance '5b3df31f-15fe-473f-992c-ddb272661c53' progress to 50 {{(pid=62824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 2425.753195] env[62824]: DEBUG oslo_concurrency.lockutils [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquiring lock "refresh_cache-df621c9c-8a68-4082-8f30-9bf3dda26fb6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2425.753378] env[62824]: DEBUG oslo_concurrency.lockutils [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquired lock "refresh_cache-df621c9c-8a68-4082-8f30-9bf3dda26fb6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2425.753556] env[62824]: DEBUG nova.network.neutron [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2426.158092] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a3f8802-308b-40ad-ba25-ff7f0f7bcd1f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2426.177359] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdab5ca1-ad3c-4e70-8d40-214987a27707 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2426.194855] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Updating instance '5b3df31f-15fe-473f-992c-ddb272661c53' progress to 67 {{(pid=62824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 2426.442108] env[62824]: DEBUG nova.objects.instance [None req-01bd391a-ace2-4614-b82d-cfb73c68f7c9 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Lazy-loading 'flavor' on Instance uuid c70880c9-a69a-49b9-b028-189c04276623 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2426.447798] env[62824]: DEBUG nova.network.neutron [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Updating instance_info_cache with network_info: [{"id": "28514853-dbb7-4a89-999c-517a28061b9b", "address": "fa:16:3e:f1:72:8d", "network": {"id": "12411632-b7f6-4544-b2e4-b04ae7337547", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-917106830-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4479921caf5f405b8fc49baad390a0e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "de7fa486-5f28-44ae-b0cf-72234ff87546", "external-id": "nsx-vlan-transportzone-229", "segmentation_id": 229, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap28514853-db", "ovs_interfaceid": "28514853-dbb7-4a89-999c-517a28061b9b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2426.745250] env[62824]: DEBUG nova.network.neutron [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Port 7be95c5e-6e98-4fec-8550-1c22a787b95b binding to destination host cpu-1 is already ACTIVE {{(pid=62824) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 2426.948511] env[62824]: DEBUG oslo_concurrency.lockutils [None req-01bd391a-ace2-4614-b82d-cfb73c68f7c9 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Lock "c70880c9-a69a-49b9-b028-189c04276623" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.244s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2426.952141] env[62824]: DEBUG oslo_concurrency.lockutils [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Releasing lock "refresh_cache-df621c9c-8a68-4082-8f30-9bf3dda26fb6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2427.263261] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-128cbbf1-0973-4bf7-8fb2-3090e28cc3aa tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2427.263484] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-128cbbf1-0973-4bf7-8fb2-3090e28cc3aa tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412] Deleting contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2427.263664] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-128cbbf1-0973-4bf7-8fb2-3090e28cc3aa tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Deleting the datastore file [datastore1] c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2427.263922] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-66877614-7e84-4131-ac9b-b6878477a405 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2427.270717] env[62824]: DEBUG oslo_vmware.api [None req-128cbbf1-0973-4bf7-8fb2-3090e28cc3aa tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Waiting for the task: (returnval){ [ 2427.270717] env[62824]: value = "task-2146995" [ 2427.270717] env[62824]: _type = "Task" [ 2427.270717] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2427.278684] env[62824]: DEBUG oslo_vmware.api [None req-128cbbf1-0973-4bf7-8fb2-3090e28cc3aa tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Task: {'id': task-2146995, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2427.295537] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-60e3c005-4f57-43b7-8102-c73553b29f5e tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2427.295995] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-60e3c005-4f57-43b7-8102-c73553b29f5e tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Deleting contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2427.296338] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-60e3c005-4f57-43b7-8102-c73553b29f5e tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Deleting the datastore file [datastore2] 3d294ee9-92f6-4115-95db-9d4b66562b56 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2427.296404] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c882bfb5-3279-4b0b-89c7-908294c07692 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2427.307696] env[62824]: DEBUG oslo_vmware.api [None req-60e3c005-4f57-43b7-8102-c73553b29f5e tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2427.307696] env[62824]: value = "task-2146996" [ 2427.307696] env[62824]: _type = "Task" [ 2427.307696] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2427.319067] env[62824]: DEBUG oslo_vmware.api [None req-60e3c005-4f57-43b7-8102-c73553b29f5e tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146996, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2427.768773] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquiring lock "5b3df31f-15fe-473f-992c-ddb272661c53-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2427.769067] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lock "5b3df31f-15fe-473f-992c-ddb272661c53-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2427.769233] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lock "5b3df31f-15fe-473f-992c-ddb272661c53-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2427.780766] env[62824]: DEBUG oslo_vmware.api [None req-128cbbf1-0973-4bf7-8fb2-3090e28cc3aa tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Task: {'id': task-2146995, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.341622} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2427.781016] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-128cbbf1-0973-4bf7-8fb2-3090e28cc3aa tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2427.781527] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-128cbbf1-0973-4bf7-8fb2-3090e28cc3aa tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412] Deleted contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2427.781769] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-128cbbf1-0973-4bf7-8fb2-3090e28cc3aa tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2427.781998] env[62824]: INFO nova.compute.manager [None req-128cbbf1-0973-4bf7-8fb2-3090e28cc3aa tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] [instance: c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412] Took 6.50 seconds to destroy the instance on the hypervisor. [ 2427.782211] env[62824]: DEBUG oslo.service.loopingcall [None req-128cbbf1-0973-4bf7-8fb2-3090e28cc3aa tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2427.782984] env[62824]: DEBUG nova.compute.manager [-] [instance: c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2427.783342] env[62824]: DEBUG nova.network.neutron [-] [instance: c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2427.816722] env[62824]: DEBUG oslo_vmware.api [None req-60e3c005-4f57-43b7-8102-c73553b29f5e tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2146996, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.151043} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2427.817017] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-60e3c005-4f57-43b7-8102-c73553b29f5e tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2427.817214] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-60e3c005-4f57-43b7-8102-c73553b29f5e tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Deleted contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2427.817395] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-60e3c005-4f57-43b7-8102-c73553b29f5e tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2427.817566] env[62824]: INFO nova.compute.manager [None req-60e3c005-4f57-43b7-8102-c73553b29f5e tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Took 7.27 seconds to destroy the instance on the hypervisor. [ 2427.817803] env[62824]: DEBUG oslo.service.loopingcall [None req-60e3c005-4f57-43b7-8102-c73553b29f5e tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2427.817990] env[62824]: DEBUG nova.compute.manager [-] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2427.818101] env[62824]: DEBUG nova.network.neutron [-] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2428.468649] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1f6bec6-37bb-48b6-9e31-da50621a2505 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2428.490649] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Updating instance 'df621c9c-8a68-4082-8f30-9bf3dda26fb6' progress to 0 {{(pid=62824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 2428.543683] env[62824]: DEBUG oslo_concurrency.lockutils [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Acquiring lock "ad9f01dd-d719-41a2-971b-b12cd3e8bdf6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2428.543929] env[62824]: DEBUG oslo_concurrency.lockutils [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Lock "ad9f01dd-d719-41a2-971b-b12cd3e8bdf6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2428.795594] env[62824]: DEBUG nova.compute.manager [req-6319e0bc-aac9-4aa4-9920-bcfc26944364 req-b779bfb3-60be-45d8-b153-3e2c6b4c2b28 service nova] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Received event network-vif-deleted-f1800205-8a92-425e-9f64-1dabd5809221 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2428.795866] env[62824]: INFO nova.compute.manager [req-6319e0bc-aac9-4aa4-9920-bcfc26944364 req-b779bfb3-60be-45d8-b153-3e2c6b4c2b28 service nova] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Neutron deleted interface f1800205-8a92-425e-9f64-1dabd5809221; detaching it from the instance and deleting it from the info cache [ 2428.796018] env[62824]: DEBUG nova.network.neutron [req-6319e0bc-aac9-4aa4-9920-bcfc26944364 req-b779bfb3-60be-45d8-b153-3e2c6b4c2b28 service nova] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2428.804974] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquiring lock "refresh_cache-5b3df31f-15fe-473f-992c-ddb272661c53" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2428.805169] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquired lock "refresh_cache-5b3df31f-15fe-473f-992c-ddb272661c53" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2428.805351] env[62824]: DEBUG nova.network.neutron [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2429.001037] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2429.001366] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7e826fab-156c-4eda-93b5-19053f088362 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2429.009389] env[62824]: DEBUG oslo_vmware.api [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2429.009389] env[62824]: value = "task-2146997" [ 2429.009389] env[62824]: _type = "Task" [ 2429.009389] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2429.023875] env[62824]: DEBUG oslo_vmware.api [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146997, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2429.046721] env[62824]: DEBUG nova.compute.manager [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: ad9f01dd-d719-41a2-971b-b12cd3e8bdf6] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2429.238815] env[62824]: DEBUG nova.network.neutron [-] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2429.247599] env[62824]: DEBUG nova.network.neutron [-] [instance: c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2429.299760] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b890c648-85cc-429e-8bf3-77830b30d778 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2429.309592] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-817b1bef-921c-473d-a2c0-4f085c335588 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2429.339894] env[62824]: DEBUG nova.compute.manager [req-6319e0bc-aac9-4aa4-9920-bcfc26944364 req-b779bfb3-60be-45d8-b153-3e2c6b4c2b28 service nova] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Detach interface failed, port_id=f1800205-8a92-425e-9f64-1dabd5809221, reason: Instance 3d294ee9-92f6-4115-95db-9d4b66562b56 could not be found. {{(pid=62824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11482}} [ 2429.521600] env[62824]: DEBUG oslo_vmware.api [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146997, 'name': PowerOffVM_Task, 'duration_secs': 0.187505} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2429.521913] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2429.522132] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Updating instance 'df621c9c-8a68-4082-8f30-9bf3dda26fb6' progress to 17 {{(pid=62824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 2429.561403] env[62824]: DEBUG nova.network.neutron [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Updating instance_info_cache with network_info: [{"id": "7be95c5e-6e98-4fec-8550-1c22a787b95b", "address": "fa:16:3e:e8:69:3a", "network": {"id": "bba9b98c-c7eb-4890-8de5-eafacfdebd7d", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1928394728-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.160", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7fcdbfc29df64a1ba8d982bdcc667b64", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b49c5024-2ced-42ca-90cc-6066766d43e6", "external-id": "nsx-vlan-transportzone-239", "segmentation_id": 239, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7be95c5e-6e", "ovs_interfaceid": "7be95c5e-6e98-4fec-8550-1c22a787b95b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2429.574943] env[62824]: DEBUG oslo_concurrency.lockutils [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2429.575234] env[62824]: DEBUG oslo_concurrency.lockutils [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2429.576808] env[62824]: INFO nova.compute.claims [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: ad9f01dd-d719-41a2-971b-b12cd3e8bdf6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2429.740310] env[62824]: INFO nova.compute.manager [-] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Took 1.92 seconds to deallocate network for instance. [ 2429.750536] env[62824]: INFO nova.compute.manager [-] [instance: c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412] Took 1.97 seconds to deallocate network for instance. [ 2430.028265] env[62824]: DEBUG nova.virt.hardware [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:24Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2430.028689] env[62824]: DEBUG nova.virt.hardware [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2430.028689] env[62824]: DEBUG nova.virt.hardware [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2430.028960] env[62824]: DEBUG nova.virt.hardware [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2430.029111] env[62824]: DEBUG nova.virt.hardware [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2430.029174] env[62824]: DEBUG nova.virt.hardware [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2430.029399] env[62824]: DEBUG nova.virt.hardware [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2430.029656] env[62824]: DEBUG nova.virt.hardware [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2430.029848] env[62824]: DEBUG nova.virt.hardware [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2430.030038] env[62824]: DEBUG nova.virt.hardware [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2430.030219] env[62824]: DEBUG nova.virt.hardware [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2430.035443] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ddd8e88a-d754-4944-8a4a-b380e886108d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2430.051620] env[62824]: DEBUG oslo_vmware.api [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2430.051620] env[62824]: value = "task-2146998" [ 2430.051620] env[62824]: _type = "Task" [ 2430.051620] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2430.059469] env[62824]: DEBUG oslo_vmware.api [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146998, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2430.068049] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Releasing lock "refresh_cache-5b3df31f-15fe-473f-992c-ddb272661c53" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2430.246514] env[62824]: DEBUG oslo_concurrency.lockutils [None req-60e3c005-4f57-43b7-8102-c73553b29f5e tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2430.256015] env[62824]: DEBUG oslo_concurrency.lockutils [None req-128cbbf1-0973-4bf7-8fb2-3090e28cc3aa tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2430.561822] env[62824]: DEBUG oslo_vmware.api [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146998, 'name': ReconfigVM_Task, 'duration_secs': 0.19694} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2430.563080] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Updating instance 'df621c9c-8a68-4082-8f30-9bf3dda26fb6' progress to 33 {{(pid=62824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 2430.591016] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f4e5aee-fa42-4a66-a529-5b8b7390bffc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2430.613624] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43ce267a-3a44-4f2f-8a34-665651d8100c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2430.622018] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Updating instance '5b3df31f-15fe-473f-992c-ddb272661c53' progress to 83 {{(pid=62824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 2430.731361] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da254834-f934-43c1-af3b-94944a77663b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2430.738997] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97c1472a-2625-4622-8e4f-07e48684240b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2430.769282] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b721e7cb-f5a9-4af5-91ce-c78eba5d7a2c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2430.777303] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-323c9701-cdd3-4ef4-aab6-70bcb3496415 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2430.790436] env[62824]: DEBUG nova.compute.provider_tree [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2430.819852] env[62824]: DEBUG nova.compute.manager [req-c50db44b-ee3b-4023-9bf2-ace94683e7e5 req-43e1eb16-49af-4e50-a045-929a3054b81d service nova] [instance: c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412] Received event network-vif-deleted-0f67546a-a44f-4c20-96e8-e7b09334b221 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2431.068737] env[62824]: DEBUG nova.virt.hardware [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2431.069221] env[62824]: DEBUG nova.virt.hardware [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2431.069221] env[62824]: DEBUG nova.virt.hardware [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2431.069326] env[62824]: DEBUG nova.virt.hardware [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2431.069429] env[62824]: DEBUG nova.virt.hardware [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2431.069578] env[62824]: DEBUG nova.virt.hardware [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2431.069802] env[62824]: DEBUG nova.virt.hardware [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2431.069986] env[62824]: DEBUG nova.virt.hardware [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2431.070176] env[62824]: DEBUG nova.virt.hardware [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2431.070341] env[62824]: DEBUG nova.virt.hardware [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2431.070530] env[62824]: DEBUG nova.virt.hardware [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2431.075934] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Reconfiguring VM instance instance-00000070 to detach disk 2000 {{(pid=62824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 2431.076217] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7cb1722a-de35-4fde-aa42-a406c4026a30 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2431.095859] env[62824]: DEBUG oslo_vmware.api [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2431.095859] env[62824]: value = "task-2146999" [ 2431.095859] env[62824]: _type = "Task" [ 2431.095859] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2431.103804] env[62824]: DEBUG oslo_vmware.api [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146999, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2431.128996] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2431.129315] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6dca322a-8c78-4d1d-b864-97c5d9542c2f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2431.135581] env[62824]: DEBUG oslo_vmware.api [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2431.135581] env[62824]: value = "task-2147000" [ 2431.135581] env[62824]: _type = "Task" [ 2431.135581] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2431.143223] env[62824]: DEBUG oslo_vmware.api [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2147000, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2431.294289] env[62824]: DEBUG nova.scheduler.client.report [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2431.606035] env[62824]: DEBUG oslo_vmware.api [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2146999, 'name': ReconfigVM_Task, 'duration_secs': 0.156596} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2431.606335] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Reconfigured VM instance instance-00000070 to detach disk 2000 {{(pid=62824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 2431.607112] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78ef9f5e-1040-4912-a0ed-5b3c7f78722e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2431.628560] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Reconfiguring VM instance instance-00000070 to attach disk [datastore2] df621c9c-8a68-4082-8f30-9bf3dda26fb6/df621c9c-8a68-4082-8f30-9bf3dda26fb6.vmdk or device None with type thin {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2431.628866] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e5d60e82-83c5-44fe-b116-363a88b6b672 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2431.649710] env[62824]: DEBUG oslo_vmware.api [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2147000, 'name': PowerOnVM_Task, 'duration_secs': 0.365647} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2431.650883] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2431.651107] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5dc13737-f6e0-41bc-8bdf-e492a799e142 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Updating instance '5b3df31f-15fe-473f-992c-ddb272661c53' progress to 100 {{(pid=62824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 2431.654731] env[62824]: DEBUG oslo_vmware.api [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2431.654731] env[62824]: value = "task-2147001" [ 2431.654731] env[62824]: _type = "Task" [ 2431.654731] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2431.664049] env[62824]: DEBUG oslo_vmware.api [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2147001, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2431.799171] env[62824]: DEBUG oslo_concurrency.lockutils [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.224s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2431.799723] env[62824]: DEBUG nova.compute.manager [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: ad9f01dd-d719-41a2-971b-b12cd3e8bdf6] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2431.802467] env[62824]: DEBUG oslo_concurrency.lockutils [None req-60e3c005-4f57-43b7-8102-c73553b29f5e tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.556s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2431.802679] env[62824]: DEBUG nova.objects.instance [None req-60e3c005-4f57-43b7-8102-c73553b29f5e tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lazy-loading 'resources' on Instance uuid 3d294ee9-92f6-4115-95db-9d4b66562b56 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2432.167865] env[62824]: DEBUG oslo_vmware.api [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2147001, 'name': ReconfigVM_Task, 'duration_secs': 0.263131} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2432.168216] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Reconfigured VM instance instance-00000070 to attach disk [datastore2] df621c9c-8a68-4082-8f30-9bf3dda26fb6/df621c9c-8a68-4082-8f30-9bf3dda26fb6.vmdk or device None with type thin {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2432.168396] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Updating instance 'df621c9c-8a68-4082-8f30-9bf3dda26fb6' progress to 50 {{(pid=62824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 2432.306339] env[62824]: DEBUG nova.compute.utils [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2432.307577] env[62824]: DEBUG nova.compute.manager [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: ad9f01dd-d719-41a2-971b-b12cd3e8bdf6] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2432.307745] env[62824]: DEBUG nova.network.neutron [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: ad9f01dd-d719-41a2-971b-b12cd3e8bdf6] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2432.349427] env[62824]: DEBUG nova.policy [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cb36c701db8a471bbd65d2b68ce3becf', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c95aa87980de4ec8a2170c2640423301', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2432.437226] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-114dc94f-d877-45d3-aac9-c72a742a1543 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2432.446722] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afe27474-6238-4b87-b039-0c5e52edf550 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2432.478404] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82cef580-98a4-4d9a-b934-420be0ad73d5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2432.485726] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ef4c637-24ba-48a7-bbae-fda926995bcd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2432.500617] env[62824]: DEBUG nova.compute.provider_tree [None req-60e3c005-4f57-43b7-8102-c73553b29f5e tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2432.608376] env[62824]: DEBUG nova.network.neutron [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: ad9f01dd-d719-41a2-971b-b12cd3e8bdf6] Successfully created port: ffa86a2f-846f-4178-8ace-b502f5cb81bb {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2432.676021] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23834034-788e-4609-9c9a-78306fc9aef9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2432.697679] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16a96419-0ea8-4080-983e-d88e59d2baf1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2432.720734] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Updating instance 'df621c9c-8a68-4082-8f30-9bf3dda26fb6' progress to 67 {{(pid=62824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 2432.813970] env[62824]: DEBUG nova.compute.manager [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: ad9f01dd-d719-41a2-971b-b12cd3e8bdf6] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2433.004431] env[62824]: DEBUG nova.scheduler.client.report [None req-60e3c005-4f57-43b7-8102-c73553b29f5e tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2433.261816] env[62824]: DEBUG nova.network.neutron [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Port 28514853-dbb7-4a89-999c-517a28061b9b binding to destination host cpu-1 is already ACTIVE {{(pid=62824) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 2433.464920] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1da85f41-630f-4353-b59d-c2769516ce54 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquiring lock "5b3df31f-15fe-473f-992c-ddb272661c53" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2433.465203] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1da85f41-630f-4353-b59d-c2769516ce54 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lock "5b3df31f-15fe-473f-992c-ddb272661c53" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2433.465441] env[62824]: DEBUG nova.compute.manager [None req-1da85f41-630f-4353-b59d-c2769516ce54 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Going to confirm migration 4 {{(pid=62824) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5250}} [ 2433.508990] env[62824]: DEBUG oslo_concurrency.lockutils [None req-60e3c005-4f57-43b7-8102-c73553b29f5e tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.706s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2433.511412] env[62824]: DEBUG oslo_concurrency.lockutils [None req-128cbbf1-0973-4bf7-8fb2-3090e28cc3aa tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.255s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2433.511646] env[62824]: DEBUG nova.objects.instance [None req-128cbbf1-0973-4bf7-8fb2-3090e28cc3aa tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Lazy-loading 'resources' on Instance uuid c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2433.534600] env[62824]: INFO nova.scheduler.client.report [None req-60e3c005-4f57-43b7-8102-c73553b29f5e tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Deleted allocations for instance 3d294ee9-92f6-4115-95db-9d4b66562b56 [ 2433.824123] env[62824]: DEBUG nova.compute.manager [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: ad9f01dd-d719-41a2-971b-b12cd3e8bdf6] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2433.852718] env[62824]: DEBUG nova.virt.hardware [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2433.853046] env[62824]: DEBUG nova.virt.hardware [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2433.853167] env[62824]: DEBUG nova.virt.hardware [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2433.853355] env[62824]: DEBUG nova.virt.hardware [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2433.853501] env[62824]: DEBUG nova.virt.hardware [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2433.853650] env[62824]: DEBUG nova.virt.hardware [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2433.853861] env[62824]: DEBUG nova.virt.hardware [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2433.854169] env[62824]: DEBUG nova.virt.hardware [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2433.854552] env[62824]: DEBUG nova.virt.hardware [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2433.854552] env[62824]: DEBUG nova.virt.hardware [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2433.854661] env[62824]: DEBUG nova.virt.hardware [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2433.855520] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b55c74f-0e4e-4e26-b98d-eb9a8ef2b498 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2433.864129] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bc1f688-5878-4f29-a37e-f75b925983e2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2434.042654] env[62824]: DEBUG oslo_concurrency.lockutils [None req-60e3c005-4f57-43b7-8102-c73553b29f5e tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "3d294ee9-92f6-4115-95db-9d4b66562b56" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.998s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2434.067900] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1da85f41-630f-4353-b59d-c2769516ce54 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquiring lock "refresh_cache-5b3df31f-15fe-473f-992c-ddb272661c53" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2434.068087] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1da85f41-630f-4353-b59d-c2769516ce54 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquired lock "refresh_cache-5b3df31f-15fe-473f-992c-ddb272661c53" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2434.068264] env[62824]: DEBUG nova.network.neutron [None req-1da85f41-630f-4353-b59d-c2769516ce54 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2434.068442] env[62824]: DEBUG nova.objects.instance [None req-1da85f41-630f-4353-b59d-c2769516ce54 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lazy-loading 'info_cache' on Instance uuid 5b3df31f-15fe-473f-992c-ddb272661c53 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2434.113847] env[62824]: DEBUG nova.compute.manager [req-f97b0dd3-1656-4ef1-a1f7-ed6d8bbbe0e4 req-9d4f56ea-c2e3-41e1-b284-83fd7bfe8f81 service nova] [instance: ad9f01dd-d719-41a2-971b-b12cd3e8bdf6] Received event network-vif-plugged-ffa86a2f-846f-4178-8ace-b502f5cb81bb {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2434.114082] env[62824]: DEBUG oslo_concurrency.lockutils [req-f97b0dd3-1656-4ef1-a1f7-ed6d8bbbe0e4 req-9d4f56ea-c2e3-41e1-b284-83fd7bfe8f81 service nova] Acquiring lock "ad9f01dd-d719-41a2-971b-b12cd3e8bdf6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2434.114331] env[62824]: DEBUG oslo_concurrency.lockutils [req-f97b0dd3-1656-4ef1-a1f7-ed6d8bbbe0e4 req-9d4f56ea-c2e3-41e1-b284-83fd7bfe8f81 service nova] Lock "ad9f01dd-d719-41a2-971b-b12cd3e8bdf6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2434.114502] env[62824]: DEBUG oslo_concurrency.lockutils [req-f97b0dd3-1656-4ef1-a1f7-ed6d8bbbe0e4 req-9d4f56ea-c2e3-41e1-b284-83fd7bfe8f81 service nova] Lock "ad9f01dd-d719-41a2-971b-b12cd3e8bdf6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2434.114666] env[62824]: DEBUG nova.compute.manager [req-f97b0dd3-1656-4ef1-a1f7-ed6d8bbbe0e4 req-9d4f56ea-c2e3-41e1-b284-83fd7bfe8f81 service nova] [instance: ad9f01dd-d719-41a2-971b-b12cd3e8bdf6] No waiting events found dispatching network-vif-plugged-ffa86a2f-846f-4178-8ace-b502f5cb81bb {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2434.114824] env[62824]: WARNING nova.compute.manager [req-f97b0dd3-1656-4ef1-a1f7-ed6d8bbbe0e4 req-9d4f56ea-c2e3-41e1-b284-83fd7bfe8f81 service nova] [instance: ad9f01dd-d719-41a2-971b-b12cd3e8bdf6] Received unexpected event network-vif-plugged-ffa86a2f-846f-4178-8ace-b502f5cb81bb for instance with vm_state building and task_state spawning. [ 2434.151449] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d65d3d38-8d64-4528-bcbc-9e9252f52110 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2434.159184] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f857e6b-91ff-4824-a8ab-5d7917ced7ba {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2434.193654] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e98ca34-e30b-4f7f-861d-1924fcd6c0b9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2434.202118] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89dbcee2-3eec-4db6-b843-38fdedfc29ee {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2434.223365] env[62824]: DEBUG nova.compute.provider_tree [None req-128cbbf1-0973-4bf7-8fb2-3090e28cc3aa tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2434.225327] env[62824]: DEBUG nova.network.neutron [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: ad9f01dd-d719-41a2-971b-b12cd3e8bdf6] Successfully updated port: ffa86a2f-846f-4178-8ace-b502f5cb81bb {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2434.284860] env[62824]: DEBUG oslo_concurrency.lockutils [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquiring lock "df621c9c-8a68-4082-8f30-9bf3dda26fb6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2434.285144] env[62824]: DEBUG oslo_concurrency.lockutils [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "df621c9c-8a68-4082-8f30-9bf3dda26fb6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2434.285375] env[62824]: DEBUG oslo_concurrency.lockutils [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "df621c9c-8a68-4082-8f30-9bf3dda26fb6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2434.728708] env[62824]: DEBUG nova.scheduler.client.report [None req-128cbbf1-0973-4bf7-8fb2-3090e28cc3aa tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2434.735605] env[62824]: DEBUG oslo_concurrency.lockutils [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Acquiring lock "refresh_cache-ad9f01dd-d719-41a2-971b-b12cd3e8bdf6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2434.735853] env[62824]: DEBUG oslo_concurrency.lockutils [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Acquired lock "refresh_cache-ad9f01dd-d719-41a2-971b-b12cd3e8bdf6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2434.736266] env[62824]: DEBUG nova.network.neutron [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: ad9f01dd-d719-41a2-971b-b12cd3e8bdf6] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2435.238664] env[62824]: DEBUG oslo_concurrency.lockutils [None req-128cbbf1-0973-4bf7-8fb2-3090e28cc3aa tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.727s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2435.293462] env[62824]: INFO nova.scheduler.client.report [None req-128cbbf1-0973-4bf7-8fb2-3090e28cc3aa tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Deleted allocations for instance c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412 [ 2435.307326] env[62824]: DEBUG nova.network.neutron [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: ad9f01dd-d719-41a2-971b-b12cd3e8bdf6] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2435.371139] env[62824]: DEBUG oslo_concurrency.lockutils [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquiring lock "refresh_cache-df621c9c-8a68-4082-8f30-9bf3dda26fb6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2435.371318] env[62824]: DEBUG oslo_concurrency.lockutils [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquired lock "refresh_cache-df621c9c-8a68-4082-8f30-9bf3dda26fb6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2435.371490] env[62824]: DEBUG nova.network.neutron [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2435.422580] env[62824]: DEBUG nova.network.neutron [None req-1da85f41-630f-4353-b59d-c2769516ce54 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Updating instance_info_cache with network_info: [{"id": "7be95c5e-6e98-4fec-8550-1c22a787b95b", "address": "fa:16:3e:e8:69:3a", "network": {"id": "bba9b98c-c7eb-4890-8de5-eafacfdebd7d", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1928394728-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.160", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7fcdbfc29df64a1ba8d982bdcc667b64", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b49c5024-2ced-42ca-90cc-6066766d43e6", "external-id": "nsx-vlan-transportzone-239", "segmentation_id": 239, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7be95c5e-6e", "ovs_interfaceid": "7be95c5e-6e98-4fec-8550-1c22a787b95b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2435.447509] env[62824]: DEBUG oslo_concurrency.lockutils [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquiring lock "4dff2a76-b0a2-438c-b52f-d6495948d041" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2435.447826] env[62824]: DEBUG oslo_concurrency.lockutils [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "4dff2a76-b0a2-438c-b52f-d6495948d041" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2435.501885] env[62824]: DEBUG nova.network.neutron [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: ad9f01dd-d719-41a2-971b-b12cd3e8bdf6] Updating instance_info_cache with network_info: [{"id": "ffa86a2f-846f-4178-8ace-b502f5cb81bb", "address": "fa:16:3e:d3:92:a1", "network": {"id": "f7936d74-eeef-4c41-8ae7-9427c43663fa", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1643883257-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c95aa87980de4ec8a2170c2640423301", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6eaa481-1f92-4851-b98e-09ed0daad7cb", "external-id": "nsx-vlan-transportzone-636", "segmentation_id": 636, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapffa86a2f-84", "ovs_interfaceid": "ffa86a2f-846f-4178-8ace-b502f5cb81bb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2435.804289] env[62824]: DEBUG oslo_concurrency.lockutils [None req-128cbbf1-0973-4bf7-8fb2-3090e28cc3aa tempest-AttachVolumeTestJSON-1894392206 tempest-AttachVolumeTestJSON-1894392206-project-member] Lock "c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.031s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2435.924946] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1da85f41-630f-4353-b59d-c2769516ce54 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Releasing lock "refresh_cache-5b3df31f-15fe-473f-992c-ddb272661c53" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2435.925243] env[62824]: DEBUG nova.objects.instance [None req-1da85f41-630f-4353-b59d-c2769516ce54 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lazy-loading 'migration_context' on Instance uuid 5b3df31f-15fe-473f-992c-ddb272661c53 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2435.949762] env[62824]: DEBUG nova.compute.manager [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 4dff2a76-b0a2-438c-b52f-d6495948d041] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2436.004209] env[62824]: DEBUG oslo_concurrency.lockutils [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Releasing lock "refresh_cache-ad9f01dd-d719-41a2-971b-b12cd3e8bdf6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2436.004534] env[62824]: DEBUG nova.compute.manager [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: ad9f01dd-d719-41a2-971b-b12cd3e8bdf6] Instance network_info: |[{"id": "ffa86a2f-846f-4178-8ace-b502f5cb81bb", "address": "fa:16:3e:d3:92:a1", "network": {"id": "f7936d74-eeef-4c41-8ae7-9427c43663fa", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1643883257-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c95aa87980de4ec8a2170c2640423301", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6eaa481-1f92-4851-b98e-09ed0daad7cb", "external-id": "nsx-vlan-transportzone-636", "segmentation_id": 636, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapffa86a2f-84", "ovs_interfaceid": "ffa86a2f-846f-4178-8ace-b502f5cb81bb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2436.005015] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: ad9f01dd-d719-41a2-971b-b12cd3e8bdf6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d3:92:a1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c6eaa481-1f92-4851-b98e-09ed0daad7cb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ffa86a2f-846f-4178-8ace-b502f5cb81bb', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2436.012572] env[62824]: DEBUG oslo.service.loopingcall [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2436.012778] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ad9f01dd-d719-41a2-971b-b12cd3e8bdf6] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2436.013044] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5e45c775-870f-43a2-bfba-551cb12699d4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2436.036556] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2436.036556] env[62824]: value = "task-2147002" [ 2436.036556] env[62824]: _type = "Task" [ 2436.036556] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2436.044546] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2147002, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2436.129540] env[62824]: DEBUG nova.network.neutron [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Updating instance_info_cache with network_info: [{"id": "28514853-dbb7-4a89-999c-517a28061b9b", "address": "fa:16:3e:f1:72:8d", "network": {"id": "12411632-b7f6-4544-b2e4-b04ae7337547", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-917106830-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4479921caf5f405b8fc49baad390a0e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "de7fa486-5f28-44ae-b0cf-72234ff87546", "external-id": "nsx-vlan-transportzone-229", "segmentation_id": 229, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap28514853-db", "ovs_interfaceid": "28514853-dbb7-4a89-999c-517a28061b9b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2436.190164] env[62824]: DEBUG nova.compute.manager [req-78106f07-372e-48db-8959-e037edc35423 req-0413d94f-cab4-4900-bbc6-a810aff9c250 service nova] [instance: ad9f01dd-d719-41a2-971b-b12cd3e8bdf6] Received event network-changed-ffa86a2f-846f-4178-8ace-b502f5cb81bb {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2436.190449] env[62824]: DEBUG nova.compute.manager [req-78106f07-372e-48db-8959-e037edc35423 req-0413d94f-cab4-4900-bbc6-a810aff9c250 service nova] [instance: ad9f01dd-d719-41a2-971b-b12cd3e8bdf6] Refreshing instance network info cache due to event network-changed-ffa86a2f-846f-4178-8ace-b502f5cb81bb. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2436.190686] env[62824]: DEBUG oslo_concurrency.lockutils [req-78106f07-372e-48db-8959-e037edc35423 req-0413d94f-cab4-4900-bbc6-a810aff9c250 service nova] Acquiring lock "refresh_cache-ad9f01dd-d719-41a2-971b-b12cd3e8bdf6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2436.190830] env[62824]: DEBUG oslo_concurrency.lockutils [req-78106f07-372e-48db-8959-e037edc35423 req-0413d94f-cab4-4900-bbc6-a810aff9c250 service nova] Acquired lock "refresh_cache-ad9f01dd-d719-41a2-971b-b12cd3e8bdf6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2436.190992] env[62824]: DEBUG nova.network.neutron [req-78106f07-372e-48db-8959-e037edc35423 req-0413d94f-cab4-4900-bbc6-a810aff9c250 service nova] [instance: ad9f01dd-d719-41a2-971b-b12cd3e8bdf6] Refreshing network info cache for port ffa86a2f-846f-4178-8ace-b502f5cb81bb {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2436.428267] env[62824]: DEBUG nova.objects.base [None req-1da85f41-630f-4353-b59d-c2769516ce54 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Object Instance<5b3df31f-15fe-473f-992c-ddb272661c53> lazy-loaded attributes: info_cache,migration_context {{(pid=62824) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 2436.429315] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bf9c9e4-fcdd-4adc-a659-7d4d61aee445 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2436.450685] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c9b6192d-f988-4ac8-b672-f5f85b968caf {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2436.460806] env[62824]: DEBUG oslo_vmware.api [None req-1da85f41-630f-4353-b59d-c2769516ce54 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2436.460806] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52966ddd-c2a5-e9bf-3557-b4423e93bdb9" [ 2436.460806] env[62824]: _type = "Task" [ 2436.460806] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2436.473071] env[62824]: DEBUG oslo_vmware.api [None req-1da85f41-630f-4353-b59d-c2769516ce54 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52966ddd-c2a5-e9bf-3557-b4423e93bdb9, 'name': SearchDatastore_Task, 'duration_secs': 0.009304} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2436.473395] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1da85f41-630f-4353-b59d-c2769516ce54 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2436.474036] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1da85f41-630f-4353-b59d-c2769516ce54 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2436.478607] env[62824]: DEBUG oslo_concurrency.lockutils [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2436.545657] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2147002, 'name': CreateVM_Task, 'duration_secs': 0.492367} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2436.545824] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ad9f01dd-d719-41a2-971b-b12cd3e8bdf6] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2436.546513] env[62824]: DEBUG oslo_concurrency.lockutils [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2436.546679] env[62824]: DEBUG oslo_concurrency.lockutils [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2436.547014] env[62824]: DEBUG oslo_concurrency.lockutils [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2436.547266] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-524d9e8a-4f56-4757-a36c-1a228dd9a104 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2436.552241] env[62824]: DEBUG oslo_vmware.api [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Waiting for the task: (returnval){ [ 2436.552241] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52404b09-4ad2-1030-7418-3bd5edc78281" [ 2436.552241] env[62824]: _type = "Task" [ 2436.552241] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2436.560970] env[62824]: DEBUG oslo_vmware.api [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52404b09-4ad2-1030-7418-3bd5edc78281, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2436.633578] env[62824]: DEBUG oslo_concurrency.lockutils [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Releasing lock "refresh_cache-df621c9c-8a68-4082-8f30-9bf3dda26fb6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2436.872110] env[62824]: DEBUG nova.network.neutron [req-78106f07-372e-48db-8959-e037edc35423 req-0413d94f-cab4-4900-bbc6-a810aff9c250 service nova] [instance: ad9f01dd-d719-41a2-971b-b12cd3e8bdf6] Updated VIF entry in instance network info cache for port ffa86a2f-846f-4178-8ace-b502f5cb81bb. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2436.872533] env[62824]: DEBUG nova.network.neutron [req-78106f07-372e-48db-8959-e037edc35423 req-0413d94f-cab4-4900-bbc6-a810aff9c250 service nova] [instance: ad9f01dd-d719-41a2-971b-b12cd3e8bdf6] Updating instance_info_cache with network_info: [{"id": "ffa86a2f-846f-4178-8ace-b502f5cb81bb", "address": "fa:16:3e:d3:92:a1", "network": {"id": "f7936d74-eeef-4c41-8ae7-9427c43663fa", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1643883257-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c95aa87980de4ec8a2170c2640423301", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6eaa481-1f92-4851-b98e-09ed0daad7cb", "external-id": "nsx-vlan-transportzone-636", "segmentation_id": 636, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapffa86a2f-84", "ovs_interfaceid": "ffa86a2f-846f-4178-8ace-b502f5cb81bb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2436.961550] env[62824]: DEBUG nova.compute.manager [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2436.962542] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ba657fa-6651-4aca-a6af-d914c3b6ca76 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2437.061745] env[62824]: DEBUG oslo_vmware.api [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52404b09-4ad2-1030-7418-3bd5edc78281, 'name': SearchDatastore_Task, 'duration_secs': 0.009444} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2437.063930] env[62824]: DEBUG oslo_concurrency.lockutils [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2437.064200] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: ad9f01dd-d719-41a2-971b-b12cd3e8bdf6] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2437.064446] env[62824]: DEBUG oslo_concurrency.lockutils [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2437.064595] env[62824]: DEBUG oslo_concurrency.lockutils [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2437.064773] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2437.065191] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e4484082-6045-4010-98e8-047b3d522033 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2437.075100] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2437.075315] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2437.075990] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ca398e29-af00-4140-9968-d83565dfbbab {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2437.081301] env[62824]: DEBUG oslo_vmware.api [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Waiting for the task: (returnval){ [ 2437.081301] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52f5dcca-c9ef-c189-a99e-c8501912464e" [ 2437.081301] env[62824]: _type = "Task" [ 2437.081301] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2437.085715] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c73ee98-53d7-490a-8877-be39e1788f2b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2437.092529] env[62824]: DEBUG oslo_vmware.api [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52f5dcca-c9ef-c189-a99e-c8501912464e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2437.094991] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-845e0140-8057-44bf-9f40-92ec71f32b59 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2437.125780] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3131e17b-5e08-4ff7-83d8-a0bd2108ef43 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2437.133202] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eddf89ec-14a3-4cbb-9ea5-b62cc667faf1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2437.149331] env[62824]: DEBUG nova.compute.provider_tree [None req-1da85f41-630f-4353-b59d-c2769516ce54 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2437.163222] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe6f6966-ea2f-4ae0-b684-f5e358578b90 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2437.181466] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7d0820e-72b2-45b6-8b78-4c8cce33b4f8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2437.187875] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Updating instance 'df621c9c-8a68-4082-8f30-9bf3dda26fb6' progress to 83 {{(pid=62824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 2437.376286] env[62824]: DEBUG oslo_concurrency.lockutils [req-78106f07-372e-48db-8959-e037edc35423 req-0413d94f-cab4-4900-bbc6-a810aff9c250 service nova] Releasing lock "refresh_cache-ad9f01dd-d719-41a2-971b-b12cd3e8bdf6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2437.472817] env[62824]: INFO nova.compute.manager [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] instance snapshotting [ 2437.473495] env[62824]: DEBUG nova.objects.instance [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lazy-loading 'flavor' on Instance uuid 7c47bb92-a58e-4537-a68a-289e2a5fc19a {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2437.591589] env[62824]: DEBUG oslo_vmware.api [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52f5dcca-c9ef-c189-a99e-c8501912464e, 'name': SearchDatastore_Task, 'duration_secs': 0.013782} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2437.592400] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7a23702a-a320-4c58-94ea-f3fe004f0077 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2437.597641] env[62824]: DEBUG oslo_vmware.api [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Waiting for the task: (returnval){ [ 2437.597641] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5239111c-1578-3418-a2b3-9ead273a6e31" [ 2437.597641] env[62824]: _type = "Task" [ 2437.597641] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2437.605724] env[62824]: DEBUG oslo_vmware.api [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5239111c-1578-3418-a2b3-9ead273a6e31, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2437.653856] env[62824]: DEBUG nova.scheduler.client.report [None req-1da85f41-630f-4353-b59d-c2769516ce54 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2437.693410] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2437.693703] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8e1a4f7d-593a-4094-86a9-a1d17d84274f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2437.701618] env[62824]: DEBUG oslo_vmware.api [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2437.701618] env[62824]: value = "task-2147004" [ 2437.701618] env[62824]: _type = "Task" [ 2437.701618] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2437.709369] env[62824]: DEBUG oslo_vmware.api [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2147004, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2437.772862] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2437.773315] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager.update_available_resource {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2437.979671] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8d5fa14-e540-4af7-8980-cf5c46bf89cc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2437.999048] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75fcd106-7350-4b1c-a015-44af036a5179 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2438.110755] env[62824]: DEBUG oslo_vmware.api [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5239111c-1578-3418-a2b3-9ead273a6e31, 'name': SearchDatastore_Task, 'duration_secs': 0.01163} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2438.111097] env[62824]: DEBUG oslo_concurrency.lockutils [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2438.111374] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] ad9f01dd-d719-41a2-971b-b12cd3e8bdf6/ad9f01dd-d719-41a2-971b-b12cd3e8bdf6.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2438.111637] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a3683e21-ead6-4c5a-94e8-f82ad6717a85 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2438.119471] env[62824]: DEBUG oslo_vmware.api [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Waiting for the task: (returnval){ [ 2438.119471] env[62824]: value = "task-2147005" [ 2438.119471] env[62824]: _type = "Task" [ 2438.119471] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2438.127266] env[62824]: DEBUG oslo_vmware.api [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147005, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2438.211353] env[62824]: DEBUG oslo_vmware.api [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2147004, 'name': PowerOnVM_Task, 'duration_secs': 0.459712} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2438.211624] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2438.211804] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-687d0e46-d363-41dc-9c2b-736227659bc4 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Updating instance 'df621c9c-8a68-4082-8f30-9bf3dda26fb6' progress to 100 {{(pid=62824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 2438.276380] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2438.509365] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Creating Snapshot of the VM instance {{(pid=62824) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 2438.509713] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-572e4e10-fb11-466c-b29b-d844ff14678b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2438.518028] env[62824]: DEBUG oslo_vmware.api [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2438.518028] env[62824]: value = "task-2147006" [ 2438.518028] env[62824]: _type = "Task" [ 2438.518028] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2438.527393] env[62824]: DEBUG oslo_vmware.api [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147006, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2438.629528] env[62824]: DEBUG oslo_vmware.api [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147005, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.457609} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2438.629798] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] ad9f01dd-d719-41a2-971b-b12cd3e8bdf6/ad9f01dd-d719-41a2-971b-b12cd3e8bdf6.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2438.630111] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: ad9f01dd-d719-41a2-971b-b12cd3e8bdf6] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2438.630391] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8050c455-f3d1-402d-831e-9c5518508d55 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2438.637505] env[62824]: DEBUG oslo_vmware.api [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Waiting for the task: (returnval){ [ 2438.637505] env[62824]: value = "task-2147007" [ 2438.637505] env[62824]: _type = "Task" [ 2438.637505] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2438.644964] env[62824]: DEBUG oslo_vmware.api [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147007, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2438.663501] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1da85f41-630f-4353-b59d-c2769516ce54 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.190s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2438.666378] env[62824]: DEBUG oslo_concurrency.lockutils [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.188s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2438.667874] env[62824]: INFO nova.compute.claims [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 4dff2a76-b0a2-438c-b52f-d6495948d041] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2439.027852] env[62824]: DEBUG oslo_vmware.api [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147006, 'name': CreateSnapshot_Task, 'duration_secs': 0.459864} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2439.028182] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Created Snapshot of the VM instance {{(pid=62824) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 2439.028907] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0a0cb03-cbaa-4d44-8009-5e8fb46e2999 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2439.147447] env[62824]: DEBUG oslo_vmware.api [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147007, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072545} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2439.147727] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: ad9f01dd-d719-41a2-971b-b12cd3e8bdf6] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2439.148504] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bfa8828-8c0d-4f59-871c-af66f27f43ce {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2439.169542] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: ad9f01dd-d719-41a2-971b-b12cd3e8bdf6] Reconfiguring VM instance instance-00000071 to attach disk [datastore1] ad9f01dd-d719-41a2-971b-b12cd3e8bdf6/ad9f01dd-d719-41a2-971b-b12cd3e8bdf6.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2439.169784] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a5bdc1f2-9454-42df-a604-21ffbfc34c52 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2439.196364] env[62824]: DEBUG oslo_vmware.api [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Waiting for the task: (returnval){ [ 2439.196364] env[62824]: value = "task-2147008" [ 2439.196364] env[62824]: _type = "Task" [ 2439.196364] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2439.205647] env[62824]: DEBUG oslo_vmware.api [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147008, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2439.243371] env[62824]: INFO nova.scheduler.client.report [None req-1da85f41-630f-4353-b59d-c2769516ce54 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Deleted allocation for migration a4f6b1ba-8f40-4ece-af3a-39164086942a [ 2439.546855] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Creating linked-clone VM from snapshot {{(pid=62824) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 2439.547564] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-0a7a3ad9-f42f-4003-96b7-9293c1c57130 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2439.560967] env[62824]: DEBUG oslo_vmware.api [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2439.560967] env[62824]: value = "task-2147009" [ 2439.560967] env[62824]: _type = "Task" [ 2439.560967] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2439.569851] env[62824]: DEBUG oslo_vmware.api [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147009, 'name': CloneVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2439.709626] env[62824]: DEBUG oslo_vmware.api [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147008, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2439.750838] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1da85f41-630f-4353-b59d-c2769516ce54 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lock "5b3df31f-15fe-473f-992c-ddb272661c53" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.285s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2439.831337] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-426511d6-cd00-4e39-8789-cbfd7198fde7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2439.839351] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ed6057c-2b11-441e-918e-08edd0f01acc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2439.871250] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acee6360-9bee-4189-98f7-69c268f737d4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2439.878964] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f8bbfd1-447e-4a63-b734-30ea9d894a11 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2440.649462] env[62824]: DEBUG oslo_concurrency.lockutils [None req-513f1eb6-422f-45f6-8be2-f708668a058a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquiring lock "df621c9c-8a68-4082-8f30-9bf3dda26fb6" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2440.649783] env[62824]: DEBUG oslo_concurrency.lockutils [None req-513f1eb6-422f-45f6-8be2-f708668a058a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "df621c9c-8a68-4082-8f30-9bf3dda26fb6" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2440.649904] env[62824]: DEBUG nova.compute.manager [None req-513f1eb6-422f-45f6-8be2-f708668a058a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Going to confirm migration 5 {{(pid=62824) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5250}} [ 2440.671049] env[62824]: DEBUG nova.compute.provider_tree [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2440.679042] env[62824]: DEBUG oslo_vmware.api [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147009, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2440.679590] env[62824]: DEBUG oslo_vmware.api [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147008, 'name': ReconfigVM_Task, 'duration_secs': 0.749609} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2440.680550] env[62824]: DEBUG nova.scheduler.client.report [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2440.683552] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: ad9f01dd-d719-41a2-971b-b12cd3e8bdf6] Reconfigured VM instance instance-00000071 to attach disk [datastore1] ad9f01dd-d719-41a2-971b-b12cd3e8bdf6/ad9f01dd-d719-41a2-971b-b12cd3e8bdf6.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2440.686826] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4f931e72-a4a3-44ee-9cf8-36ab626290c2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2440.691786] env[62824]: DEBUG oslo_vmware.api [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Waiting for the task: (returnval){ [ 2440.691786] env[62824]: value = "task-2147010" [ 2440.691786] env[62824]: _type = "Task" [ 2440.691786] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2440.700691] env[62824]: DEBUG oslo_vmware.api [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147010, 'name': Rename_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2441.162321] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a7e061cc-c794-4218-9571-b7846b71f5eb tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquiring lock "5b3df31f-15fe-473f-992c-ddb272661c53" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2441.162718] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a7e061cc-c794-4218-9571-b7846b71f5eb tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lock "5b3df31f-15fe-473f-992c-ddb272661c53" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2441.162960] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a7e061cc-c794-4218-9571-b7846b71f5eb tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquiring lock "5b3df31f-15fe-473f-992c-ddb272661c53-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2441.163224] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a7e061cc-c794-4218-9571-b7846b71f5eb tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lock "5b3df31f-15fe-473f-992c-ddb272661c53-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2441.163420] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a7e061cc-c794-4218-9571-b7846b71f5eb tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lock "5b3df31f-15fe-473f-992c-ddb272661c53-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2441.168461] env[62824]: DEBUG oslo_vmware.api [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147009, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2441.170455] env[62824]: INFO nova.compute.manager [None req-a7e061cc-c794-4218-9571-b7846b71f5eb tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Terminating instance [ 2441.186660] env[62824]: DEBUG oslo_concurrency.lockutils [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.520s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2441.187190] env[62824]: DEBUG nova.compute.manager [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 4dff2a76-b0a2-438c-b52f-d6495948d041] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2441.189745] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 2.914s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2441.189958] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2441.190186] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62824) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 2441.191453] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea17c193-d60f-4ffc-ac99-1641724c28d9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2441.206175] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0051c24-3129-4448-bd1c-e0c599226116 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2441.210058] env[62824]: DEBUG oslo_vmware.api [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147010, 'name': Rename_Task, 'duration_secs': 0.276537} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2441.210668] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: ad9f01dd-d719-41a2-971b-b12cd3e8bdf6] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2441.211235] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-66713286-6127-407e-8630-ec5402a46e95 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2441.221370] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-114c3151-c431-4603-a8db-86c4d96dc216 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2441.224934] env[62824]: DEBUG oslo_vmware.api [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Waiting for the task: (returnval){ [ 2441.224934] env[62824]: value = "task-2147011" [ 2441.224934] env[62824]: _type = "Task" [ 2441.224934] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2441.231840] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5822982-2f55-4796-8c47-ac0dab958f8b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2441.239660] env[62824]: DEBUG oslo_vmware.api [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147011, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2441.240775] env[62824]: DEBUG oslo_concurrency.lockutils [None req-513f1eb6-422f-45f6-8be2-f708668a058a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquiring lock "refresh_cache-df621c9c-8a68-4082-8f30-9bf3dda26fb6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2441.240939] env[62824]: DEBUG oslo_concurrency.lockutils [None req-513f1eb6-422f-45f6-8be2-f708668a058a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquired lock "refresh_cache-df621c9c-8a68-4082-8f30-9bf3dda26fb6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2441.241129] env[62824]: DEBUG nova.network.neutron [None req-513f1eb6-422f-45f6-8be2-f708668a058a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2441.241312] env[62824]: DEBUG nova.objects.instance [None req-513f1eb6-422f-45f6-8be2-f708668a058a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lazy-loading 'info_cache' on Instance uuid df621c9c-8a68-4082-8f30-9bf3dda26fb6 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2441.271740] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179884MB free_disk=175GB free_vcpus=48 pci_devices=None {{(pid=62824) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 2441.272037] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2441.272367] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2441.661488] env[62824]: DEBUG oslo_vmware.api [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147009, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2441.677393] env[62824]: DEBUG nova.compute.manager [None req-a7e061cc-c794-4218-9571-b7846b71f5eb tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2441.677609] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-a7e061cc-c794-4218-9571-b7846b71f5eb tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2441.678382] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9db40ed1-3ee9-446f-b402-c265957d4d5c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2441.685034] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7e061cc-c794-4218-9571-b7846b71f5eb tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2441.685268] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f8a94e8b-03c0-4ffd-8b7d-52eede5ed3c0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2441.691544] env[62824]: DEBUG oslo_vmware.api [None req-a7e061cc-c794-4218-9571-b7846b71f5eb tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2441.691544] env[62824]: value = "task-2147012" [ 2441.691544] env[62824]: _type = "Task" [ 2441.691544] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2441.695925] env[62824]: DEBUG nova.compute.utils [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2441.699828] env[62824]: DEBUG nova.compute.manager [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 4dff2a76-b0a2-438c-b52f-d6495948d041] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2441.699957] env[62824]: DEBUG nova.network.neutron [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 4dff2a76-b0a2-438c-b52f-d6495948d041] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2441.701650] env[62824]: DEBUG oslo_vmware.api [None req-a7e061cc-c794-4218-9571-b7846b71f5eb tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2147012, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2441.735207] env[62824]: DEBUG oslo_vmware.api [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147011, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2441.740698] env[62824]: DEBUG nova.policy [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9cc01135021d4a90a718dc605ce98af2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '815ecf9340934733a83d4dcb26612fd4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2441.989847] env[62824]: DEBUG nova.network.neutron [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 4dff2a76-b0a2-438c-b52f-d6495948d041] Successfully created port: 2e0a71bd-c461-4fe9-a810-fe72908cf629 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2442.162555] env[62824]: DEBUG oslo_vmware.api [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147009, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2442.201072] env[62824]: DEBUG oslo_vmware.api [None req-a7e061cc-c794-4218-9571-b7846b71f5eb tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2147012, 'name': PowerOffVM_Task, 'duration_secs': 0.205218} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2442.201365] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7e061cc-c794-4218-9571-b7846b71f5eb tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2442.201563] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-a7e061cc-c794-4218-9571-b7846b71f5eb tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2442.201831] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-439d15f2-1f6b-44fa-bc0a-23b7e2aff60a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2442.204064] env[62824]: DEBUG nova.compute.manager [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 4dff2a76-b0a2-438c-b52f-d6495948d041] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2442.236850] env[62824]: DEBUG oslo_vmware.api [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147011, 'name': PowerOnVM_Task, 'duration_secs': 0.900921} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2442.237160] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: ad9f01dd-d719-41a2-971b-b12cd3e8bdf6] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2442.237374] env[62824]: INFO nova.compute.manager [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: ad9f01dd-d719-41a2-971b-b12cd3e8bdf6] Took 8.41 seconds to spawn the instance on the hypervisor. [ 2442.237547] env[62824]: DEBUG nova.compute.manager [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: ad9f01dd-d719-41a2-971b-b12cd3e8bdf6] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2442.238368] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc9612c4-e80a-48e4-9820-7d62fe7bd57a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2442.282093] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Applying migration context for instance df621c9c-8a68-4082-8f30-9bf3dda26fb6 as it has an incoming, in-progress migration a132a1b4-4c08-4cae-bc8c-af84308301aa. Migration status is confirming {{(pid=62824) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 2442.283328] env[62824]: INFO nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Updating resource usage from migration a132a1b4-4c08-4cae-bc8c-af84308301aa [ 2442.307792] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance ca42fef0-1e90-4ab6-9d60-8ef7e4997884 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2442.307792] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance c70880c9-a69a-49b9-b028-189c04276623 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2442.307792] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 7c47bb92-a58e-4537-a68a-289e2a5fc19a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2442.307792] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 5b3df31f-15fe-473f-992c-ddb272661c53 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2442.307792] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Migration a132a1b4-4c08-4cae-bc8c-af84308301aa is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 2442.307792] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance df621c9c-8a68-4082-8f30-9bf3dda26fb6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2442.307792] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance ad9f01dd-d719-41a2-971b-b12cd3e8bdf6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2442.307792] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 4dff2a76-b0a2-438c-b52f-d6495948d041 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2442.307792] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Total usable vcpus: 48, total allocated vcpus: 8 {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 2442.307792] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2176MB phys_disk=200GB used_disk=8GB total_vcpus=48 used_vcpus=8 pci_stats=[] {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 2442.408685] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36f71211-a4c9-497a-bebc-ddadf1524e94 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2442.415972] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7c6bf46-aa31-481a-81aa-5aeda19eb164 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2442.446918] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7d86fcc-1ff5-4d7b-8376-e7ba12969bc0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2442.453628] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-533f6302-6ab2-4ebe-b91d-0a6939047ad6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2442.466290] env[62824]: DEBUG nova.compute.provider_tree [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2442.477201] env[62824]: DEBUG nova.network.neutron [None req-513f1eb6-422f-45f6-8be2-f708668a058a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Updating instance_info_cache with network_info: [{"id": "28514853-dbb7-4a89-999c-517a28061b9b", "address": "fa:16:3e:f1:72:8d", "network": {"id": "12411632-b7f6-4544-b2e4-b04ae7337547", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-917106830-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4479921caf5f405b8fc49baad390a0e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "de7fa486-5f28-44ae-b0cf-72234ff87546", "external-id": "nsx-vlan-transportzone-229", "segmentation_id": 229, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap28514853-db", "ovs_interfaceid": "28514853-dbb7-4a89-999c-517a28061b9b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2442.663235] env[62824]: DEBUG oslo_vmware.api [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147009, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2442.758599] env[62824]: INFO nova.compute.manager [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: ad9f01dd-d719-41a2-971b-b12cd3e8bdf6] Took 13.20 seconds to build instance. [ 2442.970017] env[62824]: DEBUG nova.scheduler.client.report [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2442.979185] env[62824]: DEBUG oslo_concurrency.lockutils [None req-513f1eb6-422f-45f6-8be2-f708668a058a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Releasing lock "refresh_cache-df621c9c-8a68-4082-8f30-9bf3dda26fb6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2442.979477] env[62824]: DEBUG nova.objects.instance [None req-513f1eb6-422f-45f6-8be2-f708668a058a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lazy-loading 'migration_context' on Instance uuid df621c9c-8a68-4082-8f30-9bf3dda26fb6 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2443.165032] env[62824]: DEBUG oslo_vmware.api [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147009, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2443.215579] env[62824]: DEBUG nova.compute.manager [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 4dff2a76-b0a2-438c-b52f-d6495948d041] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2443.237332] env[62824]: DEBUG nova.virt.hardware [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2443.237573] env[62824]: DEBUG nova.virt.hardware [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2443.237731] env[62824]: DEBUG nova.virt.hardware [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2443.237910] env[62824]: DEBUG nova.virt.hardware [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2443.238070] env[62824]: DEBUG nova.virt.hardware [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2443.238220] env[62824]: DEBUG nova.virt.hardware [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2443.238426] env[62824]: DEBUG nova.virt.hardware [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2443.238586] env[62824]: DEBUG nova.virt.hardware [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2443.238753] env[62824]: DEBUG nova.virt.hardware [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2443.238914] env[62824]: DEBUG nova.virt.hardware [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2443.239110] env[62824]: DEBUG nova.virt.hardware [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2443.239984] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcd337f2-9e5a-4b96-b700-ff5b6c442159 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2443.247989] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb4810f8-6d45-45ce-8099-336b6e42cd24 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2443.260726] env[62824]: DEBUG oslo_concurrency.lockutils [None req-72916ecc-df77-4ba5-845d-bdeda07b3e8e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Lock "ad9f01dd-d719-41a2-971b-b12cd3e8bdf6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.717s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2443.479720] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62824) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 2443.479720] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.205s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2443.482843] env[62824]: DEBUG nova.objects.base [None req-513f1eb6-422f-45f6-8be2-f708668a058a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=62824) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 2443.483770] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e15612b-11eb-45f5-8b77-e2b7f8bb78cd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2443.504110] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-519c4f14-0901-453d-8085-916213793a9f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2443.509589] env[62824]: DEBUG oslo_vmware.api [None req-513f1eb6-422f-45f6-8be2-f708668a058a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2443.509589] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52fb3f7e-67db-6c4c-ca50-c692868c6b37" [ 2443.509589] env[62824]: _type = "Task" [ 2443.509589] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2443.517254] env[62824]: DEBUG oslo_vmware.api [None req-513f1eb6-422f-45f6-8be2-f708668a058a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52fb3f7e-67db-6c4c-ca50-c692868c6b37, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2443.664784] env[62824]: DEBUG oslo_vmware.api [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147009, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2444.020016] env[62824]: DEBUG oslo_vmware.api [None req-513f1eb6-422f-45f6-8be2-f708668a058a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52fb3f7e-67db-6c4c-ca50-c692868c6b37, 'name': SearchDatastore_Task, 'duration_secs': 0.007739} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2444.020289] env[62824]: DEBUG oslo_concurrency.lockutils [None req-513f1eb6-422f-45f6-8be2-f708668a058a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2444.020556] env[62824]: DEBUG oslo_concurrency.lockutils [None req-513f1eb6-422f-45f6-8be2-f708668a058a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2444.060727] env[62824]: DEBUG nova.compute.manager [req-e1d329e8-8117-4fa4-991e-ceb19e7b960e req-efe829e0-ff1b-4c76-9cd2-f86cee34ef21 service nova] [instance: ad9f01dd-d719-41a2-971b-b12cd3e8bdf6] Received event network-changed-ffa86a2f-846f-4178-8ace-b502f5cb81bb {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2444.060924] env[62824]: DEBUG nova.compute.manager [req-e1d329e8-8117-4fa4-991e-ceb19e7b960e req-efe829e0-ff1b-4c76-9cd2-f86cee34ef21 service nova] [instance: ad9f01dd-d719-41a2-971b-b12cd3e8bdf6] Refreshing instance network info cache due to event network-changed-ffa86a2f-846f-4178-8ace-b502f5cb81bb. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2444.061164] env[62824]: DEBUG oslo_concurrency.lockutils [req-e1d329e8-8117-4fa4-991e-ceb19e7b960e req-efe829e0-ff1b-4c76-9cd2-f86cee34ef21 service nova] Acquiring lock "refresh_cache-ad9f01dd-d719-41a2-971b-b12cd3e8bdf6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2444.061308] env[62824]: DEBUG oslo_concurrency.lockutils [req-e1d329e8-8117-4fa4-991e-ceb19e7b960e req-efe829e0-ff1b-4c76-9cd2-f86cee34ef21 service nova] Acquired lock "refresh_cache-ad9f01dd-d719-41a2-971b-b12cd3e8bdf6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2444.061469] env[62824]: DEBUG nova.network.neutron [req-e1d329e8-8117-4fa4-991e-ceb19e7b960e req-efe829e0-ff1b-4c76-9cd2-f86cee34ef21 service nova] [instance: ad9f01dd-d719-41a2-971b-b12cd3e8bdf6] Refreshing network info cache for port ffa86a2f-846f-4178-8ace-b502f5cb81bb {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2444.166051] env[62824]: DEBUG oslo_vmware.api [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147009, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2444.640906] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54996c38-a731-4164-86b6-bfac0704a90c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2444.648619] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09af909d-8dd1-482f-bfb7-b38c183f60eb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2444.683717] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1700f69e-2f36-4ea2-b799-ed4db14d3516 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2444.691404] env[62824]: DEBUG oslo_vmware.api [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147009, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2444.694429] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b89716f7-89bd-4deb-b9b9-4b2ed4454aff {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2444.707580] env[62824]: DEBUG nova.compute.provider_tree [None req-513f1eb6-422f-45f6-8be2-f708668a058a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2444.803548] env[62824]: DEBUG nova.network.neutron [req-e1d329e8-8117-4fa4-991e-ceb19e7b960e req-efe829e0-ff1b-4c76-9cd2-f86cee34ef21 service nova] [instance: ad9f01dd-d719-41a2-971b-b12cd3e8bdf6] Updated VIF entry in instance network info cache for port ffa86a2f-846f-4178-8ace-b502f5cb81bb. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2444.804080] env[62824]: DEBUG nova.network.neutron [req-e1d329e8-8117-4fa4-991e-ceb19e7b960e req-efe829e0-ff1b-4c76-9cd2-f86cee34ef21 service nova] [instance: ad9f01dd-d719-41a2-971b-b12cd3e8bdf6] Updating instance_info_cache with network_info: [{"id": "ffa86a2f-846f-4178-8ace-b502f5cb81bb", "address": "fa:16:3e:d3:92:a1", "network": {"id": "f7936d74-eeef-4c41-8ae7-9427c43663fa", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1643883257-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.224", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c95aa87980de4ec8a2170c2640423301", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6eaa481-1f92-4851-b98e-09ed0daad7cb", "external-id": "nsx-vlan-transportzone-636", "segmentation_id": 636, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapffa86a2f-84", "ovs_interfaceid": "ffa86a2f-846f-4178-8ace-b502f5cb81bb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2445.187756] env[62824]: DEBUG oslo_vmware.api [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147009, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2445.212121] env[62824]: DEBUG nova.scheduler.client.report [None req-513f1eb6-422f-45f6-8be2-f708668a058a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2445.307616] env[62824]: DEBUG oslo_concurrency.lockutils [req-e1d329e8-8117-4fa4-991e-ceb19e7b960e req-efe829e0-ff1b-4c76-9cd2-f86cee34ef21 service nova] Releasing lock "refresh_cache-ad9f01dd-d719-41a2-971b-b12cd3e8bdf6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2445.477292] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2445.477516] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Starting heal instance info cache {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10445}} [ 2445.690031] env[62824]: DEBUG oslo_vmware.api [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147009, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2445.981186] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Skipping network cache update for instance because it is being deleted. {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10494}} [ 2446.189091] env[62824]: DEBUG oslo_vmware.api [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147009, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2446.222509] env[62824]: DEBUG oslo_concurrency.lockutils [None req-513f1eb6-422f-45f6-8be2-f708668a058a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.202s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2446.689251] env[62824]: DEBUG oslo_vmware.api [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147009, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2446.783386] env[62824]: INFO nova.scheduler.client.report [None req-513f1eb6-422f-45f6-8be2-f708668a058a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Deleted allocation for migration a132a1b4-4c08-4cae-bc8c-af84308301aa [ 2447.190183] env[62824]: DEBUG oslo_vmware.api [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147009, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2447.288911] env[62824]: DEBUG oslo_concurrency.lockutils [None req-513f1eb6-422f-45f6-8be2-f708668a058a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "df621c9c-8a68-4082-8f30-9bf3dda26fb6" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.639s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2447.692662] env[62824]: DEBUG oslo_vmware.api [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147009, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2448.191280] env[62824]: DEBUG oslo_vmware.api [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147009, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2448.193851] env[62824]: DEBUG oslo_concurrency.lockutils [None req-893bea24-d1ce-47aa-a915-ae34050c578a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquiring lock "df621c9c-8a68-4082-8f30-9bf3dda26fb6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2448.193981] env[62824]: DEBUG oslo_concurrency.lockutils [None req-893bea24-d1ce-47aa-a915-ae34050c578a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "df621c9c-8a68-4082-8f30-9bf3dda26fb6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2448.194201] env[62824]: DEBUG oslo_concurrency.lockutils [None req-893bea24-d1ce-47aa-a915-ae34050c578a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquiring lock "df621c9c-8a68-4082-8f30-9bf3dda26fb6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2448.194385] env[62824]: DEBUG oslo_concurrency.lockutils [None req-893bea24-d1ce-47aa-a915-ae34050c578a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "df621c9c-8a68-4082-8f30-9bf3dda26fb6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2448.194548] env[62824]: DEBUG oslo_concurrency.lockutils [None req-893bea24-d1ce-47aa-a915-ae34050c578a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "df621c9c-8a68-4082-8f30-9bf3dda26fb6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2448.196505] env[62824]: INFO nova.compute.manager [None req-893bea24-d1ce-47aa-a915-ae34050c578a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Terminating instance [ 2448.662540] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquiring lock "9227cc56-d55d-44ba-a619-63533dd77b66" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2448.662768] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "9227cc56-d55d-44ba-a619-63533dd77b66" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2448.691430] env[62824]: DEBUG oslo_vmware.api [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147009, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2448.699409] env[62824]: DEBUG nova.compute.manager [None req-893bea24-d1ce-47aa-a915-ae34050c578a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2448.699683] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-893bea24-d1ce-47aa-a915-ae34050c578a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2448.700445] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dbf2c62-ef8e-431a-85f2-ffe30f2b964b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2448.706938] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-893bea24-d1ce-47aa-a915-ae34050c578a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2448.707257] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e4abf081-5596-4012-b517-a2dce5c377db {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2448.714190] env[62824]: DEBUG oslo_vmware.api [None req-893bea24-d1ce-47aa-a915-ae34050c578a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2448.714190] env[62824]: value = "task-2147014" [ 2448.714190] env[62824]: _type = "Task" [ 2448.714190] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2448.721600] env[62824]: DEBUG oslo_vmware.api [None req-893bea24-d1ce-47aa-a915-ae34050c578a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2147014, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2448.996039] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Didn't find any instances for network info cache update. {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10531}} [ 2448.996335] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2448.996462] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2448.996613] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2448.996763] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2448.996907] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2448.997082] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2448.997207] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62824) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11064}} [ 2449.165444] env[62824]: DEBUG nova.compute.manager [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 9227cc56-d55d-44ba-a619-63533dd77b66] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2449.191509] env[62824]: DEBUG oslo_vmware.api [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147009, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2449.225788] env[62824]: DEBUG oslo_vmware.api [None req-893bea24-d1ce-47aa-a915-ae34050c578a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2147014, 'name': PowerOffVM_Task, 'duration_secs': 0.222752} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2449.226234] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-893bea24-d1ce-47aa-a915-ae34050c578a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2449.226486] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-893bea24-d1ce-47aa-a915-ae34050c578a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2449.226757] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a2184825-7cd2-42e8-b75b-29d160eba3e1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2449.237298] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-a7e061cc-c794-4218-9571-b7846b71f5eb tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2449.237613] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-a7e061cc-c794-4218-9571-b7846b71f5eb tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Deleting contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2449.237866] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-a7e061cc-c794-4218-9571-b7846b71f5eb tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Deleting the datastore file [datastore1] 5b3df31f-15fe-473f-992c-ddb272661c53 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2449.238212] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e70697ab-ded0-456b-9133-a3842f1594d7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2449.245723] env[62824]: DEBUG oslo_vmware.api [None req-a7e061cc-c794-4218-9571-b7846b71f5eb tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2449.245723] env[62824]: value = "task-2147016" [ 2449.245723] env[62824]: _type = "Task" [ 2449.245723] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2449.253852] env[62824]: DEBUG oslo_vmware.api [None req-a7e061cc-c794-4218-9571-b7846b71f5eb tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2147016, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2449.472114] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-893bea24-d1ce-47aa-a915-ae34050c578a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2449.472365] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-893bea24-d1ce-47aa-a915-ae34050c578a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Deleting contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2449.472574] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-893bea24-d1ce-47aa-a915-ae34050c578a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Deleting the datastore file [datastore2] df621c9c-8a68-4082-8f30-9bf3dda26fb6 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2449.472861] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ad64c50a-1a02-478b-8294-52de1ccd2019 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2449.479773] env[62824]: DEBUG oslo_vmware.api [None req-893bea24-d1ce-47aa-a915-ae34050c578a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2449.479773] env[62824]: value = "task-2147017" [ 2449.479773] env[62824]: _type = "Task" [ 2449.479773] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2449.487664] env[62824]: DEBUG oslo_vmware.api [None req-893bea24-d1ce-47aa-a915-ae34050c578a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2147017, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2449.688441] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2449.688742] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2449.690386] env[62824]: INFO nova.compute.claims [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 9227cc56-d55d-44ba-a619-63533dd77b66] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2449.698939] env[62824]: DEBUG oslo_vmware.api [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147009, 'name': CloneVM_Task} progress is 95%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2449.754737] env[62824]: DEBUG oslo_vmware.api [None req-a7e061cc-c794-4218-9571-b7846b71f5eb tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2147016, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.174359} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2449.755093] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-a7e061cc-c794-4218-9571-b7846b71f5eb tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2449.755345] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-a7e061cc-c794-4218-9571-b7846b71f5eb tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Deleted contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2449.755578] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-a7e061cc-c794-4218-9571-b7846b71f5eb tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2449.755795] env[62824]: INFO nova.compute.manager [None req-a7e061cc-c794-4218-9571-b7846b71f5eb tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Took 8.08 seconds to destroy the instance on the hypervisor. [ 2449.756066] env[62824]: DEBUG oslo.service.loopingcall [None req-a7e061cc-c794-4218-9571-b7846b71f5eb tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2449.756264] env[62824]: DEBUG nova.compute.manager [-] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2449.756361] env[62824]: DEBUG nova.network.neutron [-] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2449.989894] env[62824]: DEBUG oslo_vmware.api [None req-893bea24-d1ce-47aa-a915-ae34050c578a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2147017, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.158133} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2449.990348] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-893bea24-d1ce-47aa-a915-ae34050c578a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2449.990561] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-893bea24-d1ce-47aa-a915-ae34050c578a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Deleted contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2449.990795] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-893bea24-d1ce-47aa-a915-ae34050c578a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2449.991087] env[62824]: INFO nova.compute.manager [None req-893bea24-d1ce-47aa-a915-ae34050c578a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Took 1.29 seconds to destroy the instance on the hypervisor. [ 2449.991379] env[62824]: DEBUG oslo.service.loopingcall [None req-893bea24-d1ce-47aa-a915-ae34050c578a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2449.991594] env[62824]: DEBUG nova.compute.manager [-] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2449.991738] env[62824]: DEBUG nova.network.neutron [-] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2450.198080] env[62824]: DEBUG oslo_vmware.api [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147009, 'name': CloneVM_Task, 'duration_secs': 10.459595} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2450.198080] env[62824]: INFO nova.virt.vmwareapi.vmops [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Created linked-clone VM from snapshot [ 2450.198080] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66804fff-a352-4fae-9150-2e661dfa03f3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2450.209618] env[62824]: DEBUG nova.virt.vmwareapi.images [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Uploading image 13001af0-4e21-48f0-84af-44e4c51ff081 {{(pid=62824) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 2450.242758] env[62824]: DEBUG oslo_vmware.rw_handles [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 2450.242758] env[62824]: value = "vm-438778" [ 2450.242758] env[62824]: _type = "VirtualMachine" [ 2450.242758] env[62824]: }. {{(pid=62824) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 2450.243057] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-71ebff0d-1405-45ff-9a8b-f6b49a5d9877 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2450.250943] env[62824]: DEBUG oslo_vmware.rw_handles [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lease: (returnval){ [ 2450.250943] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52d60a01-9780-69bc-661d-521ce92ef216" [ 2450.250943] env[62824]: _type = "HttpNfcLease" [ 2450.250943] env[62824]: } obtained for exporting VM: (result){ [ 2450.250943] env[62824]: value = "vm-438778" [ 2450.250943] env[62824]: _type = "VirtualMachine" [ 2450.250943] env[62824]: }. {{(pid=62824) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 2450.251199] env[62824]: DEBUG oslo_vmware.api [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the lease: (returnval){ [ 2450.251199] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52d60a01-9780-69bc-661d-521ce92ef216" [ 2450.251199] env[62824]: _type = "HttpNfcLease" [ 2450.251199] env[62824]: } to be ready. {{(pid=62824) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 2450.257460] env[62824]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 2450.257460] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52d60a01-9780-69bc-661d-521ce92ef216" [ 2450.257460] env[62824]: _type = "HttpNfcLease" [ 2450.257460] env[62824]: } is initializing. {{(pid=62824) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 2450.374203] env[62824]: DEBUG nova.compute.manager [req-fd641e2b-ef01-4cce-bff2-18cf4d636871 req-327f62c2-8794-40cb-bca9-804435a01571 service nova] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Received event network-vif-deleted-7be95c5e-6e98-4fec-8550-1c22a787b95b {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2450.374413] env[62824]: INFO nova.compute.manager [req-fd641e2b-ef01-4cce-bff2-18cf4d636871 req-327f62c2-8794-40cb-bca9-804435a01571 service nova] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Neutron deleted interface 7be95c5e-6e98-4fec-8550-1c22a787b95b; detaching it from the instance and deleting it from the info cache [ 2450.374588] env[62824]: DEBUG nova.network.neutron [req-fd641e2b-ef01-4cce-bff2-18cf4d636871 req-327f62c2-8794-40cb-bca9-804435a01571 service nova] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2450.759103] env[62824]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 2450.759103] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52d60a01-9780-69bc-661d-521ce92ef216" [ 2450.759103] env[62824]: _type = "HttpNfcLease" [ 2450.759103] env[62824]: } is ready. {{(pid=62824) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 2450.759508] env[62824]: DEBUG oslo_vmware.rw_handles [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 2450.759508] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52d60a01-9780-69bc-661d-521ce92ef216" [ 2450.759508] env[62824]: _type = "HttpNfcLease" [ 2450.759508] env[62824]: }. {{(pid=62824) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 2450.760224] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8d6b363-ab65-4c97-b3b1-a6aa90d9377f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2450.769802] env[62824]: DEBUG oslo_vmware.rw_handles [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ce1642-2ae5-6f51-e12c-bce13ad26cd0/disk-0.vmdk from lease info. {{(pid=62824) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 2450.770253] env[62824]: DEBUG oslo_vmware.rw_handles [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ce1642-2ae5-6f51-e12c-bce13ad26cd0/disk-0.vmdk for reading. {{(pid=62824) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 2450.850496] env[62824]: DEBUG nova.network.neutron [-] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2450.861041] env[62824]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-20f8b612-de76-47ac-a9a7-a3d98a7899c7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2450.865474] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea1fed6a-98eb-42fe-82af-31359c6d1393 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2450.874568] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4969e869-5162-428d-a337-f8b69b13a90a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2450.881341] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c2e1d7ad-68c6-4b5c-9ec7-de9b8d495491 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2450.908472] env[62824]: DEBUG nova.network.neutron [-] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2450.914259] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-776a2944-c9f6-4a29-bfad-d187790353be {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2450.922879] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd48a510-c956-4c5b-941f-9d400f5c4373 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2450.942764] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d82fe22-d8d2-4964-b820-d90679865e0a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2450.956759] env[62824]: DEBUG nova.compute.provider_tree [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2450.964525] env[62824]: DEBUG nova.compute.manager [req-fd641e2b-ef01-4cce-bff2-18cf4d636871 req-327f62c2-8794-40cb-bca9-804435a01571 service nova] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Detach interface failed, port_id=7be95c5e-6e98-4fec-8550-1c22a787b95b, reason: Instance 5b3df31f-15fe-473f-992c-ddb272661c53 could not be found. {{(pid=62824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11482}} [ 2450.972533] env[62824]: DEBUG nova.network.neutron [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 4dff2a76-b0a2-438c-b52f-d6495948d041] Successfully updated port: 2e0a71bd-c461-4fe9-a810-fe72908cf629 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2451.354057] env[62824]: INFO nova.compute.manager [-] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Took 1.60 seconds to deallocate network for instance. [ 2451.411420] env[62824]: INFO nova.compute.manager [-] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Took 1.42 seconds to deallocate network for instance. [ 2451.466511] env[62824]: DEBUG nova.scheduler.client.report [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2451.474532] env[62824]: DEBUG oslo_concurrency.lockutils [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquiring lock "refresh_cache-4dff2a76-b0a2-438c-b52f-d6495948d041" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2451.474701] env[62824]: DEBUG oslo_concurrency.lockutils [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquired lock "refresh_cache-4dff2a76-b0a2-438c-b52f-d6495948d041" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2451.474947] env[62824]: DEBUG nova.network.neutron [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 4dff2a76-b0a2-438c-b52f-d6495948d041] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2451.862900] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a7e061cc-c794-4218-9571-b7846b71f5eb tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2451.918410] env[62824]: DEBUG oslo_concurrency.lockutils [None req-893bea24-d1ce-47aa-a915-ae34050c578a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2451.974664] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.286s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2451.976049] env[62824]: DEBUG nova.compute.manager [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 9227cc56-d55d-44ba-a619-63533dd77b66] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2451.978556] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a7e061cc-c794-4218-9571-b7846b71f5eb tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.116s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2451.978803] env[62824]: DEBUG nova.objects.instance [None req-a7e061cc-c794-4218-9571-b7846b71f5eb tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lazy-loading 'resources' on Instance uuid 5b3df31f-15fe-473f-992c-ddb272661c53 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2452.014748] env[62824]: DEBUG nova.network.neutron [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 4dff2a76-b0a2-438c-b52f-d6495948d041] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2452.168017] env[62824]: DEBUG nova.network.neutron [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 4dff2a76-b0a2-438c-b52f-d6495948d041] Updating instance_info_cache with network_info: [{"id": "2e0a71bd-c461-4fe9-a810-fe72908cf629", "address": "fa:16:3e:29:2e:10", "network": {"id": "b4ed285c-26a7-461b-8a71-d18e07582ab8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-997150974-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "815ecf9340934733a83d4dcb26612fd4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69f65356-c85e-4b7f-ad28-7c7b5e8cf50c", "external-id": "nsx-vlan-transportzone-281", "segmentation_id": 281, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2e0a71bd-c4", "ovs_interfaceid": "2e0a71bd-c461-4fe9-a810-fe72908cf629", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2452.408666] env[62824]: DEBUG nova.compute.manager [req-d742b84f-d477-4923-85d9-f72ea8611638 req-a76e9d33-d9c1-4c7c-a152-f06903bd35d3 service nova] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Received event network-vif-deleted-28514853-dbb7-4a89-999c-517a28061b9b {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2452.409133] env[62824]: DEBUG nova.compute.manager [req-d742b84f-d477-4923-85d9-f72ea8611638 req-a76e9d33-d9c1-4c7c-a152-f06903bd35d3 service nova] [instance: 4dff2a76-b0a2-438c-b52f-d6495948d041] Received event network-vif-plugged-2e0a71bd-c461-4fe9-a810-fe72908cf629 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2452.409513] env[62824]: DEBUG oslo_concurrency.lockutils [req-d742b84f-d477-4923-85d9-f72ea8611638 req-a76e9d33-d9c1-4c7c-a152-f06903bd35d3 service nova] Acquiring lock "4dff2a76-b0a2-438c-b52f-d6495948d041-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2452.409766] env[62824]: DEBUG oslo_concurrency.lockutils [req-d742b84f-d477-4923-85d9-f72ea8611638 req-a76e9d33-d9c1-4c7c-a152-f06903bd35d3 service nova] Lock "4dff2a76-b0a2-438c-b52f-d6495948d041-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2452.410068] env[62824]: DEBUG oslo_concurrency.lockutils [req-d742b84f-d477-4923-85d9-f72ea8611638 req-a76e9d33-d9c1-4c7c-a152-f06903bd35d3 service nova] Lock "4dff2a76-b0a2-438c-b52f-d6495948d041-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2452.410320] env[62824]: DEBUG nova.compute.manager [req-d742b84f-d477-4923-85d9-f72ea8611638 req-a76e9d33-d9c1-4c7c-a152-f06903bd35d3 service nova] [instance: 4dff2a76-b0a2-438c-b52f-d6495948d041] No waiting events found dispatching network-vif-plugged-2e0a71bd-c461-4fe9-a810-fe72908cf629 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2452.410627] env[62824]: WARNING nova.compute.manager [req-d742b84f-d477-4923-85d9-f72ea8611638 req-a76e9d33-d9c1-4c7c-a152-f06903bd35d3 service nova] [instance: 4dff2a76-b0a2-438c-b52f-d6495948d041] Received unexpected event network-vif-plugged-2e0a71bd-c461-4fe9-a810-fe72908cf629 for instance with vm_state building and task_state spawning. [ 2452.410815] env[62824]: DEBUG nova.compute.manager [req-d742b84f-d477-4923-85d9-f72ea8611638 req-a76e9d33-d9c1-4c7c-a152-f06903bd35d3 service nova] [instance: 4dff2a76-b0a2-438c-b52f-d6495948d041] Received event network-changed-2e0a71bd-c461-4fe9-a810-fe72908cf629 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2452.411119] env[62824]: DEBUG nova.compute.manager [req-d742b84f-d477-4923-85d9-f72ea8611638 req-a76e9d33-d9c1-4c7c-a152-f06903bd35d3 service nova] [instance: 4dff2a76-b0a2-438c-b52f-d6495948d041] Refreshing instance network info cache due to event network-changed-2e0a71bd-c461-4fe9-a810-fe72908cf629. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2452.411314] env[62824]: DEBUG oslo_concurrency.lockutils [req-d742b84f-d477-4923-85d9-f72ea8611638 req-a76e9d33-d9c1-4c7c-a152-f06903bd35d3 service nova] Acquiring lock "refresh_cache-4dff2a76-b0a2-438c-b52f-d6495948d041" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2452.484065] env[62824]: DEBUG nova.compute.utils [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2452.488055] env[62824]: DEBUG nova.compute.manager [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 9227cc56-d55d-44ba-a619-63533dd77b66] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2452.488297] env[62824]: DEBUG nova.network.neutron [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 9227cc56-d55d-44ba-a619-63533dd77b66] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2452.613736] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9278b27-92ea-42cc-b6ca-9496c9d8b69a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2452.622217] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5252555-e3aa-449d-892f-3b9088e00889 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2452.655465] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37d43e83-aeff-4ebb-b4fd-0df2a1d36167 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2452.760132] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f71a072b-e51f-43f1-a0c9-fdeba8809e2b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2452.760132] env[62824]: DEBUG oslo_concurrency.lockutils [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Releasing lock "refresh_cache-4dff2a76-b0a2-438c-b52f-d6495948d041" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2452.760132] env[62824]: DEBUG nova.compute.manager [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 4dff2a76-b0a2-438c-b52f-d6495948d041] Instance network_info: |[{"id": "2e0a71bd-c461-4fe9-a810-fe72908cf629", "address": "fa:16:3e:29:2e:10", "network": {"id": "b4ed285c-26a7-461b-8a71-d18e07582ab8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-997150974-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "815ecf9340934733a83d4dcb26612fd4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69f65356-c85e-4b7f-ad28-7c7b5e8cf50c", "external-id": "nsx-vlan-transportzone-281", "segmentation_id": 281, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2e0a71bd-c4", "ovs_interfaceid": "2e0a71bd-c461-4fe9-a810-fe72908cf629", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2452.760132] env[62824]: DEBUG oslo_concurrency.lockutils [req-d742b84f-d477-4923-85d9-f72ea8611638 req-a76e9d33-d9c1-4c7c-a152-f06903bd35d3 service nova] Acquired lock "refresh_cache-4dff2a76-b0a2-438c-b52f-d6495948d041" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2452.760132] env[62824]: DEBUG nova.network.neutron [req-d742b84f-d477-4923-85d9-f72ea8611638 req-a76e9d33-d9c1-4c7c-a152-f06903bd35d3 service nova] [instance: 4dff2a76-b0a2-438c-b52f-d6495948d041] Refreshing network info cache for port 2e0a71bd-c461-4fe9-a810-fe72908cf629 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2452.760132] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 4dff2a76-b0a2-438c-b52f-d6495948d041] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:29:2e:10', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '69f65356-c85e-4b7f-ad28-7c7b5e8cf50c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2e0a71bd-c461-4fe9-a810-fe72908cf629', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2452.760132] env[62824]: DEBUG oslo.service.loopingcall [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2452.760132] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4dff2a76-b0a2-438c-b52f-d6495948d041] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2452.760132] env[62824]: DEBUG nova.compute.provider_tree [None req-a7e061cc-c794-4218-9571-b7846b71f5eb tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2452.760132] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ae9f6655-5737-4f82-92d9-fc1a4252e4f3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2452.760132] env[62824]: DEBUG nova.scheduler.client.report [None req-a7e061cc-c794-4218-9571-b7846b71f5eb tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2452.760132] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2452.760132] env[62824]: value = "task-2147019" [ 2452.760132] env[62824]: _type = "Task" [ 2452.760132] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2452.760132] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2147019, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2452.774406] env[62824]: DEBUG nova.policy [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e8c10fcbfe69448fa71cfad4a7c8e179', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4479921caf5f405b8fc49baad390a0e9', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2452.992279] env[62824]: DEBUG nova.compute.manager [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 9227cc56-d55d-44ba-a619-63533dd77b66] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2453.075114] env[62824]: DEBUG nova.network.neutron [req-d742b84f-d477-4923-85d9-f72ea8611638 req-a76e9d33-d9c1-4c7c-a152-f06903bd35d3 service nova] [instance: 4dff2a76-b0a2-438c-b52f-d6495948d041] Updated VIF entry in instance network info cache for port 2e0a71bd-c461-4fe9-a810-fe72908cf629. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2453.075543] env[62824]: DEBUG nova.network.neutron [req-d742b84f-d477-4923-85d9-f72ea8611638 req-a76e9d33-d9c1-4c7c-a152-f06903bd35d3 service nova] [instance: 4dff2a76-b0a2-438c-b52f-d6495948d041] Updating instance_info_cache with network_info: [{"id": "2e0a71bd-c461-4fe9-a810-fe72908cf629", "address": "fa:16:3e:29:2e:10", "network": {"id": "b4ed285c-26a7-461b-8a71-d18e07582ab8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-997150974-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "815ecf9340934733a83d4dcb26612fd4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69f65356-c85e-4b7f-ad28-7c7b5e8cf50c", "external-id": "nsx-vlan-transportzone-281", "segmentation_id": 281, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2e0a71bd-c4", "ovs_interfaceid": "2e0a71bd-c461-4fe9-a810-fe72908cf629", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2453.171024] env[62824]: DEBUG nova.network.neutron [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 9227cc56-d55d-44ba-a619-63533dd77b66] Successfully created port: 0be28a80-9cee-483c-9d58-c3e6e610b89d {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2453.218665] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a7e061cc-c794-4218-9571-b7846b71f5eb tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.240s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2453.221855] env[62824]: DEBUG oslo_concurrency.lockutils [None req-893bea24-d1ce-47aa-a915-ae34050c578a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.303s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2453.222107] env[62824]: DEBUG oslo_concurrency.lockutils [None req-893bea24-d1ce-47aa-a915-ae34050c578a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2453.233581] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2147019, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2453.246469] env[62824]: INFO nova.scheduler.client.report [None req-893bea24-d1ce-47aa-a915-ae34050c578a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Deleted allocations for instance df621c9c-8a68-4082-8f30-9bf3dda26fb6 [ 2453.250319] env[62824]: INFO nova.scheduler.client.report [None req-a7e061cc-c794-4218-9571-b7846b71f5eb tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Deleted allocations for instance 5b3df31f-15fe-473f-992c-ddb272661c53 [ 2453.578208] env[62824]: DEBUG oslo_concurrency.lockutils [req-d742b84f-d477-4923-85d9-f72ea8611638 req-a76e9d33-d9c1-4c7c-a152-f06903bd35d3 service nova] Releasing lock "refresh_cache-4dff2a76-b0a2-438c-b52f-d6495948d041" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2453.735021] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2147019, 'name': CreateVM_Task, 'duration_secs': 0.75365} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2453.735021] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4dff2a76-b0a2-438c-b52f-d6495948d041] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2453.735356] env[62824]: DEBUG oslo_concurrency.lockutils [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2453.735512] env[62824]: DEBUG oslo_concurrency.lockutils [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2453.735841] env[62824]: DEBUG oslo_concurrency.lockutils [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2453.736116] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1fc49920-5609-4f63-82e9-f62b8895006c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2453.741390] env[62824]: DEBUG oslo_vmware.api [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2453.741390] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]523ba03a-5a19-2701-5259-b2f872cbed72" [ 2453.741390] env[62824]: _type = "Task" [ 2453.741390] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2453.748828] env[62824]: DEBUG oslo_vmware.api [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]523ba03a-5a19-2701-5259-b2f872cbed72, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2453.758838] env[62824]: DEBUG oslo_concurrency.lockutils [None req-893bea24-d1ce-47aa-a915-ae34050c578a tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "df621c9c-8a68-4082-8f30-9bf3dda26fb6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.565s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2453.762489] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a7e061cc-c794-4218-9571-b7846b71f5eb tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lock "5b3df31f-15fe-473f-992c-ddb272661c53" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.600s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2454.001152] env[62824]: DEBUG nova.compute.manager [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 9227cc56-d55d-44ba-a619-63533dd77b66] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2454.022901] env[62824]: DEBUG nova.virt.hardware [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2454.023195] env[62824]: DEBUG nova.virt.hardware [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2454.023368] env[62824]: DEBUG nova.virt.hardware [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2454.023555] env[62824]: DEBUG nova.virt.hardware [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2454.023715] env[62824]: DEBUG nova.virt.hardware [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2454.023851] env[62824]: DEBUG nova.virt.hardware [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2454.024076] env[62824]: DEBUG nova.virt.hardware [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2454.024244] env[62824]: DEBUG nova.virt.hardware [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2454.024414] env[62824]: DEBUG nova.virt.hardware [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2454.024577] env[62824]: DEBUG nova.virt.hardware [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2454.024746] env[62824]: DEBUG nova.virt.hardware [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2454.025630] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3728b700-21d5-4431-8bb6-2bf4d6438b9a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2454.033600] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d9ac28b-623e-4fdb-8662-9c972d3bea83 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2454.251501] env[62824]: DEBUG oslo_vmware.api [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]523ba03a-5a19-2701-5259-b2f872cbed72, 'name': SearchDatastore_Task, 'duration_secs': 0.012254} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2454.251800] env[62824]: DEBUG oslo_concurrency.lockutils [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2454.252044] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 4dff2a76-b0a2-438c-b52f-d6495948d041] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2454.252329] env[62824]: DEBUG oslo_concurrency.lockutils [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2454.252482] env[62824]: DEBUG oslo_concurrency.lockutils [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2454.252693] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2454.252958] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d262ff0f-f2cd-4031-ba18-5ca5f682eef4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2454.261611] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2454.261814] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2454.262820] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-531b4e31-6822-42d1-9518-e16f44365db1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2454.268347] env[62824]: DEBUG oslo_vmware.api [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2454.268347] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52563f19-2e41-3e00-eacd-9d596e2c3c16" [ 2454.268347] env[62824]: _type = "Task" [ 2454.268347] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2454.277494] env[62824]: DEBUG oslo_vmware.api [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52563f19-2e41-3e00-eacd-9d596e2c3c16, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2454.612125] env[62824]: DEBUG nova.compute.manager [req-fb862916-79cb-41cc-9c95-3e40f681fe8b req-5702a7e0-9de3-447a-bd84-248749e771e9 service nova] [instance: 9227cc56-d55d-44ba-a619-63533dd77b66] Received event network-vif-plugged-0be28a80-9cee-483c-9d58-c3e6e610b89d {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2454.612125] env[62824]: DEBUG oslo_concurrency.lockutils [req-fb862916-79cb-41cc-9c95-3e40f681fe8b req-5702a7e0-9de3-447a-bd84-248749e771e9 service nova] Acquiring lock "9227cc56-d55d-44ba-a619-63533dd77b66-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2454.612125] env[62824]: DEBUG oslo_concurrency.lockutils [req-fb862916-79cb-41cc-9c95-3e40f681fe8b req-5702a7e0-9de3-447a-bd84-248749e771e9 service nova] Lock "9227cc56-d55d-44ba-a619-63533dd77b66-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2454.612125] env[62824]: DEBUG oslo_concurrency.lockutils [req-fb862916-79cb-41cc-9c95-3e40f681fe8b req-5702a7e0-9de3-447a-bd84-248749e771e9 service nova] Lock "9227cc56-d55d-44ba-a619-63533dd77b66-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2454.612125] env[62824]: DEBUG nova.compute.manager [req-fb862916-79cb-41cc-9c95-3e40f681fe8b req-5702a7e0-9de3-447a-bd84-248749e771e9 service nova] [instance: 9227cc56-d55d-44ba-a619-63533dd77b66] No waiting events found dispatching network-vif-plugged-0be28a80-9cee-483c-9d58-c3e6e610b89d {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2454.612125] env[62824]: WARNING nova.compute.manager [req-fb862916-79cb-41cc-9c95-3e40f681fe8b req-5702a7e0-9de3-447a-bd84-248749e771e9 service nova] [instance: 9227cc56-d55d-44ba-a619-63533dd77b66] Received unexpected event network-vif-plugged-0be28a80-9cee-483c-9d58-c3e6e610b89d for instance with vm_state building and task_state spawning. [ 2454.710076] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquiring lock "80c57382-8f6f-4537-b210-70a902b02cb4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2454.710322] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lock "80c57382-8f6f-4537-b210-70a902b02cb4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2454.725406] env[62824]: DEBUG nova.network.neutron [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 9227cc56-d55d-44ba-a619-63533dd77b66] Successfully updated port: 0be28a80-9cee-483c-9d58-c3e6e610b89d {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2454.779523] env[62824]: DEBUG oslo_vmware.api [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52563f19-2e41-3e00-eacd-9d596e2c3c16, 'name': SearchDatastore_Task, 'duration_secs': 0.009656} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2454.780366] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a2a46a16-ca2b-4bef-b26f-c7e696df33e1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2454.786241] env[62824]: DEBUG oslo_vmware.api [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2454.786241] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]529d9a49-cd0e-cd04-515f-2730de262f26" [ 2454.786241] env[62824]: _type = "Task" [ 2454.786241] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2454.795120] env[62824]: DEBUG oslo_vmware.api [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]529d9a49-cd0e-cd04-515f-2730de262f26, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2455.212977] env[62824]: DEBUG nova.compute.manager [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2455.226639] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquiring lock "refresh_cache-9227cc56-d55d-44ba-a619-63533dd77b66" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2455.226869] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquired lock "refresh_cache-9227cc56-d55d-44ba-a619-63533dd77b66" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2455.227039] env[62824]: DEBUG nova.network.neutron [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 9227cc56-d55d-44ba-a619-63533dd77b66] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2455.297586] env[62824]: DEBUG oslo_vmware.api [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]529d9a49-cd0e-cd04-515f-2730de262f26, 'name': SearchDatastore_Task, 'duration_secs': 0.016584} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2455.297846] env[62824]: DEBUG oslo_concurrency.lockutils [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2455.298154] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 4dff2a76-b0a2-438c-b52f-d6495948d041/4dff2a76-b0a2-438c-b52f-d6495948d041.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2455.298440] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-db2b9df7-454d-42ad-a43c-73d16f170726 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2455.305126] env[62824]: DEBUG oslo_vmware.api [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2455.305126] env[62824]: value = "task-2147020" [ 2455.305126] env[62824]: _type = "Task" [ 2455.305126] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2455.312953] env[62824]: DEBUG oslo_vmware.api [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2147020, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2455.737319] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2455.737674] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2455.739578] env[62824]: INFO nova.compute.claims [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2455.774895] env[62824]: DEBUG nova.network.neutron [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 9227cc56-d55d-44ba-a619-63533dd77b66] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2455.815893] env[62824]: DEBUG oslo_vmware.api [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2147020, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2455.930228] env[62824]: DEBUG nova.network.neutron [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 9227cc56-d55d-44ba-a619-63533dd77b66] Updating instance_info_cache with network_info: [{"id": "0be28a80-9cee-483c-9d58-c3e6e610b89d", "address": "fa:16:3e:08:42:09", "network": {"id": "12411632-b7f6-4544-b2e4-b04ae7337547", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-917106830-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4479921caf5f405b8fc49baad390a0e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "de7fa486-5f28-44ae-b0cf-72234ff87546", "external-id": "nsx-vlan-transportzone-229", "segmentation_id": 229, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0be28a80-9c", "ovs_interfaceid": "0be28a80-9cee-483c-9d58-c3e6e610b89d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2456.288377] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2456.317823] env[62824]: DEBUG oslo_vmware.api [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2147020, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.511253} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2456.318235] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 4dff2a76-b0a2-438c-b52f-d6495948d041/4dff2a76-b0a2-438c-b52f-d6495948d041.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2456.318476] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 4dff2a76-b0a2-438c-b52f-d6495948d041] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2456.318731] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6ad5f1a0-e829-4f17-b517-63bbe4209307 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2456.328338] env[62824]: DEBUG oslo_vmware.api [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2456.328338] env[62824]: value = "task-2147021" [ 2456.328338] env[62824]: _type = "Task" [ 2456.328338] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2456.337505] env[62824]: DEBUG oslo_vmware.api [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2147021, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2456.431741] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Releasing lock "refresh_cache-9227cc56-d55d-44ba-a619-63533dd77b66" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2456.432079] env[62824]: DEBUG nova.compute.manager [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 9227cc56-d55d-44ba-a619-63533dd77b66] Instance network_info: |[{"id": "0be28a80-9cee-483c-9d58-c3e6e610b89d", "address": "fa:16:3e:08:42:09", "network": {"id": "12411632-b7f6-4544-b2e4-b04ae7337547", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-917106830-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4479921caf5f405b8fc49baad390a0e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "de7fa486-5f28-44ae-b0cf-72234ff87546", "external-id": "nsx-vlan-transportzone-229", "segmentation_id": 229, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0be28a80-9c", "ovs_interfaceid": "0be28a80-9cee-483c-9d58-c3e6e610b89d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2456.432661] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 9227cc56-d55d-44ba-a619-63533dd77b66] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:08:42:09', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'de7fa486-5f28-44ae-b0cf-72234ff87546', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0be28a80-9cee-483c-9d58-c3e6e610b89d', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2456.440235] env[62824]: DEBUG oslo.service.loopingcall [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2456.440495] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9227cc56-d55d-44ba-a619-63533dd77b66] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2456.440711] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-88e07bda-f7b0-43ba-a7f4-c28b2780fc2d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2456.462663] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2456.462663] env[62824]: value = "task-2147022" [ 2456.462663] env[62824]: _type = "Task" [ 2456.462663] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2456.470997] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2147022, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2456.629216] env[62824]: DEBUG nova.compute.manager [req-8523d3ba-a41a-40bd-8364-f06b315c8636 req-0ae83ed9-1646-4f60-b42a-cca2b24f46db service nova] [instance: 9227cc56-d55d-44ba-a619-63533dd77b66] Received event network-changed-0be28a80-9cee-483c-9d58-c3e6e610b89d {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2456.629445] env[62824]: DEBUG nova.compute.manager [req-8523d3ba-a41a-40bd-8364-f06b315c8636 req-0ae83ed9-1646-4f60-b42a-cca2b24f46db service nova] [instance: 9227cc56-d55d-44ba-a619-63533dd77b66] Refreshing instance network info cache due to event network-changed-0be28a80-9cee-483c-9d58-c3e6e610b89d. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2456.629664] env[62824]: DEBUG oslo_concurrency.lockutils [req-8523d3ba-a41a-40bd-8364-f06b315c8636 req-0ae83ed9-1646-4f60-b42a-cca2b24f46db service nova] Acquiring lock "refresh_cache-9227cc56-d55d-44ba-a619-63533dd77b66" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2456.629809] env[62824]: DEBUG oslo_concurrency.lockutils [req-8523d3ba-a41a-40bd-8364-f06b315c8636 req-0ae83ed9-1646-4f60-b42a-cca2b24f46db service nova] Acquired lock "refresh_cache-9227cc56-d55d-44ba-a619-63533dd77b66" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2456.629972] env[62824]: DEBUG nova.network.neutron [req-8523d3ba-a41a-40bd-8364-f06b315c8636 req-0ae83ed9-1646-4f60-b42a-cca2b24f46db service nova] [instance: 9227cc56-d55d-44ba-a619-63533dd77b66] Refreshing network info cache for port 0be28a80-9cee-483c-9d58-c3e6e610b89d {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2456.838280] env[62824]: DEBUG oslo_vmware.api [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2147021, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.102783} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2456.838698] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 4dff2a76-b0a2-438c-b52f-d6495948d041] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2456.839535] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf3cfa88-430f-45dc-86de-026bcc811cd3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2456.865079] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 4dff2a76-b0a2-438c-b52f-d6495948d041] Reconfiguring VM instance instance-00000072 to attach disk [datastore2] 4dff2a76-b0a2-438c-b52f-d6495948d041/4dff2a76-b0a2-438c-b52f-d6495948d041.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2456.868177] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-517d5564-78c1-420b-a9da-62e2a644bd26 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2456.890545] env[62824]: DEBUG oslo_vmware.api [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2456.890545] env[62824]: value = "task-2147023" [ 2456.890545] env[62824]: _type = "Task" [ 2456.890545] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2456.896223] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58ee5baf-42fb-4f79-b6ec-a78d7d04029c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2456.902480] env[62824]: DEBUG oslo_vmware.api [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2147023, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2456.907488] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1217dbbd-0fc5-4112-ab39-122811350aca {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2456.940263] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fd7192f-8218-4c88-a4fa-106b6f046e7f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2456.948254] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3411f422-f111-460b-8e9c-f62776e6193a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2456.963673] env[62824]: DEBUG nova.compute.provider_tree [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2456.975255] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2147022, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2457.360964] env[62824]: DEBUG nova.network.neutron [req-8523d3ba-a41a-40bd-8364-f06b315c8636 req-0ae83ed9-1646-4f60-b42a-cca2b24f46db service nova] [instance: 9227cc56-d55d-44ba-a619-63533dd77b66] Updated VIF entry in instance network info cache for port 0be28a80-9cee-483c-9d58-c3e6e610b89d. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2457.362025] env[62824]: DEBUG nova.network.neutron [req-8523d3ba-a41a-40bd-8364-f06b315c8636 req-0ae83ed9-1646-4f60-b42a-cca2b24f46db service nova] [instance: 9227cc56-d55d-44ba-a619-63533dd77b66] Updating instance_info_cache with network_info: [{"id": "0be28a80-9cee-483c-9d58-c3e6e610b89d", "address": "fa:16:3e:08:42:09", "network": {"id": "12411632-b7f6-4544-b2e4-b04ae7337547", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-917106830-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4479921caf5f405b8fc49baad390a0e9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "de7fa486-5f28-44ae-b0cf-72234ff87546", "external-id": "nsx-vlan-transportzone-229", "segmentation_id": 229, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0be28a80-9c", "ovs_interfaceid": "0be28a80-9cee-483c-9d58-c3e6e610b89d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2457.402519] env[62824]: DEBUG oslo_vmware.api [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2147023, 'name': ReconfigVM_Task, 'duration_secs': 0.473463} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2457.402854] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 4dff2a76-b0a2-438c-b52f-d6495948d041] Reconfigured VM instance instance-00000072 to attach disk [datastore2] 4dff2a76-b0a2-438c-b52f-d6495948d041/4dff2a76-b0a2-438c-b52f-d6495948d041.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2457.403905] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8bd613d4-f322-46a2-8e91-133beed123c0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2457.411582] env[62824]: DEBUG oslo_vmware.api [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2457.411582] env[62824]: value = "task-2147024" [ 2457.411582] env[62824]: _type = "Task" [ 2457.411582] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2457.420658] env[62824]: DEBUG oslo_vmware.api [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2147024, 'name': Rename_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2457.470081] env[62824]: DEBUG nova.scheduler.client.report [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2457.476952] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2147022, 'name': CreateVM_Task, 'duration_secs': 0.684794} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2457.477393] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9227cc56-d55d-44ba-a619-63533dd77b66] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2457.478134] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2457.478351] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2457.478718] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2457.479006] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-34fd86f1-b1b6-4f23-9a2f-b879c99a9369 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2457.483880] env[62824]: DEBUG oslo_vmware.api [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2457.483880] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52aa2ad1-1a6f-47d3-f381-13d3a05ab9a9" [ 2457.483880] env[62824]: _type = "Task" [ 2457.483880] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2457.493582] env[62824]: DEBUG oslo_vmware.api [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52aa2ad1-1a6f-47d3-f381-13d3a05ab9a9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2457.864676] env[62824]: DEBUG oslo_concurrency.lockutils [req-8523d3ba-a41a-40bd-8364-f06b315c8636 req-0ae83ed9-1646-4f60-b42a-cca2b24f46db service nova] Releasing lock "refresh_cache-9227cc56-d55d-44ba-a619-63533dd77b66" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2457.922203] env[62824]: DEBUG oslo_vmware.api [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2147024, 'name': Rename_Task, 'duration_secs': 0.385955} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2457.922500] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 4dff2a76-b0a2-438c-b52f-d6495948d041] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2457.922782] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-88f9d23c-8199-4729-b35f-03580539f485 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2457.929577] env[62824]: DEBUG oslo_vmware.api [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2457.929577] env[62824]: value = "task-2147025" [ 2457.929577] env[62824]: _type = "Task" [ 2457.929577] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2457.937665] env[62824]: DEBUG oslo_vmware.api [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2147025, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2457.978848] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.241s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2457.979343] env[62824]: DEBUG nova.compute.manager [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2457.994318] env[62824]: DEBUG oslo_vmware.api [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52aa2ad1-1a6f-47d3-f381-13d3a05ab9a9, 'name': SearchDatastore_Task, 'duration_secs': 0.03344} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2457.994647] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2457.994898] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 9227cc56-d55d-44ba-a619-63533dd77b66] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2457.995177] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2457.995470] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2457.995557] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2457.995875] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d69c087f-1e6f-4e32-a4c6-3a967afcaeee {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2458.007890] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2458.008165] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2458.009233] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ddec5410-cb88-4bbc-a5e1-704d16aefce4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2458.016732] env[62824]: DEBUG oslo_vmware.api [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2458.016732] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52a3d739-9067-283b-9a57-550ea91563b0" [ 2458.016732] env[62824]: _type = "Task" [ 2458.016732] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2458.028640] env[62824]: DEBUG oslo_vmware.api [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52a3d739-9067-283b-9a57-550ea91563b0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2458.441149] env[62824]: DEBUG oslo_vmware.api [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2147025, 'name': PowerOnVM_Task} progress is 78%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2458.485017] env[62824]: DEBUG nova.compute.utils [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2458.486531] env[62824]: DEBUG nova.compute.manager [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2458.487252] env[62824]: DEBUG nova.network.neutron [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2458.527910] env[62824]: DEBUG oslo_vmware.api [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52a3d739-9067-283b-9a57-550ea91563b0, 'name': SearchDatastore_Task, 'duration_secs': 0.01897} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2458.528697] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ff6e6e44-dece-4783-afe6-4469f5ae6498 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2458.535040] env[62824]: DEBUG oslo_vmware.api [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2458.535040] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]525b74bd-4849-c7b6-d12e-6fbb3c834cd1" [ 2458.535040] env[62824]: _type = "Task" [ 2458.535040] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2458.539912] env[62824]: DEBUG nova.policy [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5a5891a75fee4caf92d4bba152254ad5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7fcdbfc29df64a1ba8d982bdcc667b64', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2458.546981] env[62824]: DEBUG oslo_vmware.api [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]525b74bd-4849-c7b6-d12e-6fbb3c834cd1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2458.841498] env[62824]: DEBUG nova.network.neutron [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Successfully created port: ae875573-c95e-4613-b343-e3e5e3829964 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2458.942539] env[62824]: DEBUG oslo_vmware.api [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2147025, 'name': PowerOnVM_Task, 'duration_secs': 0.957736} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2458.942981] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 4dff2a76-b0a2-438c-b52f-d6495948d041] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2458.943100] env[62824]: INFO nova.compute.manager [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 4dff2a76-b0a2-438c-b52f-d6495948d041] Took 15.73 seconds to spawn the instance on the hypervisor. [ 2458.943214] env[62824]: DEBUG nova.compute.manager [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 4dff2a76-b0a2-438c-b52f-d6495948d041] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2458.944030] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-317678b4-079b-42d7-a6f0-18887afc4e98 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2458.990443] env[62824]: DEBUG nova.compute.manager [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2459.045909] env[62824]: DEBUG oslo_vmware.api [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]525b74bd-4849-c7b6-d12e-6fbb3c834cd1, 'name': SearchDatastore_Task, 'duration_secs': 0.013147} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2459.046203] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2459.046466] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 9227cc56-d55d-44ba-a619-63533dd77b66/9227cc56-d55d-44ba-a619-63533dd77b66.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2459.046728] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1cab71f7-378b-419b-a86f-97d78ed6f59f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2459.054141] env[62824]: DEBUG oslo_vmware.api [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2459.054141] env[62824]: value = "task-2147026" [ 2459.054141] env[62824]: _type = "Task" [ 2459.054141] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2459.062030] env[62824]: DEBUG oslo_vmware.api [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2147026, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2459.461572] env[62824]: INFO nova.compute.manager [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 4dff2a76-b0a2-438c-b52f-d6495948d041] Took 23.00 seconds to build instance. [ 2459.567253] env[62824]: DEBUG oslo_vmware.api [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2147026, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2459.940209] env[62824]: DEBUG nova.compute.manager [req-05dfe276-8c43-4e29-bae2-07bfd3c89f0b req-8dfe626f-1d61-4c4e-b7b5-5d910bb1c483 service nova] [instance: 4dff2a76-b0a2-438c-b52f-d6495948d041] Received event network-changed-2e0a71bd-c461-4fe9-a810-fe72908cf629 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2459.940410] env[62824]: DEBUG nova.compute.manager [req-05dfe276-8c43-4e29-bae2-07bfd3c89f0b req-8dfe626f-1d61-4c4e-b7b5-5d910bb1c483 service nova] [instance: 4dff2a76-b0a2-438c-b52f-d6495948d041] Refreshing instance network info cache due to event network-changed-2e0a71bd-c461-4fe9-a810-fe72908cf629. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2459.940655] env[62824]: DEBUG oslo_concurrency.lockutils [req-05dfe276-8c43-4e29-bae2-07bfd3c89f0b req-8dfe626f-1d61-4c4e-b7b5-5d910bb1c483 service nova] Acquiring lock "refresh_cache-4dff2a76-b0a2-438c-b52f-d6495948d041" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2459.940788] env[62824]: DEBUG oslo_concurrency.lockutils [req-05dfe276-8c43-4e29-bae2-07bfd3c89f0b req-8dfe626f-1d61-4c4e-b7b5-5d910bb1c483 service nova] Acquired lock "refresh_cache-4dff2a76-b0a2-438c-b52f-d6495948d041" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2459.940950] env[62824]: DEBUG nova.network.neutron [req-05dfe276-8c43-4e29-bae2-07bfd3c89f0b req-8dfe626f-1d61-4c4e-b7b5-5d910bb1c483 service nova] [instance: 4dff2a76-b0a2-438c-b52f-d6495948d041] Refreshing network info cache for port 2e0a71bd-c461-4fe9-a810-fe72908cf629 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2459.964619] env[62824]: DEBUG oslo_concurrency.lockutils [None req-299b97b4-e4fb-4e83-84fd-a0e5fb9acf81 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "4dff2a76-b0a2-438c-b52f-d6495948d041" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.517s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2460.006292] env[62824]: DEBUG nova.compute.manager [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2460.016684] env[62824]: DEBUG oslo_vmware.rw_handles [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ce1642-2ae5-6f51-e12c-bce13ad26cd0/disk-0.vmdk. {{(pid=62824) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 2460.017899] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e6a85c3-b522-4818-90c1-c3530780ad61 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2460.024691] env[62824]: DEBUG oslo_vmware.rw_handles [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ce1642-2ae5-6f51-e12c-bce13ad26cd0/disk-0.vmdk is in state: ready. {{(pid=62824) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 2460.024860] env[62824]: ERROR oslo_vmware.rw_handles [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ce1642-2ae5-6f51-e12c-bce13ad26cd0/disk-0.vmdk due to incomplete transfer. [ 2460.025095] env[62824]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-618e8271-6efd-42d4-94e6-3803d921ec8d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2460.028300] env[62824]: DEBUG nova.virt.hardware [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2460.028545] env[62824]: DEBUG nova.virt.hardware [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2460.028705] env[62824]: DEBUG nova.virt.hardware [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2460.028891] env[62824]: DEBUG nova.virt.hardware [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2460.029049] env[62824]: DEBUG nova.virt.hardware [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2460.029203] env[62824]: DEBUG nova.virt.hardware [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2460.029416] env[62824]: DEBUG nova.virt.hardware [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2460.029574] env[62824]: DEBUG nova.virt.hardware [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2460.029741] env[62824]: DEBUG nova.virt.hardware [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2460.029907] env[62824]: DEBUG nova.virt.hardware [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2460.030095] env[62824]: DEBUG nova.virt.hardware [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2460.030828] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de598525-5856-44e6-9afe-667fd1b742b3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2460.040620] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-943a7b2c-5a26-49de-9597-dcdb5f84101a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2460.045069] env[62824]: DEBUG oslo_vmware.rw_handles [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ce1642-2ae5-6f51-e12c-bce13ad26cd0/disk-0.vmdk. {{(pid=62824) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 2460.045264] env[62824]: DEBUG nova.virt.vmwareapi.images [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Uploaded image 13001af0-4e21-48f0-84af-44e4c51ff081 to the Glance image server {{(pid=62824) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 2460.047321] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Destroying the VM {{(pid=62824) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 2460.047867] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-b9419390-b0ba-4e21-9896-0d13442924be {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2460.060911] env[62824]: DEBUG oslo_vmware.api [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2460.060911] env[62824]: value = "task-2147027" [ 2460.060911] env[62824]: _type = "Task" [ 2460.060911] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2460.066884] env[62824]: DEBUG oslo_vmware.api [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2147026, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.55032} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2460.067423] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 9227cc56-d55d-44ba-a619-63533dd77b66/9227cc56-d55d-44ba-a619-63533dd77b66.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2460.067651] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 9227cc56-d55d-44ba-a619-63533dd77b66] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2460.067890] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-661950b1-00b5-444b-8e6e-d0e8e457ae79 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2460.072170] env[62824]: DEBUG oslo_vmware.api [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147027, 'name': Destroy_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2460.076496] env[62824]: DEBUG oslo_vmware.api [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2460.076496] env[62824]: value = "task-2147028" [ 2460.076496] env[62824]: _type = "Task" [ 2460.076496] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2460.085282] env[62824]: DEBUG oslo_vmware.api [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2147028, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2460.265845] env[62824]: DEBUG nova.compute.manager [req-9ba8999e-d97e-41b7-8a35-70be0793dfeb req-86276ea6-cbc5-4de6-8d17-97f697f0282d service nova] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Received event network-vif-plugged-ae875573-c95e-4613-b343-e3e5e3829964 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2460.265845] env[62824]: DEBUG oslo_concurrency.lockutils [req-9ba8999e-d97e-41b7-8a35-70be0793dfeb req-86276ea6-cbc5-4de6-8d17-97f697f0282d service nova] Acquiring lock "80c57382-8f6f-4537-b210-70a902b02cb4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2460.266171] env[62824]: DEBUG oslo_concurrency.lockutils [req-9ba8999e-d97e-41b7-8a35-70be0793dfeb req-86276ea6-cbc5-4de6-8d17-97f697f0282d service nova] Lock "80c57382-8f6f-4537-b210-70a902b02cb4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2460.266171] env[62824]: DEBUG oslo_concurrency.lockutils [req-9ba8999e-d97e-41b7-8a35-70be0793dfeb req-86276ea6-cbc5-4de6-8d17-97f697f0282d service nova] Lock "80c57382-8f6f-4537-b210-70a902b02cb4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2460.266273] env[62824]: DEBUG nova.compute.manager [req-9ba8999e-d97e-41b7-8a35-70be0793dfeb req-86276ea6-cbc5-4de6-8d17-97f697f0282d service nova] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] No waiting events found dispatching network-vif-plugged-ae875573-c95e-4613-b343-e3e5e3829964 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2460.266421] env[62824]: WARNING nova.compute.manager [req-9ba8999e-d97e-41b7-8a35-70be0793dfeb req-86276ea6-cbc5-4de6-8d17-97f697f0282d service nova] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Received unexpected event network-vif-plugged-ae875573-c95e-4613-b343-e3e5e3829964 for instance with vm_state building and task_state spawning. [ 2460.359684] env[62824]: DEBUG nova.network.neutron [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Successfully updated port: ae875573-c95e-4613-b343-e3e5e3829964 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2460.574950] env[62824]: DEBUG oslo_vmware.api [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147027, 'name': Destroy_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2460.587799] env[62824]: DEBUG oslo_vmware.api [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2147028, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066082} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2460.588038] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 9227cc56-d55d-44ba-a619-63533dd77b66] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2460.588795] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fc36684-13ca-4c22-b096-21e777072383 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2460.614021] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 9227cc56-d55d-44ba-a619-63533dd77b66] Reconfiguring VM instance instance-00000073 to attach disk [datastore2] 9227cc56-d55d-44ba-a619-63533dd77b66/9227cc56-d55d-44ba-a619-63533dd77b66.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2460.618021] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-905a9be6-3e65-4964-8db1-3deb0ea2bc9b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2460.638218] env[62824]: DEBUG oslo_vmware.api [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2460.638218] env[62824]: value = "task-2147029" [ 2460.638218] env[62824]: _type = "Task" [ 2460.638218] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2460.648542] env[62824]: DEBUG oslo_vmware.api [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2147029, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2460.672261] env[62824]: DEBUG nova.network.neutron [req-05dfe276-8c43-4e29-bae2-07bfd3c89f0b req-8dfe626f-1d61-4c4e-b7b5-5d910bb1c483 service nova] [instance: 4dff2a76-b0a2-438c-b52f-d6495948d041] Updated VIF entry in instance network info cache for port 2e0a71bd-c461-4fe9-a810-fe72908cf629. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2460.672701] env[62824]: DEBUG nova.network.neutron [req-05dfe276-8c43-4e29-bae2-07bfd3c89f0b req-8dfe626f-1d61-4c4e-b7b5-5d910bb1c483 service nova] [instance: 4dff2a76-b0a2-438c-b52f-d6495948d041] Updating instance_info_cache with network_info: [{"id": "2e0a71bd-c461-4fe9-a810-fe72908cf629", "address": "fa:16:3e:29:2e:10", "network": {"id": "b4ed285c-26a7-461b-8a71-d18e07582ab8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-997150974-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "815ecf9340934733a83d4dcb26612fd4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69f65356-c85e-4b7f-ad28-7c7b5e8cf50c", "external-id": "nsx-vlan-transportzone-281", "segmentation_id": 281, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2e0a71bd-c4", "ovs_interfaceid": "2e0a71bd-c461-4fe9-a810-fe72908cf629", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2460.863035] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquiring lock "refresh_cache-80c57382-8f6f-4537-b210-70a902b02cb4" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2460.863150] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquired lock "refresh_cache-80c57382-8f6f-4537-b210-70a902b02cb4" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2460.863231] env[62824]: DEBUG nova.network.neutron [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2461.073400] env[62824]: DEBUG oslo_vmware.api [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147027, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2461.147325] env[62824]: DEBUG oslo_vmware.api [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2147029, 'name': ReconfigVM_Task, 'duration_secs': 0.27992} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2461.147564] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 9227cc56-d55d-44ba-a619-63533dd77b66] Reconfigured VM instance instance-00000073 to attach disk [datastore2] 9227cc56-d55d-44ba-a619-63533dd77b66/9227cc56-d55d-44ba-a619-63533dd77b66.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2461.148208] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-17b3d784-51b7-42bd-9dab-994d91cfcb7d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2461.154469] env[62824]: DEBUG oslo_vmware.api [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2461.154469] env[62824]: value = "task-2147031" [ 2461.154469] env[62824]: _type = "Task" [ 2461.154469] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2461.161731] env[62824]: DEBUG oslo_vmware.api [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2147031, 'name': Rename_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2461.175322] env[62824]: DEBUG oslo_concurrency.lockutils [req-05dfe276-8c43-4e29-bae2-07bfd3c89f0b req-8dfe626f-1d61-4c4e-b7b5-5d910bb1c483 service nova] Releasing lock "refresh_cache-4dff2a76-b0a2-438c-b52f-d6495948d041" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2461.397810] env[62824]: DEBUG nova.network.neutron [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2461.537355] env[62824]: DEBUG nova.network.neutron [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Updating instance_info_cache with network_info: [{"id": "ae875573-c95e-4613-b343-e3e5e3829964", "address": "fa:16:3e:ac:dd:e4", "network": {"id": "bba9b98c-c7eb-4890-8de5-eafacfdebd7d", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1928394728-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7fcdbfc29df64a1ba8d982bdcc667b64", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b49c5024-2ced-42ca-90cc-6066766d43e6", "external-id": "nsx-vlan-transportzone-239", "segmentation_id": 239, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapae875573-c9", "ovs_interfaceid": "ae875573-c95e-4613-b343-e3e5e3829964", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2461.572314] env[62824]: DEBUG oslo_vmware.api [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147027, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2461.665420] env[62824]: DEBUG oslo_vmware.api [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2147031, 'name': Rename_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2462.040644] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Releasing lock "refresh_cache-80c57382-8f6f-4537-b210-70a902b02cb4" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2462.040976] env[62824]: DEBUG nova.compute.manager [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Instance network_info: |[{"id": "ae875573-c95e-4613-b343-e3e5e3829964", "address": "fa:16:3e:ac:dd:e4", "network": {"id": "bba9b98c-c7eb-4890-8de5-eafacfdebd7d", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1928394728-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7fcdbfc29df64a1ba8d982bdcc667b64", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b49c5024-2ced-42ca-90cc-6066766d43e6", "external-id": "nsx-vlan-transportzone-239", "segmentation_id": 239, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapae875573-c9", "ovs_interfaceid": "ae875573-c95e-4613-b343-e3e5e3829964", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2462.041455] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ac:dd:e4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b49c5024-2ced-42ca-90cc-6066766d43e6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ae875573-c95e-4613-b343-e3e5e3829964', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2462.049015] env[62824]: DEBUG oslo.service.loopingcall [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2462.049299] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2462.049565] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f7457cfb-c2a9-409a-8f20-3d2bb777dbe7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2462.072729] env[62824]: DEBUG oslo_vmware.api [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147027, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2462.073877] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2462.073877] env[62824]: value = "task-2147032" [ 2462.073877] env[62824]: _type = "Task" [ 2462.073877] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2462.080985] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2147032, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2462.163899] env[62824]: DEBUG oslo_vmware.api [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2147031, 'name': Rename_Task, 'duration_secs': 0.848793} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2462.164140] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 9227cc56-d55d-44ba-a619-63533dd77b66] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2462.164409] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0481353f-e954-49f5-bf92-96d298218da1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2462.169631] env[62824]: DEBUG oslo_vmware.api [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2462.169631] env[62824]: value = "task-2147033" [ 2462.169631] env[62824]: _type = "Task" [ 2462.169631] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2462.176726] env[62824]: DEBUG oslo_vmware.api [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2147033, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2462.291172] env[62824]: DEBUG nova.compute.manager [req-7325e9ae-3403-4491-85e8-ace4418da24a req-b1c9b0af-6d83-49ba-bc87-6d1b81786a8f service nova] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Received event network-changed-ae875573-c95e-4613-b343-e3e5e3829964 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2462.291359] env[62824]: DEBUG nova.compute.manager [req-7325e9ae-3403-4491-85e8-ace4418da24a req-b1c9b0af-6d83-49ba-bc87-6d1b81786a8f service nova] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Refreshing instance network info cache due to event network-changed-ae875573-c95e-4613-b343-e3e5e3829964. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2462.291579] env[62824]: DEBUG oslo_concurrency.lockutils [req-7325e9ae-3403-4491-85e8-ace4418da24a req-b1c9b0af-6d83-49ba-bc87-6d1b81786a8f service nova] Acquiring lock "refresh_cache-80c57382-8f6f-4537-b210-70a902b02cb4" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2462.291725] env[62824]: DEBUG oslo_concurrency.lockutils [req-7325e9ae-3403-4491-85e8-ace4418da24a req-b1c9b0af-6d83-49ba-bc87-6d1b81786a8f service nova] Acquired lock "refresh_cache-80c57382-8f6f-4537-b210-70a902b02cb4" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2462.291889] env[62824]: DEBUG nova.network.neutron [req-7325e9ae-3403-4491-85e8-ace4418da24a req-b1c9b0af-6d83-49ba-bc87-6d1b81786a8f service nova] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Refreshing network info cache for port ae875573-c95e-4613-b343-e3e5e3829964 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2462.573017] env[62824]: DEBUG oslo_vmware.api [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147027, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2462.582500] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2147032, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2462.678426] env[62824]: DEBUG oslo_vmware.api [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2147033, 'name': PowerOnVM_Task, 'duration_secs': 0.456633} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2462.678689] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 9227cc56-d55d-44ba-a619-63533dd77b66] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2462.678892] env[62824]: INFO nova.compute.manager [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 9227cc56-d55d-44ba-a619-63533dd77b66] Took 8.68 seconds to spawn the instance on the hypervisor. [ 2462.679089] env[62824]: DEBUG nova.compute.manager [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 9227cc56-d55d-44ba-a619-63533dd77b66] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2462.679837] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37a3647e-0bb0-42a3-b611-58fdb790cc15 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2462.985151] env[62824]: DEBUG nova.network.neutron [req-7325e9ae-3403-4491-85e8-ace4418da24a req-b1c9b0af-6d83-49ba-bc87-6d1b81786a8f service nova] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Updated VIF entry in instance network info cache for port ae875573-c95e-4613-b343-e3e5e3829964. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2462.985326] env[62824]: DEBUG nova.network.neutron [req-7325e9ae-3403-4491-85e8-ace4418da24a req-b1c9b0af-6d83-49ba-bc87-6d1b81786a8f service nova] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Updating instance_info_cache with network_info: [{"id": "ae875573-c95e-4613-b343-e3e5e3829964", "address": "fa:16:3e:ac:dd:e4", "network": {"id": "bba9b98c-c7eb-4890-8de5-eafacfdebd7d", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1928394728-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7fcdbfc29df64a1ba8d982bdcc667b64", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b49c5024-2ced-42ca-90cc-6066766d43e6", "external-id": "nsx-vlan-transportzone-239", "segmentation_id": 239, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapae875573-c9", "ovs_interfaceid": "ae875573-c95e-4613-b343-e3e5e3829964", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2463.077842] env[62824]: DEBUG oslo_vmware.api [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147027, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2463.088746] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2147032, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2463.195487] env[62824]: INFO nova.compute.manager [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 9227cc56-d55d-44ba-a619-63533dd77b66] Took 13.53 seconds to build instance. [ 2463.488504] env[62824]: DEBUG oslo_concurrency.lockutils [req-7325e9ae-3403-4491-85e8-ace4418da24a req-b1c9b0af-6d83-49ba-bc87-6d1b81786a8f service nova] Releasing lock "refresh_cache-80c57382-8f6f-4537-b210-70a902b02cb4" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2463.588903] env[62824]: DEBUG oslo_vmware.api [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147027, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2463.594866] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2147032, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2463.698113] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0f2d4c45-4f04-48d3-9cc0-7df15833fbc0 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "9227cc56-d55d-44ba-a619-63533dd77b66" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.035s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2464.076764] env[62824]: DEBUG oslo_vmware.api [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147027, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2464.087286] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2147032, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2464.102973] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c113ee15-cb48-4c71-a453-64177260ee09 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquiring lock "9227cc56-d55d-44ba-a619-63533dd77b66" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2464.103358] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c113ee15-cb48-4c71-a453-64177260ee09 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "9227cc56-d55d-44ba-a619-63533dd77b66" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2464.103813] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c113ee15-cb48-4c71-a453-64177260ee09 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquiring lock "9227cc56-d55d-44ba-a619-63533dd77b66-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2464.103853] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c113ee15-cb48-4c71-a453-64177260ee09 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "9227cc56-d55d-44ba-a619-63533dd77b66-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2464.104087] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c113ee15-cb48-4c71-a453-64177260ee09 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "9227cc56-d55d-44ba-a619-63533dd77b66-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2464.106482] env[62824]: INFO nova.compute.manager [None req-c113ee15-cb48-4c71-a453-64177260ee09 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 9227cc56-d55d-44ba-a619-63533dd77b66] Terminating instance [ 2464.576689] env[62824]: DEBUG oslo_vmware.api [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147027, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2464.586750] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2147032, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2464.611548] env[62824]: DEBUG nova.compute.manager [None req-c113ee15-cb48-4c71-a453-64177260ee09 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 9227cc56-d55d-44ba-a619-63533dd77b66] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2464.611813] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c113ee15-cb48-4c71-a453-64177260ee09 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 9227cc56-d55d-44ba-a619-63533dd77b66] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2464.612641] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24d44cf2-2526-4137-aab1-544612ccb89b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2464.619216] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c113ee15-cb48-4c71-a453-64177260ee09 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 9227cc56-d55d-44ba-a619-63533dd77b66] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2464.619446] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b7519d72-d540-43a1-8a85-46c34381e7cf {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2464.624797] env[62824]: DEBUG oslo_vmware.api [None req-c113ee15-cb48-4c71-a453-64177260ee09 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2464.624797] env[62824]: value = "task-2147035" [ 2464.624797] env[62824]: _type = "Task" [ 2464.624797] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2464.632185] env[62824]: DEBUG oslo_vmware.api [None req-c113ee15-cb48-4c71-a453-64177260ee09 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2147035, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2465.077629] env[62824]: DEBUG oslo_vmware.api [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147027, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2465.087628] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2147032, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2465.136013] env[62824]: DEBUG oslo_vmware.api [None req-c113ee15-cb48-4c71-a453-64177260ee09 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2147035, 'name': PowerOffVM_Task, 'duration_secs': 0.156658} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2465.136257] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c113ee15-cb48-4c71-a453-64177260ee09 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 9227cc56-d55d-44ba-a619-63533dd77b66] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2465.136427] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c113ee15-cb48-4c71-a453-64177260ee09 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 9227cc56-d55d-44ba-a619-63533dd77b66] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2465.136682] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0a855393-0596-4960-88a1-9f753340b477 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2465.579099] env[62824]: DEBUG oslo_vmware.api [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147027, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2465.589286] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2147032, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2466.079759] env[62824]: DEBUG oslo_vmware.api [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147027, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2466.088247] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2147032, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2466.581688] env[62824]: DEBUG oslo_vmware.api [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147027, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2466.589824] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2147032, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2467.081851] env[62824]: DEBUG oslo_vmware.api [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147027, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2467.090169] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2147032, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2467.582488] env[62824]: DEBUG oslo_vmware.api [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147027, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2467.590820] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2147032, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2468.082603] env[62824]: DEBUG oslo_vmware.api [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147027, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2468.089847] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2147032, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2468.586043] env[62824]: DEBUG oslo_vmware.api [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147027, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2468.590698] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2147032, 'name': CreateVM_Task} progress is 25%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2468.978909] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c113ee15-cb48-4c71-a453-64177260ee09 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 9227cc56-d55d-44ba-a619-63533dd77b66] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2468.979108] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c113ee15-cb48-4c71-a453-64177260ee09 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 9227cc56-d55d-44ba-a619-63533dd77b66] Deleting contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2468.979269] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-c113ee15-cb48-4c71-a453-64177260ee09 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Deleting the datastore file [datastore2] 9227cc56-d55d-44ba-a619-63533dd77b66 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2468.979539] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2a0dbe6a-ace9-47fa-90eb-e570a3e5ba9c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2468.986452] env[62824]: DEBUG oslo_vmware.api [None req-c113ee15-cb48-4c71-a453-64177260ee09 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for the task: (returnval){ [ 2468.986452] env[62824]: value = "task-2147039" [ 2468.986452] env[62824]: _type = "Task" [ 2468.986452] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2468.994454] env[62824]: DEBUG oslo_vmware.api [None req-c113ee15-cb48-4c71-a453-64177260ee09 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2147039, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2469.083206] env[62824]: DEBUG oslo_vmware.api [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147027, 'name': Destroy_Task, 'duration_secs': 8.809862} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2469.083544] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Destroyed the VM [ 2469.083829] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Deleting Snapshot of the VM instance {{(pid=62824) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 2469.086844] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-b892c1ae-d119-416c-bd48-a5ac2ac53f72 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2469.092781] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2147032, 'name': CreateVM_Task, 'duration_secs': 6.922799} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2469.093808] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2469.094111] env[62824]: DEBUG oslo_vmware.api [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2469.094111] env[62824]: value = "task-2147040" [ 2469.094111] env[62824]: _type = "Task" [ 2469.094111] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2469.103161] env[62824]: DEBUG oslo_vmware.api [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147040, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2469.104207] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2469.104372] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2469.104686] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2469.104896] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-839abfad-cfbd-4187-a9f4-4288a3d2a69f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2469.109221] env[62824]: DEBUG oslo_vmware.api [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2469.109221] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52d05131-e765-c4fd-ca1c-e1bf8369857f" [ 2469.109221] env[62824]: _type = "Task" [ 2469.109221] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2469.118952] env[62824]: DEBUG oslo_vmware.api [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52d05131-e765-c4fd-ca1c-e1bf8369857f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2469.497563] env[62824]: DEBUG oslo_vmware.api [None req-c113ee15-cb48-4c71-a453-64177260ee09 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Task: {'id': task-2147039, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.130001} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2469.497782] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-c113ee15-cb48-4c71-a453-64177260ee09 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2469.498038] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c113ee15-cb48-4c71-a453-64177260ee09 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 9227cc56-d55d-44ba-a619-63533dd77b66] Deleted contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2469.498263] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c113ee15-cb48-4c71-a453-64177260ee09 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 9227cc56-d55d-44ba-a619-63533dd77b66] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2469.498450] env[62824]: INFO nova.compute.manager [None req-c113ee15-cb48-4c71-a453-64177260ee09 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] [instance: 9227cc56-d55d-44ba-a619-63533dd77b66] Took 4.89 seconds to destroy the instance on the hypervisor. [ 2469.498690] env[62824]: DEBUG oslo.service.loopingcall [None req-c113ee15-cb48-4c71-a453-64177260ee09 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2469.498880] env[62824]: DEBUG nova.compute.manager [-] [instance: 9227cc56-d55d-44ba-a619-63533dd77b66] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2469.498975] env[62824]: DEBUG nova.network.neutron [-] [instance: 9227cc56-d55d-44ba-a619-63533dd77b66] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2469.604708] env[62824]: DEBUG oslo_vmware.api [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147040, 'name': RemoveSnapshot_Task, 'duration_secs': 0.506484} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2469.605016] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Deleted Snapshot of the VM instance {{(pid=62824) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 2469.605222] env[62824]: INFO nova.compute.manager [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Took 31.63 seconds to snapshot the instance on the hypervisor. [ 2469.618744] env[62824]: DEBUG oslo_vmware.api [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52d05131-e765-c4fd-ca1c-e1bf8369857f, 'name': SearchDatastore_Task, 'duration_secs': 0.010239} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2469.619178] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2469.619279] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2469.619508] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2469.619654] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2469.619832] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2469.620104] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-31101338-2093-4341-a40b-4b7cbf8c3c17 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2469.629325] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2469.629502] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2469.630223] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-968b19c5-c52c-4c3d-bd64-2c14f8845237 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2469.635475] env[62824]: DEBUG oslo_vmware.api [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2469.635475] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52577c66-4818-a5b4-4e07-3a8fa75f2ddb" [ 2469.635475] env[62824]: _type = "Task" [ 2469.635475] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2469.643441] env[62824]: DEBUG oslo_vmware.api [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52577c66-4818-a5b4-4e07-3a8fa75f2ddb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2469.755196] env[62824]: DEBUG nova.compute.manager [req-f28bcd4f-8437-450c-ab01-aca02ddf0bca req-a1fb3841-533c-428e-9eac-32ed9d82c74d service nova] [instance: 9227cc56-d55d-44ba-a619-63533dd77b66] Received event network-vif-deleted-0be28a80-9cee-483c-9d58-c3e6e610b89d {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2469.755394] env[62824]: INFO nova.compute.manager [req-f28bcd4f-8437-450c-ab01-aca02ddf0bca req-a1fb3841-533c-428e-9eac-32ed9d82c74d service nova] [instance: 9227cc56-d55d-44ba-a619-63533dd77b66] Neutron deleted interface 0be28a80-9cee-483c-9d58-c3e6e610b89d; detaching it from the instance and deleting it from the info cache [ 2469.755604] env[62824]: DEBUG nova.network.neutron [req-f28bcd4f-8437-450c-ab01-aca02ddf0bca req-a1fb3841-533c-428e-9eac-32ed9d82c74d service nova] [instance: 9227cc56-d55d-44ba-a619-63533dd77b66] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2470.147602] env[62824]: DEBUG oslo_vmware.api [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52577c66-4818-a5b4-4e07-3a8fa75f2ddb, 'name': SearchDatastore_Task, 'duration_secs': 0.009014} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2470.149520] env[62824]: DEBUG nova.compute.manager [None req-1517c73e-877d-45a1-9f21-d311c0538c90 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Found 1 images (rotation: 2) {{(pid=62824) _rotate_backups /opt/stack/nova/nova/compute/manager.py:5017}} [ 2470.151153] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-96d6f446-001c-44b7-8d67-005ce56ddb57 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2470.156474] env[62824]: DEBUG oslo_vmware.api [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2470.156474] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]524d526c-7537-0323-de29-d0cc893da72c" [ 2470.156474] env[62824]: _type = "Task" [ 2470.156474] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2470.164969] env[62824]: DEBUG oslo_vmware.api [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]524d526c-7537-0323-de29-d0cc893da72c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2470.233008] env[62824]: DEBUG nova.network.neutron [-] [instance: 9227cc56-d55d-44ba-a619-63533dd77b66] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2470.258631] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-86317458-23e9-4c0c-a17c-a6c4fefa71a5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2470.273517] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f46912dc-20dd-4719-8223-c0f9763aa6f6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2470.306891] env[62824]: DEBUG nova.compute.manager [req-f28bcd4f-8437-450c-ab01-aca02ddf0bca req-a1fb3841-533c-428e-9eac-32ed9d82c74d service nova] [instance: 9227cc56-d55d-44ba-a619-63533dd77b66] Detach interface failed, port_id=0be28a80-9cee-483c-9d58-c3e6e610b89d, reason: Instance 9227cc56-d55d-44ba-a619-63533dd77b66 could not be found. {{(pid=62824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11482}} [ 2470.667823] env[62824]: DEBUG oslo_vmware.api [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]524d526c-7537-0323-de29-d0cc893da72c, 'name': SearchDatastore_Task, 'duration_secs': 0.020461} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2470.668189] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2470.668465] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 80c57382-8f6f-4537-b210-70a902b02cb4/80c57382-8f6f-4537-b210-70a902b02cb4.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2470.668731] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-93b9fda3-2782-4bc0-a3ba-77c939403f4d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2470.675075] env[62824]: DEBUG oslo_vmware.api [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2470.675075] env[62824]: value = "task-2147042" [ 2470.675075] env[62824]: _type = "Task" [ 2470.675075] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2470.683216] env[62824]: DEBUG oslo_vmware.api [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2147042, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2470.736028] env[62824]: INFO nova.compute.manager [-] [instance: 9227cc56-d55d-44ba-a619-63533dd77b66] Took 1.24 seconds to deallocate network for instance. [ 2470.768145] env[62824]: DEBUG nova.compute.manager [None req-bddc32e6-781f-4a06-ad72-e1c63ca8c2b8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2470.768145] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c96c878-9b13-458b-80a2-e7cb790c3112 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2471.185554] env[62824]: DEBUG oslo_vmware.api [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2147042, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.434189} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2471.185811] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 80c57382-8f6f-4537-b210-70a902b02cb4/80c57382-8f6f-4537-b210-70a902b02cb4.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2471.186042] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2471.186290] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-043726bb-c351-459d-9a54-8f75b8dc8e57 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2471.192188] env[62824]: DEBUG oslo_vmware.api [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2471.192188] env[62824]: value = "task-2147043" [ 2471.192188] env[62824]: _type = "Task" [ 2471.192188] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2471.199355] env[62824]: DEBUG oslo_vmware.api [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2147043, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2471.242514] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c113ee15-cb48-4c71-a453-64177260ee09 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2471.242755] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c113ee15-cb48-4c71-a453-64177260ee09 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2471.242976] env[62824]: DEBUG nova.objects.instance [None req-c113ee15-cb48-4c71-a453-64177260ee09 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lazy-loading 'resources' on Instance uuid 9227cc56-d55d-44ba-a619-63533dd77b66 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2471.279192] env[62824]: INFO nova.compute.manager [None req-bddc32e6-781f-4a06-ad72-e1c63ca8c2b8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] instance snapshotting [ 2471.279827] env[62824]: DEBUG nova.objects.instance [None req-bddc32e6-781f-4a06-ad72-e1c63ca8c2b8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lazy-loading 'flavor' on Instance uuid 7c47bb92-a58e-4537-a68a-289e2a5fc19a {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2471.702621] env[62824]: DEBUG oslo_vmware.api [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2147043, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062654} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2471.703049] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2471.703761] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3552dfd5-64c0-4eeb-8778-1c0d83147f82 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2471.727915] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Reconfiguring VM instance instance-00000074 to attach disk [datastore1] 80c57382-8f6f-4537-b210-70a902b02cb4/80c57382-8f6f-4537-b210-70a902b02cb4.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2471.728187] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a8429c10-c0fb-4500-9355-2868bb6f63b9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2471.749265] env[62824]: DEBUG oslo_vmware.api [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2471.749265] env[62824]: value = "task-2147044" [ 2471.749265] env[62824]: _type = "Task" [ 2471.749265] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2471.758627] env[62824]: DEBUG oslo_vmware.api [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2147044, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2471.785625] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-559010df-bf9a-4467-a790-abbbe0f28bfb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2471.806247] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9378508-ea1e-48ab-861c-5cb8f10041d4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2471.860945] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd0f88e7-32b5-4ab0-b361-3fea38dfda7f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2471.868530] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5dc96cc-9306-439d-a8a8-7cb2f78d5ff7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2471.898446] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90cc86a1-d7eb-4e78-aabe-2e6ae8dc217a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2471.906064] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1553d314-97bc-4525-9365-f7a62a04efec {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2471.919102] env[62824]: DEBUG nova.compute.provider_tree [None req-c113ee15-cb48-4c71-a453-64177260ee09 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2472.259988] env[62824]: DEBUG oslo_vmware.api [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2147044, 'name': ReconfigVM_Task, 'duration_secs': 0.341584} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2472.260277] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Reconfigured VM instance instance-00000074 to attach disk [datastore1] 80c57382-8f6f-4537-b210-70a902b02cb4/80c57382-8f6f-4537-b210-70a902b02cb4.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2472.260923] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-deb85cf5-e063-432e-945c-ac0f02689848 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2472.266754] env[62824]: DEBUG oslo_vmware.api [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2472.266754] env[62824]: value = "task-2147046" [ 2472.266754] env[62824]: _type = "Task" [ 2472.266754] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2472.273924] env[62824]: DEBUG oslo_vmware.api [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2147046, 'name': Rename_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2472.316214] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-bddc32e6-781f-4a06-ad72-e1c63ca8c2b8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Creating Snapshot of the VM instance {{(pid=62824) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 2472.316521] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-07f1cd97-c1a1-42ad-af6f-10e6fef99be7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2472.323181] env[62824]: DEBUG oslo_vmware.api [None req-bddc32e6-781f-4a06-ad72-e1c63ca8c2b8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2472.323181] env[62824]: value = "task-2147047" [ 2472.323181] env[62824]: _type = "Task" [ 2472.323181] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2472.332466] env[62824]: DEBUG oslo_vmware.api [None req-bddc32e6-781f-4a06-ad72-e1c63ca8c2b8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147047, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2472.422274] env[62824]: DEBUG nova.scheduler.client.report [None req-c113ee15-cb48-4c71-a453-64177260ee09 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2472.776686] env[62824]: DEBUG oslo_vmware.api [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2147046, 'name': Rename_Task, 'duration_secs': 0.18035} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2472.777046] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2472.777231] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-049d0960-2566-426d-8e3a-198fb023efb4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2472.783727] env[62824]: DEBUG oslo_vmware.api [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2472.783727] env[62824]: value = "task-2147048" [ 2472.783727] env[62824]: _type = "Task" [ 2472.783727] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2472.791470] env[62824]: DEBUG oslo_vmware.api [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2147048, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2472.832726] env[62824]: DEBUG oslo_vmware.api [None req-bddc32e6-781f-4a06-ad72-e1c63ca8c2b8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147047, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2472.927955] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c113ee15-cb48-4c71-a453-64177260ee09 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.685s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2472.948433] env[62824]: INFO nova.scheduler.client.report [None req-c113ee15-cb48-4c71-a453-64177260ee09 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Deleted allocations for instance 9227cc56-d55d-44ba-a619-63533dd77b66 [ 2473.293745] env[62824]: DEBUG oslo_vmware.api [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2147048, 'name': PowerOnVM_Task, 'duration_secs': 0.470877} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2473.294015] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2473.294222] env[62824]: INFO nova.compute.manager [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Took 13.29 seconds to spawn the instance on the hypervisor. [ 2473.294404] env[62824]: DEBUG nova.compute.manager [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2473.295145] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a69c605-a9fc-4731-a742-71ccceac6184 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2473.333414] env[62824]: DEBUG oslo_vmware.api [None req-bddc32e6-781f-4a06-ad72-e1c63ca8c2b8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147047, 'name': CreateSnapshot_Task, 'duration_secs': 0.581643} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2473.333702] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-bddc32e6-781f-4a06-ad72-e1c63ca8c2b8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Created Snapshot of the VM instance {{(pid=62824) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 2473.334426] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-211fc8f4-7601-41dd-a447-2096f639f8a8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2473.457335] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c113ee15-cb48-4c71-a453-64177260ee09 tempest-ServerDiskConfigTestJSON-1541291917 tempest-ServerDiskConfigTestJSON-1541291917-project-member] Lock "9227cc56-d55d-44ba-a619-63533dd77b66" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.354s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2473.812985] env[62824]: INFO nova.compute.manager [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Took 18.09 seconds to build instance. [ 2473.853691] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-bddc32e6-781f-4a06-ad72-e1c63ca8c2b8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Creating linked-clone VM from snapshot {{(pid=62824) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 2473.854012] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-a0bb7589-0004-43ae-91db-a831c01745f1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2473.862478] env[62824]: DEBUG oslo_vmware.api [None req-bddc32e6-781f-4a06-ad72-e1c63ca8c2b8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2473.862478] env[62824]: value = "task-2147049" [ 2473.862478] env[62824]: _type = "Task" [ 2473.862478] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2473.870616] env[62824]: DEBUG oslo_vmware.api [None req-bddc32e6-781f-4a06-ad72-e1c63ca8c2b8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147049, 'name': CloneVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2474.314867] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c0fa4d25-644d-4207-bba6-632d66da819b tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lock "80c57382-8f6f-4537-b210-70a902b02cb4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.604s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2474.372126] env[62824]: DEBUG oslo_vmware.api [None req-bddc32e6-781f-4a06-ad72-e1c63ca8c2b8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147049, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2474.724318] env[62824]: DEBUG nova.compute.manager [req-78058206-bc89-4da7-81b6-491be28e7d81 req-d1959190-b44b-427c-ae1c-3c4cfa4e629f service nova] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Received event network-changed-ae875573-c95e-4613-b343-e3e5e3829964 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2474.724496] env[62824]: DEBUG nova.compute.manager [req-78058206-bc89-4da7-81b6-491be28e7d81 req-d1959190-b44b-427c-ae1c-3c4cfa4e629f service nova] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Refreshing instance network info cache due to event network-changed-ae875573-c95e-4613-b343-e3e5e3829964. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2474.724736] env[62824]: DEBUG oslo_concurrency.lockutils [req-78058206-bc89-4da7-81b6-491be28e7d81 req-d1959190-b44b-427c-ae1c-3c4cfa4e629f service nova] Acquiring lock "refresh_cache-80c57382-8f6f-4537-b210-70a902b02cb4" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2474.724961] env[62824]: DEBUG oslo_concurrency.lockutils [req-78058206-bc89-4da7-81b6-491be28e7d81 req-d1959190-b44b-427c-ae1c-3c4cfa4e629f service nova] Acquired lock "refresh_cache-80c57382-8f6f-4537-b210-70a902b02cb4" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2474.725065] env[62824]: DEBUG nova.network.neutron [req-78058206-bc89-4da7-81b6-491be28e7d81 req-d1959190-b44b-427c-ae1c-3c4cfa4e629f service nova] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Refreshing network info cache for port ae875573-c95e-4613-b343-e3e5e3829964 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2474.875362] env[62824]: DEBUG oslo_vmware.api [None req-bddc32e6-781f-4a06-ad72-e1c63ca8c2b8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147049, 'name': CloneVM_Task} progress is 95%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2475.377031] env[62824]: DEBUG oslo_vmware.api [None req-bddc32e6-781f-4a06-ad72-e1c63ca8c2b8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147049, 'name': CloneVM_Task, 'duration_secs': 1.328437} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2475.378392] env[62824]: INFO nova.virt.vmwareapi.vmops [None req-bddc32e6-781f-4a06-ad72-e1c63ca8c2b8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Created linked-clone VM from snapshot [ 2475.379213] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91d99cbc-29eb-4b9a-a647-eb7bbf3bb4ac {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2475.389129] env[62824]: DEBUG nova.virt.vmwareapi.images [None req-bddc32e6-781f-4a06-ad72-e1c63ca8c2b8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Uploading image 99b2c0ee-0719-430c-a611-a1fb4b76d81d {{(pid=62824) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 2475.420436] env[62824]: DEBUG oslo_vmware.rw_handles [None req-bddc32e6-781f-4a06-ad72-e1c63ca8c2b8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 2475.420436] env[62824]: value = "vm-438785" [ 2475.420436] env[62824]: _type = "VirtualMachine" [ 2475.420436] env[62824]: }. {{(pid=62824) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 2475.420716] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-ef82e607-d739-426a-b7c9-2f10f5cfc4fd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2475.428074] env[62824]: DEBUG oslo_vmware.rw_handles [None req-bddc32e6-781f-4a06-ad72-e1c63ca8c2b8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lease: (returnval){ [ 2475.428074] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5292ac4d-39c9-92a6-de5d-957d35d34034" [ 2475.428074] env[62824]: _type = "HttpNfcLease" [ 2475.428074] env[62824]: } obtained for exporting VM: (result){ [ 2475.428074] env[62824]: value = "vm-438785" [ 2475.428074] env[62824]: _type = "VirtualMachine" [ 2475.428074] env[62824]: }. {{(pid=62824) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 2475.428457] env[62824]: DEBUG oslo_vmware.api [None req-bddc32e6-781f-4a06-ad72-e1c63ca8c2b8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the lease: (returnval){ [ 2475.428457] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5292ac4d-39c9-92a6-de5d-957d35d34034" [ 2475.428457] env[62824]: _type = "HttpNfcLease" [ 2475.428457] env[62824]: } to be ready. {{(pid=62824) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 2475.435077] env[62824]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 2475.435077] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5292ac4d-39c9-92a6-de5d-957d35d34034" [ 2475.435077] env[62824]: _type = "HttpNfcLease" [ 2475.435077] env[62824]: } is initializing. {{(pid=62824) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 2475.491741] env[62824]: DEBUG nova.network.neutron [req-78058206-bc89-4da7-81b6-491be28e7d81 req-d1959190-b44b-427c-ae1c-3c4cfa4e629f service nova] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Updated VIF entry in instance network info cache for port ae875573-c95e-4613-b343-e3e5e3829964. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2475.491999] env[62824]: DEBUG nova.network.neutron [req-78058206-bc89-4da7-81b6-491be28e7d81 req-d1959190-b44b-427c-ae1c-3c4cfa4e629f service nova] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Updating instance_info_cache with network_info: [{"id": "ae875573-c95e-4613-b343-e3e5e3829964", "address": "fa:16:3e:ac:dd:e4", "network": {"id": "bba9b98c-c7eb-4890-8de5-eafacfdebd7d", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1928394728-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.160", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7fcdbfc29df64a1ba8d982bdcc667b64", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b49c5024-2ced-42ca-90cc-6066766d43e6", "external-id": "nsx-vlan-transportzone-239", "segmentation_id": 239, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapae875573-c9", "ovs_interfaceid": "ae875573-c95e-4613-b343-e3e5e3829964", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2475.937824] env[62824]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 2475.937824] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5292ac4d-39c9-92a6-de5d-957d35d34034" [ 2475.937824] env[62824]: _type = "HttpNfcLease" [ 2475.937824] env[62824]: } is ready. {{(pid=62824) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 2475.938338] env[62824]: DEBUG oslo_vmware.rw_handles [None req-bddc32e6-781f-4a06-ad72-e1c63ca8c2b8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 2475.938338] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5292ac4d-39c9-92a6-de5d-957d35d34034" [ 2475.938338] env[62824]: _type = "HttpNfcLease" [ 2475.938338] env[62824]: }. {{(pid=62824) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 2475.939111] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-011955ef-e548-40ad-b2c3-438a6d93f12f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2475.947295] env[62824]: DEBUG oslo_vmware.rw_handles [None req-bddc32e6-781f-4a06-ad72-e1c63ca8c2b8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528a8e3e-cbbc-0873-dc89-187f9f076f88/disk-0.vmdk from lease info. {{(pid=62824) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 2475.947430] env[62824]: DEBUG oslo_vmware.rw_handles [None req-bddc32e6-781f-4a06-ad72-e1c63ca8c2b8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528a8e3e-cbbc-0873-dc89-187f9f076f88/disk-0.vmdk for reading. {{(pid=62824) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 2476.004695] env[62824]: DEBUG oslo_concurrency.lockutils [req-78058206-bc89-4da7-81b6-491be28e7d81 req-d1959190-b44b-427c-ae1c-3c4cfa4e629f service nova] Releasing lock "refresh_cache-80c57382-8f6f-4537-b210-70a902b02cb4" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2476.037281] env[62824]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-f8414abe-54c3-47bb-828f-e5cd863fd3cf {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2480.221567] env[62824]: DEBUG oslo_concurrency.lockutils [None req-27a25ddf-2019-4ba0-8b69-6e86c5f78f91 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquiring lock "e524ce78-6a63-4950-b727-ff72f8df4424" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2480.221941] env[62824]: DEBUG oslo_concurrency.lockutils [None req-27a25ddf-2019-4ba0-8b69-6e86c5f78f91 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "e524ce78-6a63-4950-b727-ff72f8df4424" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2480.724855] env[62824]: DEBUG nova.compute.manager [None req-27a25ddf-2019-4ba0-8b69-6e86c5f78f91 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2481.246621] env[62824]: DEBUG oslo_concurrency.lockutils [None req-27a25ddf-2019-4ba0-8b69-6e86c5f78f91 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2481.246925] env[62824]: DEBUG oslo_concurrency.lockutils [None req-27a25ddf-2019-4ba0-8b69-6e86c5f78f91 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2481.249018] env[62824]: INFO nova.compute.claims [None req-27a25ddf-2019-4ba0-8b69-6e86c5f78f91 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2481.898037] env[62824]: DEBUG oslo_concurrency.lockutils [None req-971a8579-d52f-40b0-a5b3-0cc936973627 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Acquiring lock "ad9f01dd-d719-41a2-971b-b12cd3e8bdf6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2481.898037] env[62824]: DEBUG oslo_concurrency.lockutils [None req-971a8579-d52f-40b0-a5b3-0cc936973627 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Lock "ad9f01dd-d719-41a2-971b-b12cd3e8bdf6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2481.898037] env[62824]: DEBUG oslo_concurrency.lockutils [None req-971a8579-d52f-40b0-a5b3-0cc936973627 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Acquiring lock "ad9f01dd-d719-41a2-971b-b12cd3e8bdf6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2481.898283] env[62824]: DEBUG oslo_concurrency.lockutils [None req-971a8579-d52f-40b0-a5b3-0cc936973627 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Lock "ad9f01dd-d719-41a2-971b-b12cd3e8bdf6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2481.898412] env[62824]: DEBUG oslo_concurrency.lockutils [None req-971a8579-d52f-40b0-a5b3-0cc936973627 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Lock "ad9f01dd-d719-41a2-971b-b12cd3e8bdf6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2481.900627] env[62824]: INFO nova.compute.manager [None req-971a8579-d52f-40b0-a5b3-0cc936973627 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: ad9f01dd-d719-41a2-971b-b12cd3e8bdf6] Terminating instance [ 2482.362424] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f87f76e7-0ed8-42da-9dda-948ffb61eb7b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2482.370513] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3c708b9-1cd9-49a4-80a0-67a22f0f70ad {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2482.400860] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-422fd561-425c-4a5b-a91d-d908db44443d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2482.404008] env[62824]: DEBUG nova.compute.manager [None req-971a8579-d52f-40b0-a5b3-0cc936973627 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: ad9f01dd-d719-41a2-971b-b12cd3e8bdf6] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2482.404232] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-971a8579-d52f-40b0-a5b3-0cc936973627 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: ad9f01dd-d719-41a2-971b-b12cd3e8bdf6] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2482.405015] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b916d97-e9c4-4b67-b76d-c41d1837b357 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2482.414519] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70ee8190-c494-4af2-914a-4aa69d975655 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2482.418587] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-971a8579-d52f-40b0-a5b3-0cc936973627 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: ad9f01dd-d719-41a2-971b-b12cd3e8bdf6] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2482.418848] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-87b38c3e-b347-45ea-bf46-7b2295b646f4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2482.431025] env[62824]: DEBUG nova.compute.provider_tree [None req-27a25ddf-2019-4ba0-8b69-6e86c5f78f91 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2482.433982] env[62824]: DEBUG oslo_vmware.api [None req-971a8579-d52f-40b0-a5b3-0cc936973627 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Waiting for the task: (returnval){ [ 2482.433982] env[62824]: value = "task-2147053" [ 2482.433982] env[62824]: _type = "Task" [ 2482.433982] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2482.441677] env[62824]: DEBUG oslo_vmware.api [None req-971a8579-d52f-40b0-a5b3-0cc936973627 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147053, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2482.935360] env[62824]: DEBUG nova.scheduler.client.report [None req-27a25ddf-2019-4ba0-8b69-6e86c5f78f91 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2482.948893] env[62824]: DEBUG oslo_vmware.api [None req-971a8579-d52f-40b0-a5b3-0cc936973627 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147053, 'name': PowerOffVM_Task, 'duration_secs': 0.304154} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2482.949730] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-971a8579-d52f-40b0-a5b3-0cc936973627 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: ad9f01dd-d719-41a2-971b-b12cd3e8bdf6] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2482.949888] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-971a8579-d52f-40b0-a5b3-0cc936973627 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: ad9f01dd-d719-41a2-971b-b12cd3e8bdf6] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2482.950246] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dae3d36f-8352-467a-9993-7e02b4794fb5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2482.954447] env[62824]: DEBUG oslo_vmware.rw_handles [None req-bddc32e6-781f-4a06-ad72-e1c63ca8c2b8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528a8e3e-cbbc-0873-dc89-187f9f076f88/disk-0.vmdk. {{(pid=62824) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 2482.955461] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8c4a564-ed50-4891-b253-13127b55463c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2482.960803] env[62824]: DEBUG oslo_vmware.rw_handles [None req-bddc32e6-781f-4a06-ad72-e1c63ca8c2b8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528a8e3e-cbbc-0873-dc89-187f9f076f88/disk-0.vmdk is in state: ready. {{(pid=62824) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 2482.960967] env[62824]: ERROR oslo_vmware.rw_handles [None req-bddc32e6-781f-4a06-ad72-e1c63ca8c2b8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528a8e3e-cbbc-0873-dc89-187f9f076f88/disk-0.vmdk due to incomplete transfer. [ 2482.961177] env[62824]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-1005798f-5d85-484b-8e17-2edcf9411287 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2482.969017] env[62824]: DEBUG oslo_vmware.rw_handles [None req-bddc32e6-781f-4a06-ad72-e1c63ca8c2b8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528a8e3e-cbbc-0873-dc89-187f9f076f88/disk-0.vmdk. {{(pid=62824) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 2482.969212] env[62824]: DEBUG nova.virt.vmwareapi.images [None req-bddc32e6-781f-4a06-ad72-e1c63ca8c2b8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Uploaded image 99b2c0ee-0719-430c-a611-a1fb4b76d81d to the Glance image server {{(pid=62824) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 2482.971298] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-bddc32e6-781f-4a06-ad72-e1c63ca8c2b8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Destroying the VM {{(pid=62824) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 2482.971521] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-d066c8d8-940f-4d5e-8f4e-b85631c7c54f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2482.977404] env[62824]: DEBUG oslo_vmware.api [None req-bddc32e6-781f-4a06-ad72-e1c63ca8c2b8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2482.977404] env[62824]: value = "task-2147055" [ 2482.977404] env[62824]: _type = "Task" [ 2482.977404] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2482.984904] env[62824]: DEBUG oslo_vmware.api [None req-bddc32e6-781f-4a06-ad72-e1c63ca8c2b8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147055, 'name': Destroy_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2483.442957] env[62824]: DEBUG oslo_concurrency.lockutils [None req-27a25ddf-2019-4ba0-8b69-6e86c5f78f91 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.196s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2483.443389] env[62824]: DEBUG nova.compute.manager [None req-27a25ddf-2019-4ba0-8b69-6e86c5f78f91 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2483.487814] env[62824]: DEBUG oslo_vmware.api [None req-bddc32e6-781f-4a06-ad72-e1c63ca8c2b8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147055, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2483.949166] env[62824]: DEBUG nova.compute.utils [None req-27a25ddf-2019-4ba0-8b69-6e86c5f78f91 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2483.951195] env[62824]: DEBUG nova.compute.manager [None req-27a25ddf-2019-4ba0-8b69-6e86c5f78f91 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2483.951366] env[62824]: DEBUG nova.network.neutron [None req-27a25ddf-2019-4ba0-8b69-6e86c5f78f91 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2483.987818] env[62824]: DEBUG oslo_vmware.api [None req-bddc32e6-781f-4a06-ad72-e1c63ca8c2b8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147055, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2483.989617] env[62824]: DEBUG nova.policy [None req-27a25ddf-2019-4ba0-8b69-6e86c5f78f91 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9cc01135021d4a90a718dc605ce98af2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '815ecf9340934733a83d4dcb26612fd4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2484.257366] env[62824]: DEBUG nova.network.neutron [None req-27a25ddf-2019-4ba0-8b69-6e86c5f78f91 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Successfully created port: 62aa7735-5027-4b47-917b-7b769eac4cee {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2484.452659] env[62824]: DEBUG nova.compute.manager [None req-27a25ddf-2019-4ba0-8b69-6e86c5f78f91 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2484.488519] env[62824]: DEBUG oslo_vmware.api [None req-bddc32e6-781f-4a06-ad72-e1c63ca8c2b8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147055, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2484.958502] env[62824]: INFO nova.virt.block_device [None req-27a25ddf-2019-4ba0-8b69-6e86c5f78f91 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Booting with volume 08d8bf83-f546-44c3-8c92-1efd79d0f46e at /dev/sda [ 2484.989255] env[62824]: DEBUG oslo_vmware.api [None req-bddc32e6-781f-4a06-ad72-e1c63ca8c2b8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147055, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2484.990340] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8ff1cfd6-1990-42fa-b59c-2c7db1f0a0c0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2484.998256] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3d48068-9e7f-4a3e-89e6-55c18f8c7a8a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2485.025411] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4874495b-579c-4ca9-91bf-6ab2d4af06f6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2485.033015] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b5a7369-0074-4581-90d4-c02fa5214240 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2485.062240] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a94f079-df79-4f39-85b3-67c6ee20fa05 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2485.068743] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbd3b613-dd3a-4dd2-a3e1-189c855f866a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2485.081287] env[62824]: DEBUG nova.virt.block_device [None req-27a25ddf-2019-4ba0-8b69-6e86c5f78f91 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Updating existing volume attachment record: 6107452e-8dd1-45bb-a109-8a6a0524e7ad {{(pid=62824) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 2485.490191] env[62824]: DEBUG oslo_vmware.api [None req-bddc32e6-781f-4a06-ad72-e1c63ca8c2b8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147055, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2485.991704] env[62824]: DEBUG oslo_vmware.api [None req-bddc32e6-781f-4a06-ad72-e1c63ca8c2b8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147055, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2486.492386] env[62824]: DEBUG oslo_vmware.api [None req-bddc32e6-781f-4a06-ad72-e1c63ca8c2b8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147055, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2486.992522] env[62824]: DEBUG oslo_vmware.api [None req-bddc32e6-781f-4a06-ad72-e1c63ca8c2b8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147055, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2487.162144] env[62824]: DEBUG nova.compute.manager [None req-27a25ddf-2019-4ba0-8b69-6e86c5f78f91 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2487.162752] env[62824]: DEBUG nova.virt.hardware [None req-27a25ddf-2019-4ba0-8b69-6e86c5f78f91 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2487.162979] env[62824]: DEBUG nova.virt.hardware [None req-27a25ddf-2019-4ba0-8b69-6e86c5f78f91 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2487.163151] env[62824]: DEBUG nova.virt.hardware [None req-27a25ddf-2019-4ba0-8b69-6e86c5f78f91 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2487.163334] env[62824]: DEBUG nova.virt.hardware [None req-27a25ddf-2019-4ba0-8b69-6e86c5f78f91 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2487.163480] env[62824]: DEBUG nova.virt.hardware [None req-27a25ddf-2019-4ba0-8b69-6e86c5f78f91 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2487.163698] env[62824]: DEBUG nova.virt.hardware [None req-27a25ddf-2019-4ba0-8b69-6e86c5f78f91 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2487.163949] env[62824]: DEBUG nova.virt.hardware [None req-27a25ddf-2019-4ba0-8b69-6e86c5f78f91 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2487.164135] env[62824]: DEBUG nova.virt.hardware [None req-27a25ddf-2019-4ba0-8b69-6e86c5f78f91 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2487.164309] env[62824]: DEBUG nova.virt.hardware [None req-27a25ddf-2019-4ba0-8b69-6e86c5f78f91 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2487.164477] env[62824]: DEBUG nova.virt.hardware [None req-27a25ddf-2019-4ba0-8b69-6e86c5f78f91 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2487.164655] env[62824]: DEBUG nova.virt.hardware [None req-27a25ddf-2019-4ba0-8b69-6e86c5f78f91 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2487.165544] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38090e9a-73c8-4f40-86de-ba7a090e1ab8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2487.173656] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b57aadb-843c-43f7-88d4-c3feb517a8bf {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2487.492650] env[62824]: DEBUG oslo_vmware.api [None req-bddc32e6-781f-4a06-ad72-e1c63ca8c2b8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147055, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2487.995528] env[62824]: DEBUG oslo_vmware.api [None req-bddc32e6-781f-4a06-ad72-e1c63ca8c2b8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147055, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2488.494199] env[62824]: DEBUG oslo_vmware.api [None req-bddc32e6-781f-4a06-ad72-e1c63ca8c2b8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147055, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2488.995137] env[62824]: DEBUG oslo_vmware.api [None req-bddc32e6-781f-4a06-ad72-e1c63ca8c2b8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147055, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2489.247955] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-971a8579-d52f-40b0-a5b3-0cc936973627 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: ad9f01dd-d719-41a2-971b-b12cd3e8bdf6] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2489.248205] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-971a8579-d52f-40b0-a5b3-0cc936973627 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: ad9f01dd-d719-41a2-971b-b12cd3e8bdf6] Deleting contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2489.248403] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-971a8579-d52f-40b0-a5b3-0cc936973627 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Deleting the datastore file [datastore1] ad9f01dd-d719-41a2-971b-b12cd3e8bdf6 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2489.248676] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b7ab280d-9c75-432a-abcd-9abacac44cff {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2489.255311] env[62824]: DEBUG oslo_vmware.api [None req-971a8579-d52f-40b0-a5b3-0cc936973627 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Waiting for the task: (returnval){ [ 2489.255311] env[62824]: value = "task-2147056" [ 2489.255311] env[62824]: _type = "Task" [ 2489.255311] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2489.262948] env[62824]: DEBUG oslo_vmware.api [None req-971a8579-d52f-40b0-a5b3-0cc936973627 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147056, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2489.496337] env[62824]: DEBUG oslo_vmware.api [None req-bddc32e6-781f-4a06-ad72-e1c63ca8c2b8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147055, 'name': Destroy_Task, 'duration_secs': 6.390209} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2489.496634] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-bddc32e6-781f-4a06-ad72-e1c63ca8c2b8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Destroyed the VM [ 2489.496856] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-bddc32e6-781f-4a06-ad72-e1c63ca8c2b8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Deleting Snapshot of the VM instance {{(pid=62824) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 2489.497120] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-9c85eac8-27c0-4c68-ae9b-24455445f4d6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2489.504382] env[62824]: DEBUG oslo_vmware.api [None req-bddc32e6-781f-4a06-ad72-e1c63ca8c2b8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2489.504382] env[62824]: value = "task-2147057" [ 2489.504382] env[62824]: _type = "Task" [ 2489.504382] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2489.512238] env[62824]: DEBUG oslo_vmware.api [None req-bddc32e6-781f-4a06-ad72-e1c63ca8c2b8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147057, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2489.765009] env[62824]: DEBUG oslo_vmware.api [None req-971a8579-d52f-40b0-a5b3-0cc936973627 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147056, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.149693} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2489.765306] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-971a8579-d52f-40b0-a5b3-0cc936973627 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2489.765494] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-971a8579-d52f-40b0-a5b3-0cc936973627 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: ad9f01dd-d719-41a2-971b-b12cd3e8bdf6] Deleted contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2489.765673] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-971a8579-d52f-40b0-a5b3-0cc936973627 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: ad9f01dd-d719-41a2-971b-b12cd3e8bdf6] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2489.765886] env[62824]: INFO nova.compute.manager [None req-971a8579-d52f-40b0-a5b3-0cc936973627 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: ad9f01dd-d719-41a2-971b-b12cd3e8bdf6] Took 7.36 seconds to destroy the instance on the hypervisor. [ 2489.766098] env[62824]: DEBUG oslo.service.loopingcall [None req-971a8579-d52f-40b0-a5b3-0cc936973627 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2489.766290] env[62824]: DEBUG nova.compute.manager [-] [instance: ad9f01dd-d719-41a2-971b-b12cd3e8bdf6] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2489.766387] env[62824]: DEBUG nova.network.neutron [-] [instance: ad9f01dd-d719-41a2-971b-b12cd3e8bdf6] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2490.014490] env[62824]: DEBUG oslo_vmware.api [None req-bddc32e6-781f-4a06-ad72-e1c63ca8c2b8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147057, 'name': RemoveSnapshot_Task, 'duration_secs': 0.448468} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2490.014751] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-bddc32e6-781f-4a06-ad72-e1c63ca8c2b8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Deleted Snapshot of the VM instance {{(pid=62824) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 2490.014987] env[62824]: INFO nova.compute.manager [None req-bddc32e6-781f-4a06-ad72-e1c63ca8c2b8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Took 18.23 seconds to snapshot the instance on the hypervisor. [ 2490.274217] env[62824]: DEBUG nova.compute.manager [req-12ca3ab0-ca81-4ab5-83fa-432995ed15b7 req-b8c52be7-0d5e-4fe8-948a-66956f3fdb4e service nova] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Received event network-vif-plugged-62aa7735-5027-4b47-917b-7b769eac4cee {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2490.274792] env[62824]: DEBUG oslo_concurrency.lockutils [req-12ca3ab0-ca81-4ab5-83fa-432995ed15b7 req-b8c52be7-0d5e-4fe8-948a-66956f3fdb4e service nova] Acquiring lock "e524ce78-6a63-4950-b727-ff72f8df4424-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2490.275084] env[62824]: DEBUG oslo_concurrency.lockutils [req-12ca3ab0-ca81-4ab5-83fa-432995ed15b7 req-b8c52be7-0d5e-4fe8-948a-66956f3fdb4e service nova] Lock "e524ce78-6a63-4950-b727-ff72f8df4424-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2490.275279] env[62824]: DEBUG oslo_concurrency.lockutils [req-12ca3ab0-ca81-4ab5-83fa-432995ed15b7 req-b8c52be7-0d5e-4fe8-948a-66956f3fdb4e service nova] Lock "e524ce78-6a63-4950-b727-ff72f8df4424-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2490.275458] env[62824]: DEBUG nova.compute.manager [req-12ca3ab0-ca81-4ab5-83fa-432995ed15b7 req-b8c52be7-0d5e-4fe8-948a-66956f3fdb4e service nova] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] No waiting events found dispatching network-vif-plugged-62aa7735-5027-4b47-917b-7b769eac4cee {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2490.275627] env[62824]: WARNING nova.compute.manager [req-12ca3ab0-ca81-4ab5-83fa-432995ed15b7 req-b8c52be7-0d5e-4fe8-948a-66956f3fdb4e service nova] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Received unexpected event network-vif-plugged-62aa7735-5027-4b47-917b-7b769eac4cee for instance with vm_state building and task_state spawning. [ 2490.419686] env[62824]: DEBUG nova.network.neutron [None req-27a25ddf-2019-4ba0-8b69-6e86c5f78f91 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Successfully updated port: 62aa7735-5027-4b47-917b-7b769eac4cee {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2490.579402] env[62824]: DEBUG nova.compute.manager [None req-bddc32e6-781f-4a06-ad72-e1c63ca8c2b8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Found 2 images (rotation: 2) {{(pid=62824) _rotate_backups /opt/stack/nova/nova/compute/manager.py:5017}} [ 2490.864043] env[62824]: DEBUG nova.network.neutron [-] [instance: ad9f01dd-d719-41a2-971b-b12cd3e8bdf6] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2490.922181] env[62824]: DEBUG oslo_concurrency.lockutils [None req-27a25ddf-2019-4ba0-8b69-6e86c5f78f91 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquiring lock "refresh_cache-e524ce78-6a63-4950-b727-ff72f8df4424" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2490.922355] env[62824]: DEBUG oslo_concurrency.lockutils [None req-27a25ddf-2019-4ba0-8b69-6e86c5f78f91 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquired lock "refresh_cache-e524ce78-6a63-4950-b727-ff72f8df4424" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2490.922615] env[62824]: DEBUG nova.network.neutron [None req-27a25ddf-2019-4ba0-8b69-6e86c5f78f91 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2491.231412] env[62824]: DEBUG nova.compute.manager [None req-11b8fbfb-ae9d-4426-8f86-5f4c2fe088f8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2491.232290] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b947f59-da56-4928-911e-329e12b86234 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2491.366490] env[62824]: INFO nova.compute.manager [-] [instance: ad9f01dd-d719-41a2-971b-b12cd3e8bdf6] Took 1.60 seconds to deallocate network for instance. [ 2491.457729] env[62824]: DEBUG nova.network.neutron [None req-27a25ddf-2019-4ba0-8b69-6e86c5f78f91 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2491.595782] env[62824]: DEBUG nova.network.neutron [None req-27a25ddf-2019-4ba0-8b69-6e86c5f78f91 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Updating instance_info_cache with network_info: [{"id": "62aa7735-5027-4b47-917b-7b769eac4cee", "address": "fa:16:3e:61:22:c9", "network": {"id": "b4ed285c-26a7-461b-8a71-d18e07582ab8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-997150974-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "815ecf9340934733a83d4dcb26612fd4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69f65356-c85e-4b7f-ad28-7c7b5e8cf50c", "external-id": "nsx-vlan-transportzone-281", "segmentation_id": 281, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap62aa7735-50", "ovs_interfaceid": "62aa7735-5027-4b47-917b-7b769eac4cee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2491.743483] env[62824]: INFO nova.compute.manager [None req-11b8fbfb-ae9d-4426-8f86-5f4c2fe088f8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] instance snapshotting [ 2491.744140] env[62824]: DEBUG nova.objects.instance [None req-11b8fbfb-ae9d-4426-8f86-5f4c2fe088f8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lazy-loading 'flavor' on Instance uuid 7c47bb92-a58e-4537-a68a-289e2a5fc19a {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2491.872255] env[62824]: DEBUG oslo_concurrency.lockutils [None req-971a8579-d52f-40b0-a5b3-0cc936973627 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2491.872534] env[62824]: DEBUG oslo_concurrency.lockutils [None req-971a8579-d52f-40b0-a5b3-0cc936973627 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2491.872797] env[62824]: DEBUG nova.objects.instance [None req-971a8579-d52f-40b0-a5b3-0cc936973627 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Lazy-loading 'resources' on Instance uuid ad9f01dd-d719-41a2-971b-b12cd3e8bdf6 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2492.099042] env[62824]: DEBUG oslo_concurrency.lockutils [None req-27a25ddf-2019-4ba0-8b69-6e86c5f78f91 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Releasing lock "refresh_cache-e524ce78-6a63-4950-b727-ff72f8df4424" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2492.099406] env[62824]: DEBUG nova.compute.manager [None req-27a25ddf-2019-4ba0-8b69-6e86c5f78f91 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Instance network_info: |[{"id": "62aa7735-5027-4b47-917b-7b769eac4cee", "address": "fa:16:3e:61:22:c9", "network": {"id": "b4ed285c-26a7-461b-8a71-d18e07582ab8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-997150974-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "815ecf9340934733a83d4dcb26612fd4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69f65356-c85e-4b7f-ad28-7c7b5e8cf50c", "external-id": "nsx-vlan-transportzone-281", "segmentation_id": 281, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap62aa7735-50", "ovs_interfaceid": "62aa7735-5027-4b47-917b-7b769eac4cee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2492.099892] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-27a25ddf-2019-4ba0-8b69-6e86c5f78f91 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:61:22:c9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '69f65356-c85e-4b7f-ad28-7c7b5e8cf50c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '62aa7735-5027-4b47-917b-7b769eac4cee', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2492.108315] env[62824]: DEBUG oslo.service.loopingcall [None req-27a25ddf-2019-4ba0-8b69-6e86c5f78f91 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2492.108601] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2492.108893] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-11881e76-33e3-4c02-8af3-971cea521fa1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2492.131860] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2492.131860] env[62824]: value = "task-2147058" [ 2492.131860] env[62824]: _type = "Task" [ 2492.131860] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2492.139500] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2147058, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2492.250190] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-438d6f18-aa40-4304-97e3-de178b5307fa {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2492.269907] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a295af4-4919-4d5a-a85d-8229dbaf0857 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2492.308300] env[62824]: DEBUG nova.compute.manager [req-0a37df93-4ebd-43da-8775-a95ef8fbe5d2 req-08ccb0e2-c4c7-4af2-b303-32cb4e4c6940 service nova] [instance: ad9f01dd-d719-41a2-971b-b12cd3e8bdf6] Received event network-vif-deleted-ffa86a2f-846f-4178-8ace-b502f5cb81bb {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2492.308497] env[62824]: DEBUG nova.compute.manager [req-0a37df93-4ebd-43da-8775-a95ef8fbe5d2 req-08ccb0e2-c4c7-4af2-b303-32cb4e4c6940 service nova] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Received event network-changed-62aa7735-5027-4b47-917b-7b769eac4cee {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2492.308672] env[62824]: DEBUG nova.compute.manager [req-0a37df93-4ebd-43da-8775-a95ef8fbe5d2 req-08ccb0e2-c4c7-4af2-b303-32cb4e4c6940 service nova] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Refreshing instance network info cache due to event network-changed-62aa7735-5027-4b47-917b-7b769eac4cee. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2492.308886] env[62824]: DEBUG oslo_concurrency.lockutils [req-0a37df93-4ebd-43da-8775-a95ef8fbe5d2 req-08ccb0e2-c4c7-4af2-b303-32cb4e4c6940 service nova] Acquiring lock "refresh_cache-e524ce78-6a63-4950-b727-ff72f8df4424" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2492.309041] env[62824]: DEBUG oslo_concurrency.lockutils [req-0a37df93-4ebd-43da-8775-a95ef8fbe5d2 req-08ccb0e2-c4c7-4af2-b303-32cb4e4c6940 service nova] Acquired lock "refresh_cache-e524ce78-6a63-4950-b727-ff72f8df4424" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2492.309206] env[62824]: DEBUG nova.network.neutron [req-0a37df93-4ebd-43da-8775-a95ef8fbe5d2 req-08ccb0e2-c4c7-4af2-b303-32cb4e4c6940 service nova] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Refreshing network info cache for port 62aa7735-5027-4b47-917b-7b769eac4cee {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2492.472429] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b191a899-96f2-46da-b36b-0139758b7aee {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2492.480567] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f922e4ed-1e1d-4eb2-ba5b-abf5ff633668 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2492.512273] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7a2ee3d-d120-4137-a041-ff82a9ceb085 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2492.519575] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05317e70-938f-47eb-a435-06be9773c8c5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2492.532947] env[62824]: DEBUG nova.compute.provider_tree [None req-971a8579-d52f-40b0-a5b3-0cc936973627 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2492.641371] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2147058, 'name': CreateVM_Task, 'duration_secs': 0.4387} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2492.641741] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2492.642245] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-27a25ddf-2019-4ba0-8b69-6e86c5f78f91 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'device_type': None, 'boot_index': 0, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-438783', 'volume_id': '08d8bf83-f546-44c3-8c92-1efd79d0f46e', 'name': 'volume-08d8bf83-f546-44c3-8c92-1efd79d0f46e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'e524ce78-6a63-4950-b727-ff72f8df4424', 'attached_at': '', 'detached_at': '', 'volume_id': '08d8bf83-f546-44c3-8c92-1efd79d0f46e', 'serial': '08d8bf83-f546-44c3-8c92-1efd79d0f46e'}, 'disk_bus': None, 'mount_device': '/dev/sda', 'delete_on_termination': True, 'attachment_id': '6107452e-8dd1-45bb-a109-8a6a0524e7ad', 'guest_format': None, 'volume_type': None}], 'swap': None} {{(pid=62824) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 2492.642460] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-27a25ddf-2019-4ba0-8b69-6e86c5f78f91 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Root volume attach. Driver type: vmdk {{(pid=62824) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 2492.643260] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff33282e-f49a-4269-82e1-4273af9ee1df {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2492.650474] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8bbe1a2-b3c0-4fe8-bc63-3986c2dccc5c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2492.656189] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ab2efc8-f72f-4092-9e8a-2b2ef06b8c1c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2492.661570] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-09276755-f012-445a-a460-2580288ff0fa {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2492.667545] env[62824]: DEBUG oslo_vmware.api [None req-27a25ddf-2019-4ba0-8b69-6e86c5f78f91 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2492.667545] env[62824]: value = "task-2147059" [ 2492.667545] env[62824]: _type = "Task" [ 2492.667545] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2492.675086] env[62824]: DEBUG oslo_vmware.api [None req-27a25ddf-2019-4ba0-8b69-6e86c5f78f91 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2147059, 'name': RelocateVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2492.780809] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-11b8fbfb-ae9d-4426-8f86-5f4c2fe088f8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Creating Snapshot of the VM instance {{(pid=62824) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 2492.781318] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-19acba3f-be26-4366-ada0-bac48a3c73ec {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2492.789242] env[62824]: DEBUG oslo_vmware.api [None req-11b8fbfb-ae9d-4426-8f86-5f4c2fe088f8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2492.789242] env[62824]: value = "task-2147060" [ 2492.789242] env[62824]: _type = "Task" [ 2492.789242] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2492.797137] env[62824]: DEBUG oslo_vmware.api [None req-11b8fbfb-ae9d-4426-8f86-5f4c2fe088f8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147060, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2493.009690] env[62824]: DEBUG nova.network.neutron [req-0a37df93-4ebd-43da-8775-a95ef8fbe5d2 req-08ccb0e2-c4c7-4af2-b303-32cb4e4c6940 service nova] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Updated VIF entry in instance network info cache for port 62aa7735-5027-4b47-917b-7b769eac4cee. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2493.010129] env[62824]: DEBUG nova.network.neutron [req-0a37df93-4ebd-43da-8775-a95ef8fbe5d2 req-08ccb0e2-c4c7-4af2-b303-32cb4e4c6940 service nova] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Updating instance_info_cache with network_info: [{"id": "62aa7735-5027-4b47-917b-7b769eac4cee", "address": "fa:16:3e:61:22:c9", "network": {"id": "b4ed285c-26a7-461b-8a71-d18e07582ab8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-997150974-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "815ecf9340934733a83d4dcb26612fd4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69f65356-c85e-4b7f-ad28-7c7b5e8cf50c", "external-id": "nsx-vlan-transportzone-281", "segmentation_id": 281, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap62aa7735-50", "ovs_interfaceid": "62aa7735-5027-4b47-917b-7b769eac4cee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2493.038015] env[62824]: DEBUG nova.scheduler.client.report [None req-971a8579-d52f-40b0-a5b3-0cc936973627 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2493.178991] env[62824]: DEBUG oslo_vmware.api [None req-27a25ddf-2019-4ba0-8b69-6e86c5f78f91 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2147059, 'name': RelocateVM_Task, 'duration_secs': 0.356764} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2493.179277] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-27a25ddf-2019-4ba0-8b69-6e86c5f78f91 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Volume attach. Driver type: vmdk {{(pid=62824) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 2493.179480] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-27a25ddf-2019-4ba0-8b69-6e86c5f78f91 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-438783', 'volume_id': '08d8bf83-f546-44c3-8c92-1efd79d0f46e', 'name': 'volume-08d8bf83-f546-44c3-8c92-1efd79d0f46e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'e524ce78-6a63-4950-b727-ff72f8df4424', 'attached_at': '', 'detached_at': '', 'volume_id': '08d8bf83-f546-44c3-8c92-1efd79d0f46e', 'serial': '08d8bf83-f546-44c3-8c92-1efd79d0f46e'} {{(pid=62824) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 2493.180244] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e1745cd-6faa-4f3a-bfa5-c3c8f187c667 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2493.196132] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abe87321-4118-43f2-806a-590353a5e9b0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2493.217614] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-27a25ddf-2019-4ba0-8b69-6e86c5f78f91 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Reconfiguring VM instance instance-00000075 to attach disk [datastore2] volume-08d8bf83-f546-44c3-8c92-1efd79d0f46e/volume-08d8bf83-f546-44c3-8c92-1efd79d0f46e.vmdk or device None with type thin {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2493.217851] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f6d2e155-d64d-4657-8cfd-1b27fcb09264 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2493.236632] env[62824]: DEBUG oslo_vmware.api [None req-27a25ddf-2019-4ba0-8b69-6e86c5f78f91 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2493.236632] env[62824]: value = "task-2147061" [ 2493.236632] env[62824]: _type = "Task" [ 2493.236632] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2493.243766] env[62824]: DEBUG oslo_vmware.api [None req-27a25ddf-2019-4ba0-8b69-6e86c5f78f91 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2147061, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2493.298722] env[62824]: DEBUG oslo_vmware.api [None req-11b8fbfb-ae9d-4426-8f86-5f4c2fe088f8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147060, 'name': CreateSnapshot_Task, 'duration_secs': 0.430397} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2493.298937] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-11b8fbfb-ae9d-4426-8f86-5f4c2fe088f8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Created Snapshot of the VM instance {{(pid=62824) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 2493.299705] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-880d06c2-8684-4fa1-91f9-f3238a40e2b1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2493.512688] env[62824]: DEBUG oslo_concurrency.lockutils [req-0a37df93-4ebd-43da-8775-a95ef8fbe5d2 req-08ccb0e2-c4c7-4af2-b303-32cb4e4c6940 service nova] Releasing lock "refresh_cache-e524ce78-6a63-4950-b727-ff72f8df4424" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2493.540927] env[62824]: DEBUG oslo_concurrency.lockutils [None req-971a8579-d52f-40b0-a5b3-0cc936973627 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.668s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2493.561598] env[62824]: INFO nova.scheduler.client.report [None req-971a8579-d52f-40b0-a5b3-0cc936973627 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Deleted allocations for instance ad9f01dd-d719-41a2-971b-b12cd3e8bdf6 [ 2493.746621] env[62824]: DEBUG oslo_vmware.api [None req-27a25ddf-2019-4ba0-8b69-6e86c5f78f91 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2147061, 'name': ReconfigVM_Task, 'duration_secs': 0.244694} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2493.746910] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-27a25ddf-2019-4ba0-8b69-6e86c5f78f91 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Reconfigured VM instance instance-00000075 to attach disk [datastore2] volume-08d8bf83-f546-44c3-8c92-1efd79d0f46e/volume-08d8bf83-f546-44c3-8c92-1efd79d0f46e.vmdk or device None with type thin {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2493.751646] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-985cb662-ee80-4504-93b2-289ea52c25c0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2493.766131] env[62824]: DEBUG oslo_vmware.api [None req-27a25ddf-2019-4ba0-8b69-6e86c5f78f91 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2493.766131] env[62824]: value = "task-2147062" [ 2493.766131] env[62824]: _type = "Task" [ 2493.766131] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2493.774030] env[62824]: DEBUG oslo_vmware.api [None req-27a25ddf-2019-4ba0-8b69-6e86c5f78f91 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2147062, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2493.816325] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-11b8fbfb-ae9d-4426-8f86-5f4c2fe088f8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Creating linked-clone VM from snapshot {{(pid=62824) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 2493.816575] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-72fb25a1-db46-4b72-ae9b-e2a63ddc3bb7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2493.825114] env[62824]: DEBUG oslo_vmware.api [None req-11b8fbfb-ae9d-4426-8f86-5f4c2fe088f8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2493.825114] env[62824]: value = "task-2147063" [ 2493.825114] env[62824]: _type = "Task" [ 2493.825114] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2493.833833] env[62824]: DEBUG oslo_vmware.api [None req-11b8fbfb-ae9d-4426-8f86-5f4c2fe088f8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147063, 'name': CloneVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2494.069368] env[62824]: DEBUG oslo_concurrency.lockutils [None req-971a8579-d52f-40b0-a5b3-0cc936973627 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Lock "ad9f01dd-d719-41a2-971b-b12cd3e8bdf6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.171s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2494.275584] env[62824]: DEBUG oslo_vmware.api [None req-27a25ddf-2019-4ba0-8b69-6e86c5f78f91 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2147062, 'name': ReconfigVM_Task, 'duration_secs': 0.127911} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2494.275754] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-27a25ddf-2019-4ba0-8b69-6e86c5f78f91 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-438783', 'volume_id': '08d8bf83-f546-44c3-8c92-1efd79d0f46e', 'name': 'volume-08d8bf83-f546-44c3-8c92-1efd79d0f46e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'e524ce78-6a63-4950-b727-ff72f8df4424', 'attached_at': '', 'detached_at': '', 'volume_id': '08d8bf83-f546-44c3-8c92-1efd79d0f46e', 'serial': '08d8bf83-f546-44c3-8c92-1efd79d0f46e'} {{(pid=62824) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 2494.276233] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a8103bfc-f0ec-44b3-be6f-04e64cbccdb3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2494.282454] env[62824]: DEBUG oslo_vmware.api [None req-27a25ddf-2019-4ba0-8b69-6e86c5f78f91 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2494.282454] env[62824]: value = "task-2147064" [ 2494.282454] env[62824]: _type = "Task" [ 2494.282454] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2494.289948] env[62824]: DEBUG oslo_vmware.api [None req-27a25ddf-2019-4ba0-8b69-6e86c5f78f91 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2147064, 'name': Rename_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2494.334472] env[62824]: DEBUG oslo_vmware.api [None req-11b8fbfb-ae9d-4426-8f86-5f4c2fe088f8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147063, 'name': CloneVM_Task} progress is 93%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2494.714727] env[62824]: DEBUG oslo_concurrency.lockutils [None req-39599a5b-c5b7-475d-a413-db2b94d67502 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Acquiring lock "c70880c9-a69a-49b9-b028-189c04276623" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2494.714998] env[62824]: DEBUG oslo_concurrency.lockutils [None req-39599a5b-c5b7-475d-a413-db2b94d67502 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Lock "c70880c9-a69a-49b9-b028-189c04276623" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2494.792747] env[62824]: DEBUG oslo_vmware.api [None req-27a25ddf-2019-4ba0-8b69-6e86c5f78f91 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2147064, 'name': Rename_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2494.835511] env[62824]: DEBUG oslo_vmware.api [None req-11b8fbfb-ae9d-4426-8f86-5f4c2fe088f8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147063, 'name': CloneVM_Task} progress is 93%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2495.219093] env[62824]: INFO nova.compute.manager [None req-39599a5b-c5b7-475d-a413-db2b94d67502 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c70880c9-a69a-49b9-b028-189c04276623] Detaching volume 42f33a8d-0a1f-4132-8b0d-90a607343a57 [ 2495.251307] env[62824]: INFO nova.virt.block_device [None req-39599a5b-c5b7-475d-a413-db2b94d67502 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c70880c9-a69a-49b9-b028-189c04276623] Attempting to driver detach volume 42f33a8d-0a1f-4132-8b0d-90a607343a57 from mountpoint /dev/sdb [ 2495.251916] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-39599a5b-c5b7-475d-a413-db2b94d67502 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c70880c9-a69a-49b9-b028-189c04276623] Volume detach. Driver type: vmdk {{(pid=62824) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 2495.252209] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-39599a5b-c5b7-475d-a413-db2b94d67502 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c70880c9-a69a-49b9-b028-189c04276623] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-438775', 'volume_id': '42f33a8d-0a1f-4132-8b0d-90a607343a57', 'name': 'volume-42f33a8d-0a1f-4132-8b0d-90a607343a57', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'c70880c9-a69a-49b9-b028-189c04276623', 'attached_at': '', 'detached_at': '', 'volume_id': '42f33a8d-0a1f-4132-8b0d-90a607343a57', 'serial': '42f33a8d-0a1f-4132-8b0d-90a607343a57'} {{(pid=62824) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 2495.253236] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17e43440-19dd-438b-9aa4-27d909555e39 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2495.274726] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d12d2c90-6a2b-4faa-88df-b85dfd57dc77 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2495.281817] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e54275fe-4e5b-4c3f-bcbb-9fec9cfa8fe2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2495.292050] env[62824]: DEBUG oslo_vmware.api [None req-27a25ddf-2019-4ba0-8b69-6e86c5f78f91 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2147064, 'name': Rename_Task, 'duration_secs': 0.804172} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2495.306843] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-27a25ddf-2019-4ba0-8b69-6e86c5f78f91 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2495.307169] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9401de26-cf53-4f8b-b4de-985a293ab36e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2495.308994] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-043c2641-d483-4da2-b465-f8b902602769 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2495.325583] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-39599a5b-c5b7-475d-a413-db2b94d67502 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] The volume has not been displaced from its original location: [datastore1] volume-42f33a8d-0a1f-4132-8b0d-90a607343a57/volume-42f33a8d-0a1f-4132-8b0d-90a607343a57.vmdk. No consolidation needed. {{(pid=62824) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 2495.331106] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-39599a5b-c5b7-475d-a413-db2b94d67502 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c70880c9-a69a-49b9-b028-189c04276623] Reconfiguring VM instance instance-0000006e to detach disk 2001 {{(pid=62824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 2495.332363] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e4ab3b68-1916-4985-80fb-33bb7d0132ac {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2495.345363] env[62824]: DEBUG oslo_vmware.api [None req-27a25ddf-2019-4ba0-8b69-6e86c5f78f91 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2495.345363] env[62824]: value = "task-2147065" [ 2495.345363] env[62824]: _type = "Task" [ 2495.345363] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2495.353722] env[62824]: DEBUG oslo_vmware.api [None req-39599a5b-c5b7-475d-a413-db2b94d67502 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Waiting for the task: (returnval){ [ 2495.353722] env[62824]: value = "task-2147066" [ 2495.353722] env[62824]: _type = "Task" [ 2495.353722] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2495.359641] env[62824]: DEBUG oslo_vmware.api [None req-27a25ddf-2019-4ba0-8b69-6e86c5f78f91 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2147065, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2495.359860] env[62824]: DEBUG oslo_vmware.api [None req-11b8fbfb-ae9d-4426-8f86-5f4c2fe088f8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147063, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2495.367610] env[62824]: DEBUG oslo_vmware.api [None req-39599a5b-c5b7-475d-a413-db2b94d67502 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147066, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2495.841420] env[62824]: DEBUG oslo_vmware.api [None req-11b8fbfb-ae9d-4426-8f86-5f4c2fe088f8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147063, 'name': CloneVM_Task, 'duration_secs': 1.764617} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2495.841829] env[62824]: INFO nova.virt.vmwareapi.vmops [None req-11b8fbfb-ae9d-4426-8f86-5f4c2fe088f8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Created linked-clone VM from snapshot [ 2495.842533] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67d6dadd-0551-4513-b6fc-c64e9839a39c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2495.849888] env[62824]: DEBUG nova.virt.vmwareapi.images [None req-11b8fbfb-ae9d-4426-8f86-5f4c2fe088f8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Uploading image d7446151-5eb0-4242-b945-0e231bbb833b {{(pid=62824) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 2495.859182] env[62824]: DEBUG oslo_vmware.api [None req-27a25ddf-2019-4ba0-8b69-6e86c5f78f91 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2147065, 'name': PowerOnVM_Task, 'duration_secs': 0.513161} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2495.859472] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-27a25ddf-2019-4ba0-8b69-6e86c5f78f91 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2495.859732] env[62824]: INFO nova.compute.manager [None req-27a25ddf-2019-4ba0-8b69-6e86c5f78f91 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Took 8.70 seconds to spawn the instance on the hypervisor. [ 2495.859972] env[62824]: DEBUG nova.compute.manager [None req-27a25ddf-2019-4ba0-8b69-6e86c5f78f91 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2495.860817] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2299829d-5d07-45c7-a15e-e185755c9b39 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2495.870380] env[62824]: DEBUG oslo_vmware.api [None req-39599a5b-c5b7-475d-a413-db2b94d67502 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147066, 'name': ReconfigVM_Task, 'duration_secs': 0.231066} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2495.873276] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-39599a5b-c5b7-475d-a413-db2b94d67502 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c70880c9-a69a-49b9-b028-189c04276623] Reconfigured VM instance instance-0000006e to detach disk 2001 {{(pid=62824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 2495.880555] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-92cc75e1-f1b9-4d18-aca0-7a432cf2b021 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2495.892985] env[62824]: DEBUG oslo_vmware.rw_handles [None req-11b8fbfb-ae9d-4426-8f86-5f4c2fe088f8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 2495.892985] env[62824]: value = "vm-438788" [ 2495.892985] env[62824]: _type = "VirtualMachine" [ 2495.892985] env[62824]: }. {{(pid=62824) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 2495.893471] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-aa79e96d-cbb2-4306-9d3a-dec33a9da7a6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2495.899942] env[62824]: DEBUG oslo_vmware.rw_handles [None req-11b8fbfb-ae9d-4426-8f86-5f4c2fe088f8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lease: (returnval){ [ 2495.899942] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5208b142-4289-2e5a-95e5-dbb0e5fbb32d" [ 2495.899942] env[62824]: _type = "HttpNfcLease" [ 2495.899942] env[62824]: } obtained for exporting VM: (result){ [ 2495.899942] env[62824]: value = "vm-438788" [ 2495.899942] env[62824]: _type = "VirtualMachine" [ 2495.899942] env[62824]: }. {{(pid=62824) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 2495.900245] env[62824]: DEBUG oslo_vmware.api [None req-11b8fbfb-ae9d-4426-8f86-5f4c2fe088f8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the lease: (returnval){ [ 2495.900245] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5208b142-4289-2e5a-95e5-dbb0e5fbb32d" [ 2495.900245] env[62824]: _type = "HttpNfcLease" [ 2495.900245] env[62824]: } to be ready. {{(pid=62824) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 2495.901368] env[62824]: DEBUG oslo_vmware.api [None req-39599a5b-c5b7-475d-a413-db2b94d67502 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Waiting for the task: (returnval){ [ 2495.901368] env[62824]: value = "task-2147067" [ 2495.901368] env[62824]: _type = "Task" [ 2495.901368] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2495.910706] env[62824]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 2495.910706] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5208b142-4289-2e5a-95e5-dbb0e5fbb32d" [ 2495.910706] env[62824]: _type = "HttpNfcLease" [ 2495.910706] env[62824]: } is initializing. {{(pid=62824) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 2495.913750] env[62824]: DEBUG oslo_vmware.api [None req-39599a5b-c5b7-475d-a413-db2b94d67502 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147067, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2496.405535] env[62824]: INFO nova.compute.manager [None req-27a25ddf-2019-4ba0-8b69-6e86c5f78f91 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Took 15.18 seconds to build instance. [ 2496.416304] env[62824]: DEBUG oslo_vmware.api [None req-39599a5b-c5b7-475d-a413-db2b94d67502 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147067, 'name': ReconfigVM_Task, 'duration_secs': 0.142582} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2496.416514] env[62824]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 2496.416514] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5208b142-4289-2e5a-95e5-dbb0e5fbb32d" [ 2496.416514] env[62824]: _type = "HttpNfcLease" [ 2496.416514] env[62824]: } is ready. {{(pid=62824) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 2496.416786] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-39599a5b-c5b7-475d-a413-db2b94d67502 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c70880c9-a69a-49b9-b028-189c04276623] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-438775', 'volume_id': '42f33a8d-0a1f-4132-8b0d-90a607343a57', 'name': 'volume-42f33a8d-0a1f-4132-8b0d-90a607343a57', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'c70880c9-a69a-49b9-b028-189c04276623', 'attached_at': '', 'detached_at': '', 'volume_id': '42f33a8d-0a1f-4132-8b0d-90a607343a57', 'serial': '42f33a8d-0a1f-4132-8b0d-90a607343a57'} {{(pid=62824) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 2496.418975] env[62824]: DEBUG oslo_vmware.rw_handles [None req-11b8fbfb-ae9d-4426-8f86-5f4c2fe088f8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 2496.418975] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5208b142-4289-2e5a-95e5-dbb0e5fbb32d" [ 2496.418975] env[62824]: _type = "HttpNfcLease" [ 2496.418975] env[62824]: }. {{(pid=62824) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 2496.420409] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c00d63ba-959d-4d4e-aec9-265461549906 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2496.428395] env[62824]: DEBUG oslo_vmware.rw_handles [None req-11b8fbfb-ae9d-4426-8f86-5f4c2fe088f8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52213157-03c3-3e5b-7183-402215ecd4fa/disk-0.vmdk from lease info. {{(pid=62824) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 2496.428569] env[62824]: DEBUG oslo_vmware.rw_handles [None req-11b8fbfb-ae9d-4426-8f86-5f4c2fe088f8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52213157-03c3-3e5b-7183-402215ecd4fa/disk-0.vmdk for reading. {{(pid=62824) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 2496.518114] env[62824]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-bf824a33-d508-4f0d-bbd4-f09158718412 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2496.911246] env[62824]: DEBUG oslo_concurrency.lockutils [None req-27a25ddf-2019-4ba0-8b69-6e86c5f78f91 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "e524ce78-6a63-4950-b727-ff72f8df4424" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.689s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2496.961164] env[62824]: DEBUG nova.objects.instance [None req-39599a5b-c5b7-475d-a413-db2b94d67502 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Lazy-loading 'flavor' on Instance uuid c70880c9-a69a-49b9-b028-189c04276623 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2497.166178] env[62824]: DEBUG nova.compute.manager [req-28fb6313-e8a1-4b7e-8585-1b08ba13b525 req-3559b2f9-571c-4aa0-a800-50b2e21a20c4 service nova] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] Received event network-changed-3eb965bf-7ceb-4a51-bc6f-00803bc6b99e {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2497.168855] env[62824]: DEBUG nova.compute.manager [req-28fb6313-e8a1-4b7e-8585-1b08ba13b525 req-3559b2f9-571c-4aa0-a800-50b2e21a20c4 service nova] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] Refreshing instance network info cache due to event network-changed-3eb965bf-7ceb-4a51-bc6f-00803bc6b99e. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2497.168855] env[62824]: DEBUG oslo_concurrency.lockutils [req-28fb6313-e8a1-4b7e-8585-1b08ba13b525 req-3559b2f9-571c-4aa0-a800-50b2e21a20c4 service nova] Acquiring lock "refresh_cache-ca42fef0-1e90-4ab6-9d60-8ef7e4997884" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2497.168855] env[62824]: DEBUG oslo_concurrency.lockutils [req-28fb6313-e8a1-4b7e-8585-1b08ba13b525 req-3559b2f9-571c-4aa0-a800-50b2e21a20c4 service nova] Acquired lock "refresh_cache-ca42fef0-1e90-4ab6-9d60-8ef7e4997884" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2497.168855] env[62824]: DEBUG nova.network.neutron [req-28fb6313-e8a1-4b7e-8585-1b08ba13b525 req-3559b2f9-571c-4aa0-a800-50b2e21a20c4 service nova] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] Refreshing network info cache for port 3eb965bf-7ceb-4a51-bc6f-00803bc6b99e {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2497.772986] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2497.774188] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager.update_available_resource {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2497.876382] env[62824]: DEBUG nova.network.neutron [req-28fb6313-e8a1-4b7e-8585-1b08ba13b525 req-3559b2f9-571c-4aa0-a800-50b2e21a20c4 service nova] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] Updated VIF entry in instance network info cache for port 3eb965bf-7ceb-4a51-bc6f-00803bc6b99e. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2497.877038] env[62824]: DEBUG nova.network.neutron [req-28fb6313-e8a1-4b7e-8585-1b08ba13b525 req-3559b2f9-571c-4aa0-a800-50b2e21a20c4 service nova] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] Updating instance_info_cache with network_info: [{"id": "3eb965bf-7ceb-4a51-bc6f-00803bc6b99e", "address": "fa:16:3e:3f:e9:09", "network": {"id": "b4ed285c-26a7-461b-8a71-d18e07582ab8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-997150974-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "815ecf9340934733a83d4dcb26612fd4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69f65356-c85e-4b7f-ad28-7c7b5e8cf50c", "external-id": "nsx-vlan-transportzone-281", "segmentation_id": 281, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3eb965bf-7c", "ovs_interfaceid": "3eb965bf-7ceb-4a51-bc6f-00803bc6b99e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2497.970511] env[62824]: DEBUG oslo_concurrency.lockutils [None req-39599a5b-c5b7-475d-a413-db2b94d67502 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Lock "c70880c9-a69a-49b9-b028-189c04276623" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.255s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2498.276727] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2498.276961] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2498.277180] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2498.277373] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62824) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 2498.278313] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54b4db37-9f32-4675-b55e-26eec80bd47f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2498.287212] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06b42da2-01ea-4f01-9732-5e4cf3b55310 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2498.303999] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcacc8be-0f89-4441-adca-835634f3bbdb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2498.312922] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90c547bc-baa6-4620-8a6d-d9686dd517c3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2498.346922] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180528MB free_disk=175GB free_vcpus=48 pci_devices=None {{(pid=62824) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 2498.347102] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2498.347345] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2498.379731] env[62824]: DEBUG oslo_concurrency.lockutils [req-28fb6313-e8a1-4b7e-8585-1b08ba13b525 req-3559b2f9-571c-4aa0-a800-50b2e21a20c4 service nova] Releasing lock "refresh_cache-ca42fef0-1e90-4ab6-9d60-8ef7e4997884" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2498.611424] env[62824]: DEBUG nova.compute.manager [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Stashing vm_state: active {{(pid=62824) _prep_resize /opt/stack/nova/nova/compute/manager.py:6136}} [ 2498.991996] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2138c6dc-2a53-4862-8675-50ad9ff4176e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Acquiring lock "c70880c9-a69a-49b9-b028-189c04276623" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2498.992439] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2138c6dc-2a53-4862-8675-50ad9ff4176e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Lock "c70880c9-a69a-49b9-b028-189c04276623" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2498.993057] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2138c6dc-2a53-4862-8675-50ad9ff4176e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Acquiring lock "c70880c9-a69a-49b9-b028-189c04276623-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2498.993269] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2138c6dc-2a53-4862-8675-50ad9ff4176e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Lock "c70880c9-a69a-49b9-b028-189c04276623-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2498.993449] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2138c6dc-2a53-4862-8675-50ad9ff4176e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Lock "c70880c9-a69a-49b9-b028-189c04276623-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2498.995722] env[62824]: INFO nova.compute.manager [None req-2138c6dc-2a53-4862-8675-50ad9ff4176e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c70880c9-a69a-49b9-b028-189c04276623] Terminating instance [ 2499.134853] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2499.197241] env[62824]: DEBUG nova.compute.manager [req-ba291e5b-4b68-473e-98e4-4f1120973145 req-02ce6412-49d5-4640-b42a-bbb870f984ef service nova] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Received event network-changed-62aa7735-5027-4b47-917b-7b769eac4cee {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2499.197397] env[62824]: DEBUG nova.compute.manager [req-ba291e5b-4b68-473e-98e4-4f1120973145 req-02ce6412-49d5-4640-b42a-bbb870f984ef service nova] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Refreshing instance network info cache due to event network-changed-62aa7735-5027-4b47-917b-7b769eac4cee. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2499.197618] env[62824]: DEBUG oslo_concurrency.lockutils [req-ba291e5b-4b68-473e-98e4-4f1120973145 req-02ce6412-49d5-4640-b42a-bbb870f984ef service nova] Acquiring lock "refresh_cache-e524ce78-6a63-4950-b727-ff72f8df4424" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2499.197767] env[62824]: DEBUG oslo_concurrency.lockutils [req-ba291e5b-4b68-473e-98e4-4f1120973145 req-02ce6412-49d5-4640-b42a-bbb870f984ef service nova] Acquired lock "refresh_cache-e524ce78-6a63-4950-b727-ff72f8df4424" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2499.197932] env[62824]: DEBUG nova.network.neutron [req-ba291e5b-4b68-473e-98e4-4f1120973145 req-02ce6412-49d5-4640-b42a-bbb870f984ef service nova] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Refreshing network info cache for port 62aa7735-5027-4b47-917b-7b769eac4cee {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2499.358735] env[62824]: INFO nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Updating resource usage from migration 4c723d9e-36f1-43f6-b8f2-091f294e414f [ 2499.376136] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance ca42fef0-1e90-4ab6-9d60-8ef7e4997884 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2499.376297] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance c70880c9-a69a-49b9-b028-189c04276623 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2499.376423] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 7c47bb92-a58e-4537-a68a-289e2a5fc19a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2499.376543] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 4dff2a76-b0a2-438c-b52f-d6495948d041 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2499.376659] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 80c57382-8f6f-4537-b210-70a902b02cb4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2499.376871] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Migration 4c723d9e-36f1-43f6-b8f2-091f294e414f is active on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 2499.376997] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance e524ce78-6a63-4950-b727-ff72f8df4424 actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2499.377205] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Total usable vcpus: 48, total allocated vcpus: 6 {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 2499.377340] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1664MB phys_disk=200GB used_disk=5GB total_vcpus=48 used_vcpus=6 pci_stats=[] {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 2499.471509] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12bb0a54-4959-4bb9-9284-61d32a483023 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2499.479506] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92ed6580-07ff-47db-b570-82692d49eb4b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2499.510433] env[62824]: DEBUG nova.compute.manager [None req-2138c6dc-2a53-4862-8675-50ad9ff4176e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c70880c9-a69a-49b9-b028-189c04276623] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2499.510622] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-2138c6dc-2a53-4862-8675-50ad9ff4176e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c70880c9-a69a-49b9-b028-189c04276623] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2499.511564] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d6e7874-9cb7-452b-abb7-92450e6fae62 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2499.514669] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85eb72ab-dfa8-441f-9e12-29e7fc5dc6d5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2499.523616] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31987b72-3ad5-462e-a0e5-a50773f3a1ba {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2499.527271] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-2138c6dc-2a53-4862-8675-50ad9ff4176e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c70880c9-a69a-49b9-b028-189c04276623] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2499.527515] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-53a8e2cd-005c-451e-8e41-a8288801cf74 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2499.538417] env[62824]: DEBUG nova.compute.provider_tree [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2499.540961] env[62824]: DEBUG oslo_vmware.api [None req-2138c6dc-2a53-4862-8675-50ad9ff4176e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Waiting for the task: (returnval){ [ 2499.540961] env[62824]: value = "task-2147069" [ 2499.540961] env[62824]: _type = "Task" [ 2499.540961] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2499.548804] env[62824]: DEBUG oslo_vmware.api [None req-2138c6dc-2a53-4862-8675-50ad9ff4176e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147069, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2499.933463] env[62824]: DEBUG nova.network.neutron [req-ba291e5b-4b68-473e-98e4-4f1120973145 req-02ce6412-49d5-4640-b42a-bbb870f984ef service nova] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Updated VIF entry in instance network info cache for port 62aa7735-5027-4b47-917b-7b769eac4cee. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2499.933875] env[62824]: DEBUG nova.network.neutron [req-ba291e5b-4b68-473e-98e4-4f1120973145 req-02ce6412-49d5-4640-b42a-bbb870f984ef service nova] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Updating instance_info_cache with network_info: [{"id": "62aa7735-5027-4b47-917b-7b769eac4cee", "address": "fa:16:3e:61:22:c9", "network": {"id": "b4ed285c-26a7-461b-8a71-d18e07582ab8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-997150974-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.185", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "815ecf9340934733a83d4dcb26612fd4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69f65356-c85e-4b7f-ad28-7c7b5e8cf50c", "external-id": "nsx-vlan-transportzone-281", "segmentation_id": 281, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap62aa7735-50", "ovs_interfaceid": "62aa7735-5027-4b47-917b-7b769eac4cee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2500.043130] env[62824]: DEBUG nova.scheduler.client.report [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2500.055656] env[62824]: DEBUG oslo_vmware.api [None req-2138c6dc-2a53-4862-8675-50ad9ff4176e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147069, 'name': PowerOffVM_Task, 'duration_secs': 0.236368} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2500.055967] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-2138c6dc-2a53-4862-8675-50ad9ff4176e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c70880c9-a69a-49b9-b028-189c04276623] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2500.056234] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-2138c6dc-2a53-4862-8675-50ad9ff4176e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c70880c9-a69a-49b9-b028-189c04276623] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2500.056478] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-84ef8809-dc42-4b41-a53d-06e66a4e0cd8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2500.436658] env[62824]: DEBUG oslo_concurrency.lockutils [req-ba291e5b-4b68-473e-98e4-4f1120973145 req-02ce6412-49d5-4640-b42a-bbb870f984ef service nova] Releasing lock "refresh_cache-e524ce78-6a63-4950-b727-ff72f8df4424" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2500.550975] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62824) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 2500.551280] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.204s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2500.551520] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 1.417s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2501.056814] env[62824]: INFO nova.compute.claims [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2501.564175] env[62824]: INFO nova.compute.resource_tracker [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Updating resource usage from migration 4c723d9e-36f1-43f6-b8f2-091f294e414f [ 2501.653865] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7439fb8-3d68-4381-b29a-45281adeaacd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2501.661360] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48147d3d-4802-4fc5-a196-dad51bb03da5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2501.691741] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-449772e7-6f50-471a-a930-2ca88c49086a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2501.698539] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90550242-457a-46a8-a4a8-fa53241b8ea3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2501.711309] env[62824]: DEBUG nova.compute.provider_tree [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2502.214818] env[62824]: DEBUG nova.scheduler.client.report [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2502.720317] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.169s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2502.720565] env[62824]: INFO nova.compute.manager [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Migrating [ 2503.237528] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquiring lock "refresh_cache-e524ce78-6a63-4950-b727-ff72f8df4424" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2503.237863] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquired lock "refresh_cache-e524ce78-6a63-4950-b727-ff72f8df4424" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2503.237930] env[62824]: DEBUG nova.network.neutron [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2503.552168] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2503.552341] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Starting heal instance info cache {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10445}} [ 2503.552475] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Rebuilding the list of instances to heal {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10449}} [ 2503.952994] env[62824]: DEBUG nova.network.neutron [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Updating instance_info_cache with network_info: [{"id": "62aa7735-5027-4b47-917b-7b769eac4cee", "address": "fa:16:3e:61:22:c9", "network": {"id": "b4ed285c-26a7-461b-8a71-d18e07582ab8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-997150974-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.185", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "815ecf9340934733a83d4dcb26612fd4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69f65356-c85e-4b7f-ad28-7c7b5e8cf50c", "external-id": "nsx-vlan-transportzone-281", "segmentation_id": 281, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap62aa7735-50", "ovs_interfaceid": "62aa7735-5027-4b47-917b-7b769eac4cee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2504.057122] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: c70880c9-a69a-49b9-b028-189c04276623] Skipping network cache update for instance because it is being deleted. {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10462}} [ 2504.084619] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "refresh_cache-ca42fef0-1e90-4ab6-9d60-8ef7e4997884" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2504.084772] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquired lock "refresh_cache-ca42fef0-1e90-4ab6-9d60-8ef7e4997884" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2504.084936] env[62824]: DEBUG nova.network.neutron [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] Forcefully refreshing network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 2504.085100] env[62824]: DEBUG nova.objects.instance [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lazy-loading 'info_cache' on Instance uuid ca42fef0-1e90-4ab6-9d60-8ef7e4997884 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2504.456086] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Releasing lock "refresh_cache-e524ce78-6a63-4950-b727-ff72f8df4424" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2504.782700] env[62824]: DEBUG oslo_vmware.rw_handles [None req-11b8fbfb-ae9d-4426-8f86-5f4c2fe088f8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52213157-03c3-3e5b-7183-402215ecd4fa/disk-0.vmdk. {{(pid=62824) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 2504.783695] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59427b4f-c479-42e3-8d63-f8269fca4f5e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2504.790445] env[62824]: DEBUG oslo_vmware.rw_handles [None req-11b8fbfb-ae9d-4426-8f86-5f4c2fe088f8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52213157-03c3-3e5b-7183-402215ecd4fa/disk-0.vmdk is in state: ready. {{(pid=62824) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 2504.790609] env[62824]: ERROR oslo_vmware.rw_handles [None req-11b8fbfb-ae9d-4426-8f86-5f4c2fe088f8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52213157-03c3-3e5b-7183-402215ecd4fa/disk-0.vmdk due to incomplete transfer. [ 2504.790813] env[62824]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-6e26779e-7e91-44b0-9131-91ffddca49da {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2504.797542] env[62824]: DEBUG oslo_vmware.rw_handles [None req-11b8fbfb-ae9d-4426-8f86-5f4c2fe088f8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52213157-03c3-3e5b-7183-402215ecd4fa/disk-0.vmdk. {{(pid=62824) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 2504.797733] env[62824]: DEBUG nova.virt.vmwareapi.images [None req-11b8fbfb-ae9d-4426-8f86-5f4c2fe088f8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Uploaded image d7446151-5eb0-4242-b945-0e231bbb833b to the Glance image server {{(pid=62824) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 2504.799831] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-11b8fbfb-ae9d-4426-8f86-5f4c2fe088f8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Destroying the VM {{(pid=62824) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 2504.800077] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-629b49ce-4744-47db-ad25-d5df2fd1daea {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2504.805817] env[62824]: DEBUG oslo_vmware.api [None req-11b8fbfb-ae9d-4426-8f86-5f4c2fe088f8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2504.805817] env[62824]: value = "task-2147071" [ 2504.805817] env[62824]: _type = "Task" [ 2504.805817] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2504.813673] env[62824]: DEBUG oslo_vmware.api [None req-11b8fbfb-ae9d-4426-8f86-5f4c2fe088f8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147071, 'name': Destroy_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2505.316532] env[62824]: DEBUG oslo_vmware.api [None req-11b8fbfb-ae9d-4426-8f86-5f4c2fe088f8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147071, 'name': Destroy_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2505.793821] env[62824]: DEBUG nova.network.neutron [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] Updating instance_info_cache with network_info: [{"id": "3eb965bf-7ceb-4a51-bc6f-00803bc6b99e", "address": "fa:16:3e:3f:e9:09", "network": {"id": "b4ed285c-26a7-461b-8a71-d18e07582ab8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-997150974-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "815ecf9340934733a83d4dcb26612fd4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69f65356-c85e-4b7f-ad28-7c7b5e8cf50c", "external-id": "nsx-vlan-transportzone-281", "segmentation_id": 281, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3eb965bf-7c", "ovs_interfaceid": "3eb965bf-7ceb-4a51-bc6f-00803bc6b99e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2505.816594] env[62824]: DEBUG oslo_vmware.api [None req-11b8fbfb-ae9d-4426-8f86-5f4c2fe088f8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147071, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2505.971938] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c7a7834-e41c-480e-a28c-97458a70189c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2505.990972] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Updating instance 'e524ce78-6a63-4950-b727-ff72f8df4424' progress to 0 {{(pid=62824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 2506.296541] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Releasing lock "refresh_cache-ca42fef0-1e90-4ab6-9d60-8ef7e4997884" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2506.296768] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] Updated the network info_cache for instance {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10516}} [ 2506.296881] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2506.297065] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2506.297225] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2506.297377] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2506.297519] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2506.316902] env[62824]: DEBUG oslo_vmware.api [None req-11b8fbfb-ae9d-4426-8f86-5f4c2fe088f8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147071, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2506.497364] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2506.497500] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-471b0401-1602-4579-abbf-946971a15758 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2506.504542] env[62824]: DEBUG oslo_vmware.api [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2506.504542] env[62824]: value = "task-2147072" [ 2506.504542] env[62824]: _type = "Task" [ 2506.504542] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2506.514466] env[62824]: DEBUG oslo_vmware.api [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2147072, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2506.522444] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-2138c6dc-2a53-4862-8675-50ad9ff4176e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c70880c9-a69a-49b9-b028-189c04276623] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2506.522655] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-2138c6dc-2a53-4862-8675-50ad9ff4176e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c70880c9-a69a-49b9-b028-189c04276623] Deleting contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2506.522896] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-2138c6dc-2a53-4862-8675-50ad9ff4176e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Deleting the datastore file [datastore2] c70880c9-a69a-49b9-b028-189c04276623 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2506.523174] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-847df6e9-7688-41c8-b9dc-dcdfe6a27309 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2506.530053] env[62824]: DEBUG oslo_vmware.api [None req-2138c6dc-2a53-4862-8675-50ad9ff4176e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Waiting for the task: (returnval){ [ 2506.530053] env[62824]: value = "task-2147073" [ 2506.530053] env[62824]: _type = "Task" [ 2506.530053] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2506.537315] env[62824]: DEBUG oslo_vmware.api [None req-2138c6dc-2a53-4862-8675-50ad9ff4176e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147073, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2506.772611] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2506.772834] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62824) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11064}} [ 2506.817752] env[62824]: DEBUG oslo_vmware.api [None req-11b8fbfb-ae9d-4426-8f86-5f4c2fe088f8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147071, 'name': Destroy_Task, 'duration_secs': 1.72853} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2506.818173] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-11b8fbfb-ae9d-4426-8f86-5f4c2fe088f8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Destroyed the VM [ 2506.818271] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-11b8fbfb-ae9d-4426-8f86-5f4c2fe088f8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Deleting Snapshot of the VM instance {{(pid=62824) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 2506.818530] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-41d28f19-c8ac-49e9-9d8c-bfba418ac360 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2506.824884] env[62824]: DEBUG oslo_vmware.api [None req-11b8fbfb-ae9d-4426-8f86-5f4c2fe088f8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2506.824884] env[62824]: value = "task-2147074" [ 2506.824884] env[62824]: _type = "Task" [ 2506.824884] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2506.832335] env[62824]: DEBUG oslo_vmware.api [None req-11b8fbfb-ae9d-4426-8f86-5f4c2fe088f8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147074, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2507.014601] env[62824]: DEBUG oslo_vmware.api [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2147072, 'name': PowerOffVM_Task, 'duration_secs': 0.186896} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2507.014889] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2507.015106] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Updating instance 'e524ce78-6a63-4950-b727-ff72f8df4424' progress to 17 {{(pid=62824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 2507.040907] env[62824]: DEBUG oslo_vmware.api [None req-2138c6dc-2a53-4862-8675-50ad9ff4176e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147073, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.139562} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2507.041255] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-2138c6dc-2a53-4862-8675-50ad9ff4176e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2507.041504] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-2138c6dc-2a53-4862-8675-50ad9ff4176e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c70880c9-a69a-49b9-b028-189c04276623] Deleted contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2507.041707] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-2138c6dc-2a53-4862-8675-50ad9ff4176e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c70880c9-a69a-49b9-b028-189c04276623] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2507.041956] env[62824]: INFO nova.compute.manager [None req-2138c6dc-2a53-4862-8675-50ad9ff4176e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c70880c9-a69a-49b9-b028-189c04276623] Took 7.53 seconds to destroy the instance on the hypervisor. [ 2507.042288] env[62824]: DEBUG oslo.service.loopingcall [None req-2138c6dc-2a53-4862-8675-50ad9ff4176e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2507.042538] env[62824]: DEBUG nova.compute.manager [-] [instance: c70880c9-a69a-49b9-b028-189c04276623] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2507.042666] env[62824]: DEBUG nova.network.neutron [-] [instance: c70880c9-a69a-49b9-b028-189c04276623] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2507.334775] env[62824]: DEBUG oslo_vmware.api [None req-11b8fbfb-ae9d-4426-8f86-5f4c2fe088f8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147074, 'name': RemoveSnapshot_Task, 'duration_secs': 0.42167} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2507.335124] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-11b8fbfb-ae9d-4426-8f86-5f4c2fe088f8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Deleted Snapshot of the VM instance {{(pid=62824) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 2507.335407] env[62824]: INFO nova.compute.manager [None req-11b8fbfb-ae9d-4426-8f86-5f4c2fe088f8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Took 15.09 seconds to snapshot the instance on the hypervisor. [ 2507.478286] env[62824]: DEBUG nova.compute.manager [req-79f614fb-4931-4bbb-b200-9448c669c4dc req-ea5ac8ed-65a6-47cc-8ba9-237675d3ea74 service nova] [instance: c70880c9-a69a-49b9-b028-189c04276623] Received event network-vif-deleted-f71d9cbc-c160-47f9-9771-af38641d3683 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2507.478402] env[62824]: INFO nova.compute.manager [req-79f614fb-4931-4bbb-b200-9448c669c4dc req-ea5ac8ed-65a6-47cc-8ba9-237675d3ea74 service nova] [instance: c70880c9-a69a-49b9-b028-189c04276623] Neutron deleted interface f71d9cbc-c160-47f9-9771-af38641d3683; detaching it from the instance and deleting it from the info cache [ 2507.478583] env[62824]: DEBUG nova.network.neutron [req-79f614fb-4931-4bbb-b200-9448c669c4dc req-ea5ac8ed-65a6-47cc-8ba9-237675d3ea74 service nova] [instance: c70880c9-a69a-49b9-b028-189c04276623] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2507.521929] env[62824]: DEBUG nova.virt.hardware [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:24Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2507.522204] env[62824]: DEBUG nova.virt.hardware [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2507.522362] env[62824]: DEBUG nova.virt.hardware [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2507.522546] env[62824]: DEBUG nova.virt.hardware [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2507.522691] env[62824]: DEBUG nova.virt.hardware [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2507.522872] env[62824]: DEBUG nova.virt.hardware [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2507.523148] env[62824]: DEBUG nova.virt.hardware [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2507.523329] env[62824]: DEBUG nova.virt.hardware [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2507.523498] env[62824]: DEBUG nova.virt.hardware [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2507.523664] env[62824]: DEBUG nova.virt.hardware [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2507.523840] env[62824]: DEBUG nova.virt.hardware [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2507.529342] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e752f429-8825-4fed-8fcb-531cef364010 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2507.545537] env[62824]: DEBUG oslo_vmware.api [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2507.545537] env[62824]: value = "task-2147075" [ 2507.545537] env[62824]: _type = "Task" [ 2507.545537] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2507.554451] env[62824]: DEBUG oslo_vmware.api [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2147075, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2507.885055] env[62824]: DEBUG nova.compute.manager [None req-11b8fbfb-ae9d-4426-8f86-5f4c2fe088f8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Found 3 images (rotation: 2) {{(pid=62824) _rotate_backups /opt/stack/nova/nova/compute/manager.py:5017}} [ 2507.885297] env[62824]: DEBUG nova.compute.manager [None req-11b8fbfb-ae9d-4426-8f86-5f4c2fe088f8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Rotating out 1 backups {{(pid=62824) _rotate_backups /opt/stack/nova/nova/compute/manager.py:5025}} [ 2507.885457] env[62824]: DEBUG nova.compute.manager [None req-11b8fbfb-ae9d-4426-8f86-5f4c2fe088f8 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Deleting image 13001af0-4e21-48f0-84af-44e4c51ff081 {{(pid=62824) _rotate_backups /opt/stack/nova/nova/compute/manager.py:5030}} [ 2507.956786] env[62824]: DEBUG nova.network.neutron [-] [instance: c70880c9-a69a-49b9-b028-189c04276623] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2507.981829] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-998c4c31-5a85-48e6-b35f-30629b070055 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2507.991623] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ad4a164-da0b-4360-b964-36b84c80d7ce {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2508.020537] env[62824]: DEBUG nova.compute.manager [req-79f614fb-4931-4bbb-b200-9448c669c4dc req-ea5ac8ed-65a6-47cc-8ba9-237675d3ea74 service nova] [instance: c70880c9-a69a-49b9-b028-189c04276623] Detach interface failed, port_id=f71d9cbc-c160-47f9-9771-af38641d3683, reason: Instance c70880c9-a69a-49b9-b028-189c04276623 could not be found. {{(pid=62824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11482}} [ 2508.055402] env[62824]: DEBUG oslo_vmware.api [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2147075, 'name': ReconfigVM_Task, 'duration_secs': 0.425356} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2508.055701] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Updating instance 'e524ce78-6a63-4950-b727-ff72f8df4424' progress to 33 {{(pid=62824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 2508.459944] env[62824]: INFO nova.compute.manager [-] [instance: c70880c9-a69a-49b9-b028-189c04276623] Took 1.42 seconds to deallocate network for instance. [ 2508.561673] env[62824]: DEBUG nova.virt.hardware [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2508.561941] env[62824]: DEBUG nova.virt.hardware [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2508.562275] env[62824]: DEBUG nova.virt.hardware [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2508.562517] env[62824]: DEBUG nova.virt.hardware [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2508.562677] env[62824]: DEBUG nova.virt.hardware [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2508.562922] env[62824]: DEBUG nova.virt.hardware [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2508.563241] env[62824]: DEBUG nova.virt.hardware [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2508.563495] env[62824]: DEBUG nova.virt.hardware [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2508.563870] env[62824]: DEBUG nova.virt.hardware [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2508.564180] env[62824]: DEBUG nova.virt.hardware [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2508.564466] env[62824]: DEBUG nova.virt.hardware [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2508.569987] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Reconfiguring VM instance instance-00000075 to detach disk 2000 {{(pid=62824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 2508.570279] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8703d429-f30e-4a11-a9da-bc9895ea00b7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2508.589335] env[62824]: DEBUG oslo_vmware.api [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2508.589335] env[62824]: value = "task-2147076" [ 2508.589335] env[62824]: _type = "Task" [ 2508.589335] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2508.597333] env[62824]: DEBUG oslo_vmware.api [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2147076, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2508.967487] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2138c6dc-2a53-4862-8675-50ad9ff4176e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2508.967869] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2138c6dc-2a53-4862-8675-50ad9ff4176e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2508.968143] env[62824]: DEBUG nova.objects.instance [None req-2138c6dc-2a53-4862-8675-50ad9ff4176e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Lazy-loading 'resources' on Instance uuid c70880c9-a69a-49b9-b028-189c04276623 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2509.099042] env[62824]: DEBUG oslo_vmware.api [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2147076, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2509.568109] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ddad5a7-38af-4987-b72e-4c35818846c7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2509.576215] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55a3cd9b-641b-41a4-8089-01db58e63531 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2509.611827] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9a9ab4a-a9c7-46c9-9b64-5337c4978513 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2509.619538] env[62824]: DEBUG oslo_vmware.api [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2147076, 'name': ReconfigVM_Task, 'duration_secs': 0.755002} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2509.621617] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Reconfigured VM instance instance-00000075 to detach disk 2000 {{(pid=62824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 2509.622414] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcde245d-8e8b-4d2e-ab16-4c0d546a8b8f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2509.625786] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85827f58-a6a0-47e8-96ff-4bccf2e5a765 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2509.639365] env[62824]: DEBUG nova.compute.provider_tree [None req-2138c6dc-2a53-4862-8675-50ad9ff4176e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2509.658959] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Reconfiguring VM instance instance-00000075 to attach disk [datastore2] volume-08d8bf83-f546-44c3-8c92-1efd79d0f46e/volume-08d8bf83-f546-44c3-8c92-1efd79d0f46e.vmdk or device None with type thin {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2509.659945] env[62824]: DEBUG nova.scheduler.client.report [None req-2138c6dc-2a53-4862-8675-50ad9ff4176e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2509.663047] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d90fbcce-0a77-44a3-ad42-511abe2f488d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2509.681493] env[62824]: DEBUG oslo_vmware.api [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2509.681493] env[62824]: value = "task-2147077" [ 2509.681493] env[62824]: _type = "Task" [ 2509.681493] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2509.689300] env[62824]: DEBUG oslo_vmware.api [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2147077, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2510.050563] env[62824]: DEBUG oslo_concurrency.lockutils [None req-66195b79-c107-47ef-a529-3e9640305871 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquiring lock "7c47bb92-a58e-4537-a68a-289e2a5fc19a" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2510.050903] env[62824]: DEBUG oslo_concurrency.lockutils [None req-66195b79-c107-47ef-a529-3e9640305871 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lock "7c47bb92-a58e-4537-a68a-289e2a5fc19a" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2510.050903] env[62824]: DEBUG nova.compute.manager [None req-66195b79-c107-47ef-a529-3e9640305871 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2510.051757] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8c7c06f-52eb-4c62-b971-664e968b4ff3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2510.058797] env[62824]: DEBUG nova.compute.manager [None req-66195b79-c107-47ef-a529-3e9640305871 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62824) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3404}} [ 2510.059382] env[62824]: DEBUG nova.objects.instance [None req-66195b79-c107-47ef-a529-3e9640305871 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lazy-loading 'flavor' on Instance uuid 7c47bb92-a58e-4537-a68a-289e2a5fc19a {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2510.177825] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2138c6dc-2a53-4862-8675-50ad9ff4176e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.210s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2510.191754] env[62824]: DEBUG oslo_vmware.api [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2147077, 'name': ReconfigVM_Task, 'duration_secs': 0.228677} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2510.192145] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Reconfigured VM instance instance-00000075 to attach disk [datastore2] volume-08d8bf83-f546-44c3-8c92-1efd79d0f46e/volume-08d8bf83-f546-44c3-8c92-1efd79d0f46e.vmdk or device None with type thin {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2510.192444] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Updating instance 'e524ce78-6a63-4950-b727-ff72f8df4424' progress to 50 {{(pid=62824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 2510.203368] env[62824]: INFO nova.scheduler.client.report [None req-2138c6dc-2a53-4862-8675-50ad9ff4176e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Deleted allocations for instance c70880c9-a69a-49b9-b028-189c04276623 [ 2510.699175] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dec7af3-f402-4f5f-b47f-bf363ab42141 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2510.721712] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddb89688-fcfd-45c3-aa57-0d41676dc152 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2510.725052] env[62824]: DEBUG oslo_concurrency.lockutils [None req-2138c6dc-2a53-4862-8675-50ad9ff4176e tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Lock "c70880c9-a69a-49b9-b028-189c04276623" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.732s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2510.741461] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Updating instance 'e524ce78-6a63-4950-b727-ff72f8df4424' progress to 67 {{(pid=62824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 2511.067035] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-66195b79-c107-47ef-a529-3e9640305871 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2511.067035] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-198eb696-34a1-49ad-a752-ec380969d063 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2511.074253] env[62824]: DEBUG oslo_vmware.api [None req-66195b79-c107-47ef-a529-3e9640305871 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2511.074253] env[62824]: value = "task-2147078" [ 2511.074253] env[62824]: _type = "Task" [ 2511.074253] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2511.083208] env[62824]: DEBUG oslo_vmware.api [None req-66195b79-c107-47ef-a529-3e9640305871 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147078, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2511.584218] env[62824]: DEBUG oslo_vmware.api [None req-66195b79-c107-47ef-a529-3e9640305871 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147078, 'name': PowerOffVM_Task, 'duration_secs': 0.144853} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2511.584495] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-66195b79-c107-47ef-a529-3e9640305871 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2511.584690] env[62824]: DEBUG nova.compute.manager [None req-66195b79-c107-47ef-a529-3e9640305871 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2511.585544] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-402ecf15-b184-4e43-b7b2-8ba3d6d81166 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2512.098512] env[62824]: DEBUG oslo_concurrency.lockutils [None req-66195b79-c107-47ef-a529-3e9640305871 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lock "7c47bb92-a58e-4537-a68a-289e2a5fc19a" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.047s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2512.395932] env[62824]: DEBUG nova.network.neutron [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Port 62aa7735-5027-4b47-917b-7b769eac4cee binding to destination host cpu-1 is already ACTIVE {{(pid=62824) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 2512.925065] env[62824]: DEBUG nova.compute.manager [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Stashing vm_state: active {{(pid=62824) _prep_resize /opt/stack/nova/nova/compute/manager.py:6136}} [ 2513.047440] env[62824]: DEBUG oslo_concurrency.lockutils [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Acquiring lock "8bdfe1c4-1c23-4360-9bb0-90dfc9123d79" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2513.047681] env[62824]: DEBUG oslo_concurrency.lockutils [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Lock "8bdfe1c4-1c23-4360-9bb0-90dfc9123d79" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2513.300049] env[62824]: DEBUG nova.compute.manager [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Stashing vm_state: stopped {{(pid=62824) _prep_resize /opt/stack/nova/nova/compute/manager.py:6136}} [ 2513.418409] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquiring lock "e524ce78-6a63-4950-b727-ff72f8df4424-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2513.418409] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "e524ce78-6a63-4950-b727-ff72f8df4424-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2513.418600] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "e524ce78-6a63-4950-b727-ff72f8df4424-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2513.442257] env[62824]: DEBUG oslo_concurrency.lockutils [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2513.442574] env[62824]: DEBUG oslo_concurrency.lockutils [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2513.550484] env[62824]: DEBUG nova.compute.manager [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2513.816056] env[62824]: DEBUG oslo_concurrency.lockutils [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2513.947724] env[62824]: INFO nova.compute.claims [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2514.068638] env[62824]: DEBUG oslo_concurrency.lockutils [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2514.452041] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquiring lock "refresh_cache-e524ce78-6a63-4950-b727-ff72f8df4424" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2514.452268] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquired lock "refresh_cache-e524ce78-6a63-4950-b727-ff72f8df4424" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2514.452425] env[62824]: DEBUG nova.network.neutron [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2514.454884] env[62824]: INFO nova.compute.resource_tracker [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Updating resource usage from migration 2f209b10-6d6f-4fc7-bc13-333520807ce6 [ 2514.567999] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a6fb809-78af-4103-a711-5c4bd860088f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2514.576317] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23bd4da2-2e28-44a1-893e-715141162462 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2514.607450] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4f5af34-ad77-41e9-9c96-f2c3204d3c84 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2514.614852] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b3785f3-2864-413c-8fc7-04772a750bdb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2514.627744] env[62824]: DEBUG nova.compute.provider_tree [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2515.130779] env[62824]: DEBUG nova.scheduler.client.report [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2515.156556] env[62824]: DEBUG nova.network.neutron [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Updating instance_info_cache with network_info: [{"id": "62aa7735-5027-4b47-917b-7b769eac4cee", "address": "fa:16:3e:61:22:c9", "network": {"id": "b4ed285c-26a7-461b-8a71-d18e07582ab8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-997150974-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.185", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "815ecf9340934733a83d4dcb26612fd4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69f65356-c85e-4b7f-ad28-7c7b5e8cf50c", "external-id": "nsx-vlan-transportzone-281", "segmentation_id": 281, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap62aa7735-50", "ovs_interfaceid": "62aa7735-5027-4b47-917b-7b769eac4cee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2515.635736] env[62824]: DEBUG oslo_concurrency.lockutils [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.193s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2515.636055] env[62824]: INFO nova.compute.manager [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Migrating [ 2515.642438] env[62824]: DEBUG oslo_concurrency.lockutils [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 1.826s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2515.658768] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Releasing lock "refresh_cache-e524ce78-6a63-4950-b727-ff72f8df4424" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2516.153783] env[62824]: INFO nova.compute.claims [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2516.157592] env[62824]: DEBUG oslo_concurrency.lockutils [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquiring lock "refresh_cache-80c57382-8f6f-4537-b210-70a902b02cb4" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2516.157754] env[62824]: DEBUG oslo_concurrency.lockutils [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquired lock "refresh_cache-80c57382-8f6f-4537-b210-70a902b02cb4" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2516.157925] env[62824]: DEBUG nova.network.neutron [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2516.167233] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24d1dabe-b11c-4858-8f72-d6624045bd0e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2516.175116] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-714af8fa-df18-4368-a2e5-26e0e3d54b89 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2516.663101] env[62824]: INFO nova.compute.resource_tracker [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Updating resource usage from migration f2cec4b7-a19c-4954-8ad9-934ba57984e8 [ 2516.785009] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d50cbc4-d74e-4a04-aafe-98f401ac9fca {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2516.792385] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a818f51e-08c6-4120-bd02-99654d442277 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2516.827234] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92f3405c-20f2-4532-ad83-e8ccf4c48a80 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2516.836205] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0a536f8-f828-4a88-9b9c-efa7bd5bc16b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2516.848955] env[62824]: DEBUG nova.compute.provider_tree [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2516.891778] env[62824]: DEBUG nova.network.neutron [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Updating instance_info_cache with network_info: [{"id": "ae875573-c95e-4613-b343-e3e5e3829964", "address": "fa:16:3e:ac:dd:e4", "network": {"id": "bba9b98c-c7eb-4890-8de5-eafacfdebd7d", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1928394728-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.160", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7fcdbfc29df64a1ba8d982bdcc667b64", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b49c5024-2ced-42ca-90cc-6066766d43e6", "external-id": "nsx-vlan-transportzone-239", "segmentation_id": 239, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapae875573-c9", "ovs_interfaceid": "ae875573-c95e-4613-b343-e3e5e3829964", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2517.263415] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-447c70af-e07e-4c40-96e8-4c61b38898a1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2517.282873] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-404f7b89-40c7-497d-80c5-78fe9e5627a5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2517.289971] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Updating instance 'e524ce78-6a63-4950-b727-ff72f8df4424' progress to 83 {{(pid=62824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 2517.352204] env[62824]: DEBUG nova.scheduler.client.report [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2517.394562] env[62824]: DEBUG oslo_concurrency.lockutils [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Releasing lock "refresh_cache-80c57382-8f6f-4537-b210-70a902b02cb4" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2517.796519] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2517.796894] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b75d2689-bf89-44db-a7b1-cd5835c29cb4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2517.805617] env[62824]: DEBUG oslo_vmware.api [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2517.805617] env[62824]: value = "task-2147079" [ 2517.805617] env[62824]: _type = "Task" [ 2517.805617] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2517.814564] env[62824]: DEBUG oslo_vmware.api [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2147079, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2517.857837] env[62824]: DEBUG oslo_concurrency.lockutils [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.215s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2517.858080] env[62824]: INFO nova.compute.manager [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Migrating [ 2517.864773] env[62824]: DEBUG oslo_concurrency.lockutils [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.796s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2517.866455] env[62824]: INFO nova.compute.claims [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2518.317522] env[62824]: DEBUG oslo_vmware.api [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2147079, 'name': PowerOnVM_Task, 'duration_secs': 0.376391} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2518.317805] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2518.317969] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-5e99ff0a-b94a-41a4-a42b-c52953b8dbe7 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Updating instance 'e524ce78-6a63-4950-b727-ff72f8df4424' progress to 100 {{(pid=62824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 2518.378965] env[62824]: DEBUG oslo_concurrency.lockutils [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquiring lock "refresh_cache-7c47bb92-a58e-4537-a68a-289e2a5fc19a" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2518.379184] env[62824]: DEBUG oslo_concurrency.lockutils [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquired lock "refresh_cache-7c47bb92-a58e-4537-a68a-289e2a5fc19a" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2518.379315] env[62824]: DEBUG nova.network.neutron [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2518.911640] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acb0d474-1e6c-489c-a5ae-3165ffe4f6e4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2518.932757] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Updating instance '80c57382-8f6f-4537-b210-70a902b02cb4' progress to 0 {{(pid=62824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 2519.032919] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f964622d-af65-47e3-b43d-ef248b833741 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2519.039835] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afde3c4c-1197-4823-853f-5f1bb9c0ddd9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2519.071407] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ee63454-872b-4930-8bf2-18f67c9dec89 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2519.078461] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31cea29b-c594-480e-98a7-7cc32f377adb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2519.092542] env[62824]: DEBUG nova.compute.provider_tree [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2519.145723] env[62824]: DEBUG nova.network.neutron [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Updating instance_info_cache with network_info: [{"id": "821db814-549f-4b95-befb-cdb81b9bcb70", "address": "fa:16:3e:29:1d:69", "network": {"id": "5233a794-3514-4912-a857-7e76a66b1e8f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-910685188-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.129", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0eb41836305847fca773afebdafa01db", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd98a25d-a7a9-4fb5-8fef-e8df4dbbbf11", "external-id": "nsx-vlan-transportzone-707", "segmentation_id": 707, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap821db814-54", "ovs_interfaceid": "821db814-549f-4b95-befb-cdb81b9bcb70", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2519.438877] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2519.439201] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-48cb8559-8649-476d-95b2-1f076ada424b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2519.446871] env[62824]: DEBUG oslo_vmware.api [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2519.446871] env[62824]: value = "task-2147080" [ 2519.446871] env[62824]: _type = "Task" [ 2519.446871] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2519.454618] env[62824]: DEBUG oslo_vmware.api [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2147080, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2519.595432] env[62824]: DEBUG nova.scheduler.client.report [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2519.648447] env[62824]: DEBUG oslo_concurrency.lockutils [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Releasing lock "refresh_cache-7c47bb92-a58e-4537-a68a-289e2a5fc19a" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2519.957646] env[62824]: DEBUG oslo_vmware.api [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2147080, 'name': PowerOffVM_Task, 'duration_secs': 0.182064} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2519.957933] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2519.958137] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Updating instance '80c57382-8f6f-4537-b210-70a902b02cb4' progress to 17 {{(pid=62824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 2520.100407] env[62824]: DEBUG oslo_concurrency.lockutils [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.236s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2520.101079] env[62824]: DEBUG nova.compute.manager [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2520.465015] env[62824]: DEBUG nova.virt.hardware [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:24Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2520.465315] env[62824]: DEBUG nova.virt.hardware [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2520.465509] env[62824]: DEBUG nova.virt.hardware [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2520.465713] env[62824]: DEBUG nova.virt.hardware [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2520.465890] env[62824]: DEBUG nova.virt.hardware [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2520.466058] env[62824]: DEBUG nova.virt.hardware [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2520.466299] env[62824]: DEBUG nova.virt.hardware [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2520.466485] env[62824]: DEBUG nova.virt.hardware [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2520.466678] env[62824]: DEBUG nova.virt.hardware [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2520.466851] env[62824]: DEBUG nova.virt.hardware [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2520.467071] env[62824]: DEBUG nova.virt.hardware [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2520.472766] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-93aaac83-485f-4bfb-8767-b686b87fe11e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2520.489540] env[62824]: DEBUG oslo_vmware.api [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2520.489540] env[62824]: value = "task-2147081" [ 2520.489540] env[62824]: _type = "Task" [ 2520.489540] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2520.497858] env[62824]: DEBUG oslo_vmware.api [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2147081, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2520.606785] env[62824]: DEBUG nova.compute.utils [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2520.608274] env[62824]: DEBUG nova.compute.manager [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2520.608449] env[62824]: DEBUG nova.network.neutron [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2520.657043] env[62824]: DEBUG nova.policy [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cb36c701db8a471bbd65d2b68ce3becf', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c95aa87980de4ec8a2170c2640423301', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2520.704237] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5485d1bc-caa6-421f-8a60-298e72c01f2c tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquiring lock "e524ce78-6a63-4950-b727-ff72f8df4424" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2520.704497] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5485d1bc-caa6-421f-8a60-298e72c01f2c tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "e524ce78-6a63-4950-b727-ff72f8df4424" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2520.704687] env[62824]: DEBUG nova.compute.manager [None req-5485d1bc-caa6-421f-8a60-298e72c01f2c tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Going to confirm migration 6 {{(pid=62824) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5250}} [ 2520.953528] env[62824]: DEBUG nova.network.neutron [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] Successfully created port: 1013ac0c-ff48-42e0-8fe5-07f23015ac08 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2521.000666] env[62824]: DEBUG oslo_vmware.api [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2147081, 'name': ReconfigVM_Task, 'duration_secs': 0.151297} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2521.001065] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Updating instance '80c57382-8f6f-4537-b210-70a902b02cb4' progress to 33 {{(pid=62824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 2521.114663] env[62824]: DEBUG nova.compute.manager [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2521.167225] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f087ad40-7144-45ad-9937-c1980b8b5a8f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2521.186542] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Updating instance '7c47bb92-a58e-4537-a68a-289e2a5fc19a' progress to 0 {{(pid=62824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 2521.277367] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5485d1bc-caa6-421f-8a60-298e72c01f2c tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquiring lock "refresh_cache-e524ce78-6a63-4950-b727-ff72f8df4424" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2521.277554] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5485d1bc-caa6-421f-8a60-298e72c01f2c tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquired lock "refresh_cache-e524ce78-6a63-4950-b727-ff72f8df4424" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2521.277734] env[62824]: DEBUG nova.network.neutron [None req-5485d1bc-caa6-421f-8a60-298e72c01f2c tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2521.277925] env[62824]: DEBUG nova.objects.instance [None req-5485d1bc-caa6-421f-8a60-298e72c01f2c tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lazy-loading 'info_cache' on Instance uuid e524ce78-6a63-4950-b727-ff72f8df4424 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2521.509743] env[62824]: DEBUG nova.virt.hardware [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2521.510009] env[62824]: DEBUG nova.virt.hardware [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2521.510182] env[62824]: DEBUG nova.virt.hardware [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2521.510372] env[62824]: DEBUG nova.virt.hardware [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2521.510523] env[62824]: DEBUG nova.virt.hardware [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2521.510673] env[62824]: DEBUG nova.virt.hardware [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2521.510891] env[62824]: DEBUG nova.virt.hardware [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2521.511103] env[62824]: DEBUG nova.virt.hardware [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2521.511292] env[62824]: DEBUG nova.virt.hardware [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2521.511462] env[62824]: DEBUG nova.virt.hardware [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2521.511643] env[62824]: DEBUG nova.virt.hardware [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2521.517081] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Reconfiguring VM instance instance-00000074 to detach disk 2000 {{(pid=62824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 2521.517376] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7f37dcf1-db65-42a2-90ea-0b8aa744001b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2521.536857] env[62824]: DEBUG oslo_vmware.api [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2521.536857] env[62824]: value = "task-2147082" [ 2521.536857] env[62824]: _type = "Task" [ 2521.536857] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2521.545006] env[62824]: DEBUG oslo_vmware.api [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2147082, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2521.692061] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2521.692612] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-408fafff-aeca-4e37-a1af-546e631558c0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2521.700169] env[62824]: DEBUG oslo_vmware.api [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2521.700169] env[62824]: value = "task-2147083" [ 2521.700169] env[62824]: _type = "Task" [ 2521.700169] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2521.708250] env[62824]: DEBUG oslo_vmware.api [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147083, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2522.046194] env[62824]: DEBUG oslo_vmware.api [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2147082, 'name': ReconfigVM_Task, 'duration_secs': 0.183886} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2522.046493] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Reconfigured VM instance instance-00000074 to detach disk 2000 {{(pid=62824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 2522.047256] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cd83975-1f48-4920-864e-dd39c9666488 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2522.070238] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Reconfiguring VM instance instance-00000074 to attach disk [datastore1] 80c57382-8f6f-4537-b210-70a902b02cb4/80c57382-8f6f-4537-b210-70a902b02cb4.vmdk or device None with type thin {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2522.070464] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cc83f314-80c6-4242-a234-dafaae6156c8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2522.087254] env[62824]: DEBUG oslo_vmware.api [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2522.087254] env[62824]: value = "task-2147084" [ 2522.087254] env[62824]: _type = "Task" [ 2522.087254] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2522.094860] env[62824]: DEBUG oslo_vmware.api [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2147084, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2522.125575] env[62824]: DEBUG nova.compute.manager [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2522.152743] env[62824]: DEBUG nova.virt.hardware [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2522.153081] env[62824]: DEBUG nova.virt.hardware [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2522.153262] env[62824]: DEBUG nova.virt.hardware [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2522.153451] env[62824]: DEBUG nova.virt.hardware [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2522.153601] env[62824]: DEBUG nova.virt.hardware [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2522.153749] env[62824]: DEBUG nova.virt.hardware [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2522.153967] env[62824]: DEBUG nova.virt.hardware [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2522.154204] env[62824]: DEBUG nova.virt.hardware [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2522.154356] env[62824]: DEBUG nova.virt.hardware [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2522.154573] env[62824]: DEBUG nova.virt.hardware [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2522.154756] env[62824]: DEBUG nova.virt.hardware [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2522.155629] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59011f76-f8cc-4177-bae0-167f42e5a846 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2522.163358] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d0962d8-c42d-4002-a0f1-4554206e6266 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2522.209732] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] VM already powered off {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 2522.209902] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Updating instance '7c47bb92-a58e-4537-a68a-289e2a5fc19a' progress to 17 {{(pid=62824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 2522.498137] env[62824]: DEBUG nova.network.neutron [None req-5485d1bc-caa6-421f-8a60-298e72c01f2c tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Updating instance_info_cache with network_info: [{"id": "62aa7735-5027-4b47-917b-7b769eac4cee", "address": "fa:16:3e:61:22:c9", "network": {"id": "b4ed285c-26a7-461b-8a71-d18e07582ab8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-997150974-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.185", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "815ecf9340934733a83d4dcb26612fd4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69f65356-c85e-4b7f-ad28-7c7b5e8cf50c", "external-id": "nsx-vlan-transportzone-281", "segmentation_id": 281, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap62aa7735-50", "ovs_interfaceid": "62aa7735-5027-4b47-917b-7b769eac4cee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2522.600850] env[62824]: DEBUG oslo_vmware.api [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2147084, 'name': ReconfigVM_Task, 'duration_secs': 0.250462} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2522.600850] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Reconfigured VM instance instance-00000074 to attach disk [datastore1] 80c57382-8f6f-4537-b210-70a902b02cb4/80c57382-8f6f-4537-b210-70a902b02cb4.vmdk or device None with type thin {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2522.600850] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Updating instance '80c57382-8f6f-4537-b210-70a902b02cb4' progress to 50 {{(pid=62824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 2522.717255] env[62824]: DEBUG nova.virt.hardware [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:24Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2522.717459] env[62824]: DEBUG nova.virt.hardware [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2522.717613] env[62824]: DEBUG nova.virt.hardware [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2522.717801] env[62824]: DEBUG nova.virt.hardware [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2522.717950] env[62824]: DEBUG nova.virt.hardware [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2522.718127] env[62824]: DEBUG nova.virt.hardware [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2522.718330] env[62824]: DEBUG nova.virt.hardware [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2522.718490] env[62824]: DEBUG nova.virt.hardware [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2522.718661] env[62824]: DEBUG nova.virt.hardware [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2522.718825] env[62824]: DEBUG nova.virt.hardware [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2522.719012] env[62824]: DEBUG nova.virt.hardware [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2522.725031] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eaa9330c-6fa5-4077-959e-f6be5b68d441 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2522.742434] env[62824]: DEBUG oslo_vmware.api [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2522.742434] env[62824]: value = "task-2147085" [ 2522.742434] env[62824]: _type = "Task" [ 2522.742434] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2522.749981] env[62824]: DEBUG oslo_vmware.api [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147085, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2523.001213] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5485d1bc-caa6-421f-8a60-298e72c01f2c tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Releasing lock "refresh_cache-e524ce78-6a63-4950-b727-ff72f8df4424" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2523.001536] env[62824]: DEBUG nova.objects.instance [None req-5485d1bc-caa6-421f-8a60-298e72c01f2c tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lazy-loading 'migration_context' on Instance uuid e524ce78-6a63-4950-b727-ff72f8df4424 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2523.106324] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eeb4f6e9-4667-462c-a5a7-98ebb4ff9985 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2523.125200] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0142631b-5f3b-460f-87bd-57ab5ccee140 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2523.144258] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Updating instance '80c57382-8f6f-4537-b210-70a902b02cb4' progress to 67 {{(pid=62824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 2523.254247] env[62824]: DEBUG oslo_vmware.api [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147085, 'name': ReconfigVM_Task, 'duration_secs': 0.138919} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2523.254530] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Updating instance '7c47bb92-a58e-4537-a68a-289e2a5fc19a' progress to 33 {{(pid=62824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 2523.504987] env[62824]: DEBUG nova.objects.base [None req-5485d1bc-caa6-421f-8a60-298e72c01f2c tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=62824) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 2523.506433] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46e2250b-afee-4aee-8c95-73d70d33f91d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2523.529773] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-89ca13c3-c0c9-420e-80fa-cca935045f7b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2523.535346] env[62824]: DEBUG oslo_vmware.api [None req-5485d1bc-caa6-421f-8a60-298e72c01f2c tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2523.535346] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52adcb98-1045-ab4b-2883-1dc5a07d9252" [ 2523.535346] env[62824]: _type = "Task" [ 2523.535346] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2523.543336] env[62824]: DEBUG oslo_vmware.api [None req-5485d1bc-caa6-421f-8a60-298e72c01f2c tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52adcb98-1045-ab4b-2883-1dc5a07d9252, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2523.683831] env[62824]: DEBUG nova.network.neutron [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Port ae875573-c95e-4613-b343-e3e5e3829964 binding to destination host cpu-1 is already ACTIVE {{(pid=62824) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 2523.761263] env[62824]: DEBUG nova.virt.hardware [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2523.761496] env[62824]: DEBUG nova.virt.hardware [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2523.761656] env[62824]: DEBUG nova.virt.hardware [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2523.761837] env[62824]: DEBUG nova.virt.hardware [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2523.761983] env[62824]: DEBUG nova.virt.hardware [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2523.762151] env[62824]: DEBUG nova.virt.hardware [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2523.762357] env[62824]: DEBUG nova.virt.hardware [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2523.762515] env[62824]: DEBUG nova.virt.hardware [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2523.762685] env[62824]: DEBUG nova.virt.hardware [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2523.762851] env[62824]: DEBUG nova.virt.hardware [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2523.763112] env[62824]: DEBUG nova.virt.hardware [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2523.768399] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Reconfiguring VM instance instance-0000006f to detach disk 2000 {{(pid=62824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 2523.768917] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eea9c972-8b8c-4239-ba57-f0851ed37f63 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2523.789580] env[62824]: DEBUG oslo_vmware.api [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2523.789580] env[62824]: value = "task-2147086" [ 2523.789580] env[62824]: _type = "Task" [ 2523.789580] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2523.799884] env[62824]: DEBUG oslo_vmware.api [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147086, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2524.046107] env[62824]: DEBUG oslo_vmware.api [None req-5485d1bc-caa6-421f-8a60-298e72c01f2c tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52adcb98-1045-ab4b-2883-1dc5a07d9252, 'name': SearchDatastore_Task, 'duration_secs': 0.00874} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2524.046429] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5485d1bc-caa6-421f-8a60-298e72c01f2c tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2524.046667] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5485d1bc-caa6-421f-8a60-298e72c01f2c tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2524.299063] env[62824]: DEBUG oslo_vmware.api [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147086, 'name': ReconfigVM_Task, 'duration_secs': 0.150716} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2524.299306] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Reconfigured VM instance instance-0000006f to detach disk 2000 {{(pid=62824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 2524.300085] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b33edfc-c9b1-493e-bd19-3c427942b9ed {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2524.323264] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Reconfiguring VM instance instance-0000006f to attach disk [datastore2] 7c47bb92-a58e-4537-a68a-289e2a5fc19a/7c47bb92-a58e-4537-a68a-289e2a5fc19a.vmdk or device None with type thin {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2524.323526] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-49df58f1-5e30-4c1a-93c0-9b97ba7bab8a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2524.342749] env[62824]: DEBUG oslo_vmware.api [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2524.342749] env[62824]: value = "task-2147087" [ 2524.342749] env[62824]: _type = "Task" [ 2524.342749] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2524.350848] env[62824]: DEBUG oslo_vmware.api [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147087, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2524.665620] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6808ed9-ed33-4e48-9b8e-e1d7b238d11b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2524.673517] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b73e2a1-5fd6-4206-88f5-85b7f9d7d0a5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2524.710187] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43a246ec-f1df-4e2f-955d-5d3278d67cfa {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2524.718899] env[62824]: DEBUG oslo_concurrency.lockutils [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquiring lock "80c57382-8f6f-4537-b210-70a902b02cb4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2524.719141] env[62824]: DEBUG oslo_concurrency.lockutils [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lock "80c57382-8f6f-4537-b210-70a902b02cb4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2524.719322] env[62824]: DEBUG oslo_concurrency.lockutils [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lock "80c57382-8f6f-4537-b210-70a902b02cb4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2524.725752] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4962919-8eb7-499e-a6a8-55f329b2c454 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2524.740603] env[62824]: DEBUG nova.compute.provider_tree [None req-5485d1bc-caa6-421f-8a60-298e72c01f2c tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2524.852986] env[62824]: DEBUG oslo_vmware.api [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147087, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2525.245075] env[62824]: DEBUG nova.scheduler.client.report [None req-5485d1bc-caa6-421f-8a60-298e72c01f2c tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2525.354126] env[62824]: DEBUG oslo_vmware.api [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147087, 'name': ReconfigVM_Task, 'duration_secs': 0.801947} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2525.354513] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Reconfigured VM instance instance-0000006f to attach disk [datastore2] 7c47bb92-a58e-4537-a68a-289e2a5fc19a/7c47bb92-a58e-4537-a68a-289e2a5fc19a.vmdk or device None with type thin {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2525.354637] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Updating instance '7c47bb92-a58e-4537-a68a-289e2a5fc19a' progress to 50 {{(pid=62824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 2525.765388] env[62824]: DEBUG oslo_concurrency.lockutils [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquiring lock "refresh_cache-80c57382-8f6f-4537-b210-70a902b02cb4" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2525.765587] env[62824]: DEBUG oslo_concurrency.lockutils [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquired lock "refresh_cache-80c57382-8f6f-4537-b210-70a902b02cb4" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2525.765771] env[62824]: DEBUG nova.network.neutron [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2525.862197] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3315fda5-bf64-413d-ba0b-3b0703664351 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2525.882375] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e098025d-af35-49cd-b9d7-61e7f44d3a61 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2525.886363] env[62824]: DEBUG nova.compute.manager [req-123c5139-1190-4f99-95a5-6a1ae38e7069 req-59f76160-9d74-4c77-89fd-03c7bee8fd57 service nova] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] Received event network-vif-plugged-1013ac0c-ff48-42e0-8fe5-07f23015ac08 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2525.886745] env[62824]: DEBUG oslo_concurrency.lockutils [req-123c5139-1190-4f99-95a5-6a1ae38e7069 req-59f76160-9d74-4c77-89fd-03c7bee8fd57 service nova] Acquiring lock "8bdfe1c4-1c23-4360-9bb0-90dfc9123d79-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2525.888464] env[62824]: DEBUG oslo_concurrency.lockutils [req-123c5139-1190-4f99-95a5-6a1ae38e7069 req-59f76160-9d74-4c77-89fd-03c7bee8fd57 service nova] Lock "8bdfe1c4-1c23-4360-9bb0-90dfc9123d79-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2525.888464] env[62824]: DEBUG oslo_concurrency.lockutils [req-123c5139-1190-4f99-95a5-6a1ae38e7069 req-59f76160-9d74-4c77-89fd-03c7bee8fd57 service nova] Lock "8bdfe1c4-1c23-4360-9bb0-90dfc9123d79-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2525.888464] env[62824]: DEBUG nova.compute.manager [req-123c5139-1190-4f99-95a5-6a1ae38e7069 req-59f76160-9d74-4c77-89fd-03c7bee8fd57 service nova] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] No waiting events found dispatching network-vif-plugged-1013ac0c-ff48-42e0-8fe5-07f23015ac08 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2525.888464] env[62824]: WARNING nova.compute.manager [req-123c5139-1190-4f99-95a5-6a1ae38e7069 req-59f76160-9d74-4c77-89fd-03c7bee8fd57 service nova] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] Received unexpected event network-vif-plugged-1013ac0c-ff48-42e0-8fe5-07f23015ac08 for instance with vm_state building and task_state spawning. [ 2525.907378] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Updating instance '7c47bb92-a58e-4537-a68a-289e2a5fc19a' progress to 67 {{(pid=62824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 2525.955281] env[62824]: DEBUG nova.network.neutron [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] Successfully updated port: 1013ac0c-ff48-42e0-8fe5-07f23015ac08 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2526.253992] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5485d1bc-caa6-421f-8a60-298e72c01f2c tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.207s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2526.458597] env[62824]: DEBUG oslo_concurrency.lockutils [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Acquiring lock "refresh_cache-8bdfe1c4-1c23-4360-9bb0-90dfc9123d79" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2526.458974] env[62824]: DEBUG oslo_concurrency.lockutils [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Acquired lock "refresh_cache-8bdfe1c4-1c23-4360-9bb0-90dfc9123d79" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2526.458974] env[62824]: DEBUG nova.network.neutron [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2526.474500] env[62824]: DEBUG nova.network.neutron [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Port 821db814-549f-4b95-befb-cdb81b9bcb70 binding to destination host cpu-1 is already ACTIVE {{(pid=62824) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 2526.513139] env[62824]: DEBUG nova.network.neutron [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Updating instance_info_cache with network_info: [{"id": "ae875573-c95e-4613-b343-e3e5e3829964", "address": "fa:16:3e:ac:dd:e4", "network": {"id": "bba9b98c-c7eb-4890-8de5-eafacfdebd7d", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1928394728-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.160", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7fcdbfc29df64a1ba8d982bdcc667b64", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b49c5024-2ced-42ca-90cc-6066766d43e6", "external-id": "nsx-vlan-transportzone-239", "segmentation_id": 239, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapae875573-c9", "ovs_interfaceid": "ae875573-c95e-4613-b343-e3e5e3829964", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2526.813936] env[62824]: INFO nova.scheduler.client.report [None req-5485d1bc-caa6-421f-8a60-298e72c01f2c tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Deleted allocation for migration 4c723d9e-36f1-43f6-b8f2-091f294e414f [ 2526.988589] env[62824]: DEBUG nova.network.neutron [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2527.016259] env[62824]: DEBUG oslo_concurrency.lockutils [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Releasing lock "refresh_cache-80c57382-8f6f-4537-b210-70a902b02cb4" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2527.042361] env[62824]: INFO nova.compute.manager [None req-5dfe3cd1-6815-4ac6-8d55-f1bdf4abc7e1 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Get console output [ 2527.042569] env[62824]: WARNING nova.virt.vmwareapi.driver [None req-5dfe3cd1-6815-4ac6-8d55-f1bdf4abc7e1 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] The console log is missing. Check your VSPC configuration [ 2527.118297] env[62824]: DEBUG nova.network.neutron [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] Updating instance_info_cache with network_info: [{"id": "1013ac0c-ff48-42e0-8fe5-07f23015ac08", "address": "fa:16:3e:df:4e:8c", "network": {"id": "f7936d74-eeef-4c41-8ae7-9427c43663fa", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1643883257-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c95aa87980de4ec8a2170c2640423301", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6eaa481-1f92-4851-b98e-09ed0daad7cb", "external-id": "nsx-vlan-transportzone-636", "segmentation_id": 636, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1013ac0c-ff", "ovs_interfaceid": "1013ac0c-ff48-42e0-8fe5-07f23015ac08", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2527.319470] env[62824]: DEBUG oslo_concurrency.lockutils [None req-5485d1bc-caa6-421f-8a60-298e72c01f2c tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "e524ce78-6a63-4950-b727-ff72f8df4424" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.615s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2527.497811] env[62824]: DEBUG oslo_concurrency.lockutils [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquiring lock "7c47bb92-a58e-4537-a68a-289e2a5fc19a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2527.498192] env[62824]: DEBUG oslo_concurrency.lockutils [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lock "7c47bb92-a58e-4537-a68a-289e2a5fc19a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2527.498238] env[62824]: DEBUG oslo_concurrency.lockutils [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lock "7c47bb92-a58e-4537-a68a-289e2a5fc19a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2527.536395] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f55f4107-ed84-401e-912e-208e4e8d2b19 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2527.555996] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acd5f5ed-277c-40d8-bc5c-88483ff6b927 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2527.562680] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Updating instance '80c57382-8f6f-4537-b210-70a902b02cb4' progress to 83 {{(pid=62824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 2527.620985] env[62824]: DEBUG oslo_concurrency.lockutils [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Releasing lock "refresh_cache-8bdfe1c4-1c23-4360-9bb0-90dfc9123d79" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2527.621284] env[62824]: DEBUG nova.compute.manager [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] Instance network_info: |[{"id": "1013ac0c-ff48-42e0-8fe5-07f23015ac08", "address": "fa:16:3e:df:4e:8c", "network": {"id": "f7936d74-eeef-4c41-8ae7-9427c43663fa", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1643883257-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c95aa87980de4ec8a2170c2640423301", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6eaa481-1f92-4851-b98e-09ed0daad7cb", "external-id": "nsx-vlan-transportzone-636", "segmentation_id": 636, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1013ac0c-ff", "ovs_interfaceid": "1013ac0c-ff48-42e0-8fe5-07f23015ac08", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2527.621688] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:df:4e:8c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c6eaa481-1f92-4851-b98e-09ed0daad7cb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1013ac0c-ff48-42e0-8fe5-07f23015ac08', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2527.629613] env[62824]: DEBUG oslo.service.loopingcall [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2527.629816] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2527.630051] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8b6f9cf4-fa53-4a84-b0be-653d8ff4d3f1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2527.651041] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2527.651041] env[62824]: value = "task-2147088" [ 2527.651041] env[62824]: _type = "Task" [ 2527.651041] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2527.662536] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2147088, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2527.895597] env[62824]: DEBUG nova.compute.manager [req-82ed4f0a-d745-42c5-b224-0ff789be1cec req-bd71b0f0-ed37-4905-b06d-908c42ddb982 service nova] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] Received event network-changed-1013ac0c-ff48-42e0-8fe5-07f23015ac08 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2527.895811] env[62824]: DEBUG nova.compute.manager [req-82ed4f0a-d745-42c5-b224-0ff789be1cec req-bd71b0f0-ed37-4905-b06d-908c42ddb982 service nova] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] Refreshing instance network info cache due to event network-changed-1013ac0c-ff48-42e0-8fe5-07f23015ac08. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2527.896046] env[62824]: DEBUG oslo_concurrency.lockutils [req-82ed4f0a-d745-42c5-b224-0ff789be1cec req-bd71b0f0-ed37-4905-b06d-908c42ddb982 service nova] Acquiring lock "refresh_cache-8bdfe1c4-1c23-4360-9bb0-90dfc9123d79" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2527.896198] env[62824]: DEBUG oslo_concurrency.lockutils [req-82ed4f0a-d745-42c5-b224-0ff789be1cec req-bd71b0f0-ed37-4905-b06d-908c42ddb982 service nova] Acquired lock "refresh_cache-8bdfe1c4-1c23-4360-9bb0-90dfc9123d79" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2527.896363] env[62824]: DEBUG nova.network.neutron [req-82ed4f0a-d745-42c5-b224-0ff789be1cec req-bd71b0f0-ed37-4905-b06d-908c42ddb982 service nova] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] Refreshing network info cache for port 1013ac0c-ff48-42e0-8fe5-07f23015ac08 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2528.069350] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2528.069723] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1dd554c8-f9bb-4fae-915d-ba54aad61beb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2528.078143] env[62824]: DEBUG oslo_vmware.api [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2528.078143] env[62824]: value = "task-2147089" [ 2528.078143] env[62824]: _type = "Task" [ 2528.078143] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2528.085863] env[62824]: DEBUG oslo_vmware.api [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2147089, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2528.160851] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2147088, 'name': CreateVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2528.546910] env[62824]: DEBUG oslo_concurrency.lockutils [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquiring lock "refresh_cache-7c47bb92-a58e-4537-a68a-289e2a5fc19a" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2528.547174] env[62824]: DEBUG oslo_concurrency.lockutils [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquired lock "refresh_cache-7c47bb92-a58e-4537-a68a-289e2a5fc19a" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2528.547323] env[62824]: DEBUG nova.network.neutron [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2528.589586] env[62824]: DEBUG oslo_vmware.api [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2147089, 'name': PowerOnVM_Task, 'duration_secs': 0.379207} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2528.589842] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2528.590064] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-995eab40-a717-49e6-b87d-33ba79e062f1 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Updating instance '80c57382-8f6f-4537-b210-70a902b02cb4' progress to 100 {{(pid=62824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 2528.622488] env[62824]: DEBUG nova.network.neutron [req-82ed4f0a-d745-42c5-b224-0ff789be1cec req-bd71b0f0-ed37-4905-b06d-908c42ddb982 service nova] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] Updated VIF entry in instance network info cache for port 1013ac0c-ff48-42e0-8fe5-07f23015ac08. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2528.622876] env[62824]: DEBUG nova.network.neutron [req-82ed4f0a-d745-42c5-b224-0ff789be1cec req-bd71b0f0-ed37-4905-b06d-908c42ddb982 service nova] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] Updating instance_info_cache with network_info: [{"id": "1013ac0c-ff48-42e0-8fe5-07f23015ac08", "address": "fa:16:3e:df:4e:8c", "network": {"id": "f7936d74-eeef-4c41-8ae7-9427c43663fa", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1643883257-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c95aa87980de4ec8a2170c2640423301", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6eaa481-1f92-4851-b98e-09ed0daad7cb", "external-id": "nsx-vlan-transportzone-636", "segmentation_id": 636, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1013ac0c-ff", "ovs_interfaceid": "1013ac0c-ff48-42e0-8fe5-07f23015ac08", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2528.662171] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2147088, 'name': CreateVM_Task, 'duration_secs': 0.514965} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2528.662348] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2528.663065] env[62824]: DEBUG oslo_concurrency.lockutils [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2528.664153] env[62824]: DEBUG oslo_concurrency.lockutils [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2528.664153] env[62824]: DEBUG oslo_concurrency.lockutils [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2528.664153] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f663e014-1cce-4861-97b0-14a21a995a46 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2528.668802] env[62824]: DEBUG oslo_vmware.api [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Waiting for the task: (returnval){ [ 2528.668802] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52341b44-34e4-0c73-9d7b-758049b9085d" [ 2528.668802] env[62824]: _type = "Task" [ 2528.668802] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2528.676159] env[62824]: DEBUG oslo_vmware.api [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52341b44-34e4-0c73-9d7b-758049b9085d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2529.126579] env[62824]: DEBUG oslo_concurrency.lockutils [req-82ed4f0a-d745-42c5-b224-0ff789be1cec req-bd71b0f0-ed37-4905-b06d-908c42ddb982 service nova] Releasing lock "refresh_cache-8bdfe1c4-1c23-4360-9bb0-90dfc9123d79" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2529.179304] env[62824]: DEBUG oslo_vmware.api [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52341b44-34e4-0c73-9d7b-758049b9085d, 'name': SearchDatastore_Task, 'duration_secs': 0.010193} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2529.179604] env[62824]: DEBUG oslo_concurrency.lockutils [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2529.179837] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2529.180074] env[62824]: DEBUG oslo_concurrency.lockutils [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2529.180226] env[62824]: DEBUG oslo_concurrency.lockutils [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2529.180402] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2529.180651] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c9a3712a-5311-45f8-a3b3-9122a0afd2cd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2529.188468] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2529.188640] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2529.189339] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-91d9cda0-cf38-4b2a-841d-59849ccd9265 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2529.193824] env[62824]: DEBUG oslo_vmware.api [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Waiting for the task: (returnval){ [ 2529.193824] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52dfd837-61c3-446b-3a9b-807e622d9497" [ 2529.193824] env[62824]: _type = "Task" [ 2529.193824] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2529.200854] env[62824]: DEBUG oslo_vmware.api [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52dfd837-61c3-446b-3a9b-807e622d9497, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2529.247344] env[62824]: DEBUG nova.network.neutron [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Updating instance_info_cache with network_info: [{"id": "821db814-549f-4b95-befb-cdb81b9bcb70", "address": "fa:16:3e:29:1d:69", "network": {"id": "5233a794-3514-4912-a857-7e76a66b1e8f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-910685188-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.129", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0eb41836305847fca773afebdafa01db", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd98a25d-a7a9-4fb5-8fef-e8df4dbbbf11", "external-id": "nsx-vlan-transportzone-707", "segmentation_id": 707, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap821db814-54", "ovs_interfaceid": "821db814-549f-4b95-befb-cdb81b9bcb70", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2529.704265] env[62824]: DEBUG oslo_vmware.api [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52dfd837-61c3-446b-3a9b-807e622d9497, 'name': SearchDatastore_Task, 'duration_secs': 0.008134} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2529.705089] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-485c15bb-b163-4ccd-9514-92890ca1423d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2529.710332] env[62824]: DEBUG oslo_vmware.api [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Waiting for the task: (returnval){ [ 2529.710332] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5281f36e-9f8f-3489-0e1c-d087fc22a7b9" [ 2529.710332] env[62824]: _type = "Task" [ 2529.710332] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2529.717564] env[62824]: DEBUG oslo_vmware.api [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5281f36e-9f8f-3489-0e1c-d087fc22a7b9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2529.750331] env[62824]: DEBUG oslo_concurrency.lockutils [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Releasing lock "refresh_cache-7c47bb92-a58e-4537-a68a-289e2a5fc19a" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2530.221234] env[62824]: DEBUG oslo_vmware.api [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5281f36e-9f8f-3489-0e1c-d087fc22a7b9, 'name': SearchDatastore_Task, 'duration_secs': 0.009767} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2530.221441] env[62824]: DEBUG oslo_concurrency.lockutils [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2530.221695] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79/8bdfe1c4-1c23-4360-9bb0-90dfc9123d79.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2530.221946] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5bdd4fa7-ebf7-4b93-b16e-3822d445d972 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2530.228440] env[62824]: DEBUG oslo_vmware.api [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Waiting for the task: (returnval){ [ 2530.228440] env[62824]: value = "task-2147090" [ 2530.228440] env[62824]: _type = "Task" [ 2530.228440] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2530.235787] env[62824]: DEBUG oslo_vmware.api [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147090, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2530.274303] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fba0807f-03a3-459c-89b6-dd0132966c79 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2530.292760] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-596616b5-5bce-4793-9d69-e80e37d576fb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2530.299545] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Updating instance '7c47bb92-a58e-4537-a68a-289e2a5fc19a' progress to 83 {{(pid=62824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 2530.490821] env[62824]: DEBUG nova.network.neutron [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Port ae875573-c95e-4613-b343-e3e5e3829964 binding to destination host cpu-1 is already ACTIVE {{(pid=62824) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 2530.491262] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquiring lock "refresh_cache-80c57382-8f6f-4537-b210-70a902b02cb4" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2530.491434] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquired lock "refresh_cache-80c57382-8f6f-4537-b210-70a902b02cb4" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2530.491621] env[62824]: DEBUG nova.network.neutron [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2530.739707] env[62824]: DEBUG oslo_vmware.api [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147090, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.448611} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2530.740149] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79/8bdfe1c4-1c23-4360-9bb0-90dfc9123d79.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2530.740245] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2530.740436] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cbecef92-7a92-42e4-a4e5-9156c501789f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2530.747132] env[62824]: DEBUG oslo_vmware.api [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Waiting for the task: (returnval){ [ 2530.747132] env[62824]: value = "task-2147091" [ 2530.747132] env[62824]: _type = "Task" [ 2530.747132] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2530.754609] env[62824]: DEBUG oslo_vmware.api [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147091, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2530.806191] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-131e2cf1-1171-4314-9873-cb5fdb166511 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Updating instance '7c47bb92-a58e-4537-a68a-289e2a5fc19a' progress to 100 {{(pid=62824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 2531.205988] env[62824]: DEBUG nova.network.neutron [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Updating instance_info_cache with network_info: [{"id": "ae875573-c95e-4613-b343-e3e5e3829964", "address": "fa:16:3e:ac:dd:e4", "network": {"id": "bba9b98c-c7eb-4890-8de5-eafacfdebd7d", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1928394728-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.160", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7fcdbfc29df64a1ba8d982bdcc667b64", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b49c5024-2ced-42ca-90cc-6066766d43e6", "external-id": "nsx-vlan-transportzone-239", "segmentation_id": 239, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapae875573-c9", "ovs_interfaceid": "ae875573-c95e-4613-b343-e3e5e3829964", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2531.257552] env[62824]: DEBUG oslo_vmware.api [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147091, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.13094} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2531.257801] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2531.258604] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6299366b-f768-4cd7-8d73-395015cdab7a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2531.280449] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] Reconfiguring VM instance instance-00000076 to attach disk [datastore1] 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79/8bdfe1c4-1c23-4360-9bb0-90dfc9123d79.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2531.280702] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c970d58a-ac55-46c8-8a3e-85ca56849d92 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2531.301030] env[62824]: DEBUG oslo_vmware.api [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Waiting for the task: (returnval){ [ 2531.301030] env[62824]: value = "task-2147092" [ 2531.301030] env[62824]: _type = "Task" [ 2531.301030] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2531.313622] env[62824]: DEBUG oslo_vmware.api [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147092, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2531.708981] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Releasing lock "refresh_cache-80c57382-8f6f-4537-b210-70a902b02cb4" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2531.811565] env[62824]: DEBUG oslo_vmware.api [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147092, 'name': ReconfigVM_Task, 'duration_secs': 0.325245} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2531.811931] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] Reconfigured VM instance instance-00000076 to attach disk [datastore1] 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79/8bdfe1c4-1c23-4360-9bb0-90dfc9123d79.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2531.812484] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cdd2ce2a-da0b-4d24-8f9c-240d6926fdd9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2531.819057] env[62824]: DEBUG oslo_vmware.api [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Waiting for the task: (returnval){ [ 2531.819057] env[62824]: value = "task-2147093" [ 2531.819057] env[62824]: _type = "Task" [ 2531.819057] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2531.826686] env[62824]: DEBUG oslo_vmware.api [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147093, 'name': Rename_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2532.213277] env[62824]: DEBUG nova.compute.manager [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=62824) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:901}} [ 2532.213524] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2532.213759] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2532.335832] env[62824]: DEBUG oslo_vmware.api [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147093, 'name': Rename_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2532.717053] env[62824]: DEBUG nova.objects.instance [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lazy-loading 'migration_context' on Instance uuid 80c57382-8f6f-4537-b210-70a902b02cb4 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2532.835582] env[62824]: DEBUG oslo_vmware.api [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147093, 'name': Rename_Task, 'duration_secs': 0.886736} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2532.835950] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2532.836058] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5954c039-9b82-4679-8072-d6032c16806a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2532.841562] env[62824]: DEBUG oslo_vmware.api [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Waiting for the task: (returnval){ [ 2532.841562] env[62824]: value = "task-2147094" [ 2532.841562] env[62824]: _type = "Task" [ 2532.841562] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2532.848721] env[62824]: DEBUG oslo_vmware.api [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147094, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2533.081746] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f0b48ff0-09f4-4baf-bd37-eafcaf373e83 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquiring lock "7c47bb92-a58e-4537-a68a-289e2a5fc19a" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2533.082053] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f0b48ff0-09f4-4baf-bd37-eafcaf373e83 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lock "7c47bb92-a58e-4537-a68a-289e2a5fc19a" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2533.082288] env[62824]: DEBUG nova.compute.manager [None req-f0b48ff0-09f4-4baf-bd37-eafcaf373e83 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Going to confirm migration 8 {{(pid=62824) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5250}} [ 2533.329855] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-760c8366-9e8b-4687-8a50-07d333a6784a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2533.337917] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c39e0dd8-8335-4bbd-a218-44c568f398bc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2533.369477] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89d8d2e3-57f7-4a24-bd05-45e1de2cdcf3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2533.374533] env[62824]: DEBUG oslo_vmware.api [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147094, 'name': PowerOnVM_Task, 'duration_secs': 0.413843} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2533.375152] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2533.375417] env[62824]: INFO nova.compute.manager [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] Took 11.25 seconds to spawn the instance on the hypervisor. [ 2533.375616] env[62824]: DEBUG nova.compute.manager [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2533.376362] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf58b260-2847-49ba-9e2b-f532db94a851 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2533.383064] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e1feadc-1341-4b73-9fc1-49143dfa9ad8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2533.399779] env[62824]: DEBUG nova.compute.provider_tree [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2533.650100] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f0b48ff0-09f4-4baf-bd37-eafcaf373e83 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquiring lock "refresh_cache-7c47bb92-a58e-4537-a68a-289e2a5fc19a" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2533.650320] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f0b48ff0-09f4-4baf-bd37-eafcaf373e83 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquired lock "refresh_cache-7c47bb92-a58e-4537-a68a-289e2a5fc19a" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2533.650496] env[62824]: DEBUG nova.network.neutron [None req-f0b48ff0-09f4-4baf-bd37-eafcaf373e83 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2533.650687] env[62824]: DEBUG nova.objects.instance [None req-f0b48ff0-09f4-4baf-bd37-eafcaf373e83 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lazy-loading 'info_cache' on Instance uuid 7c47bb92-a58e-4537-a68a-289e2a5fc19a {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2533.899259] env[62824]: INFO nova.compute.manager [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] Took 19.84 seconds to build instance. [ 2533.902957] env[62824]: DEBUG nova.scheduler.client.report [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2534.127876] env[62824]: DEBUG nova.compute.manager [req-97863c9a-1705-4ac3-ac5f-24b332207a39 req-13eff329-9c0d-4a81-a147-ea1a9d7fa0a3 service nova] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] Received event network-changed-1013ac0c-ff48-42e0-8fe5-07f23015ac08 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2534.128154] env[62824]: DEBUG nova.compute.manager [req-97863c9a-1705-4ac3-ac5f-24b332207a39 req-13eff329-9c0d-4a81-a147-ea1a9d7fa0a3 service nova] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] Refreshing instance network info cache due to event network-changed-1013ac0c-ff48-42e0-8fe5-07f23015ac08. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2534.128309] env[62824]: DEBUG oslo_concurrency.lockutils [req-97863c9a-1705-4ac3-ac5f-24b332207a39 req-13eff329-9c0d-4a81-a147-ea1a9d7fa0a3 service nova] Acquiring lock "refresh_cache-8bdfe1c4-1c23-4360-9bb0-90dfc9123d79" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2534.128446] env[62824]: DEBUG oslo_concurrency.lockutils [req-97863c9a-1705-4ac3-ac5f-24b332207a39 req-13eff329-9c0d-4a81-a147-ea1a9d7fa0a3 service nova] Acquired lock "refresh_cache-8bdfe1c4-1c23-4360-9bb0-90dfc9123d79" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2534.128606] env[62824]: DEBUG nova.network.neutron [req-97863c9a-1705-4ac3-ac5f-24b332207a39 req-13eff329-9c0d-4a81-a147-ea1a9d7fa0a3 service nova] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] Refreshing network info cache for port 1013ac0c-ff48-42e0-8fe5-07f23015ac08 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2534.401241] env[62824]: DEBUG oslo_concurrency.lockutils [None req-51280f85-6712-42d7-a859-0dd2e88a2109 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Lock "8bdfe1c4-1c23-4360-9bb0-90dfc9123d79" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.353s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2534.914034] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.700s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2535.150700] env[62824]: DEBUG nova.network.neutron [req-97863c9a-1705-4ac3-ac5f-24b332207a39 req-13eff329-9c0d-4a81-a147-ea1a9d7fa0a3 service nova] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] Updated VIF entry in instance network info cache for port 1013ac0c-ff48-42e0-8fe5-07f23015ac08. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2535.151093] env[62824]: DEBUG nova.network.neutron [req-97863c9a-1705-4ac3-ac5f-24b332207a39 req-13eff329-9c0d-4a81-a147-ea1a9d7fa0a3 service nova] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] Updating instance_info_cache with network_info: [{"id": "1013ac0c-ff48-42e0-8fe5-07f23015ac08", "address": "fa:16:3e:df:4e:8c", "network": {"id": "f7936d74-eeef-4c41-8ae7-9427c43663fa", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1643883257-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.250", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c95aa87980de4ec8a2170c2640423301", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6eaa481-1f92-4851-b98e-09ed0daad7cb", "external-id": "nsx-vlan-transportzone-636", "segmentation_id": 636, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1013ac0c-ff", "ovs_interfaceid": "1013ac0c-ff48-42e0-8fe5-07f23015ac08", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2535.164869] env[62824]: DEBUG nova.network.neutron [None req-f0b48ff0-09f4-4baf-bd37-eafcaf373e83 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Updating instance_info_cache with network_info: [{"id": "821db814-549f-4b95-befb-cdb81b9bcb70", "address": "fa:16:3e:29:1d:69", "network": {"id": "5233a794-3514-4912-a857-7e76a66b1e8f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-910685188-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.129", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0eb41836305847fca773afebdafa01db", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd98a25d-a7a9-4fb5-8fef-e8df4dbbbf11", "external-id": "nsx-vlan-transportzone-707", "segmentation_id": 707, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap821db814-54", "ovs_interfaceid": "821db814-549f-4b95-befb-cdb81b9bcb70", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2535.657664] env[62824]: DEBUG oslo_concurrency.lockutils [req-97863c9a-1705-4ac3-ac5f-24b332207a39 req-13eff329-9c0d-4a81-a147-ea1a9d7fa0a3 service nova] Releasing lock "refresh_cache-8bdfe1c4-1c23-4360-9bb0-90dfc9123d79" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2535.667577] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f0b48ff0-09f4-4baf-bd37-eafcaf373e83 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Releasing lock "refresh_cache-7c47bb92-a58e-4537-a68a-289e2a5fc19a" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2535.667822] env[62824]: DEBUG nova.objects.instance [None req-f0b48ff0-09f4-4baf-bd37-eafcaf373e83 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lazy-loading 'migration_context' on Instance uuid 7c47bb92-a58e-4537-a68a-289e2a5fc19a {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2536.171035] env[62824]: DEBUG nova.objects.base [None req-f0b48ff0-09f4-4baf-bd37-eafcaf373e83 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Object Instance<7c47bb92-a58e-4537-a68a-289e2a5fc19a> lazy-loaded attributes: info_cache,migration_context {{(pid=62824) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 2536.171816] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60e854d3-ac2f-407e-bd9f-d3c46fe0c5d1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2536.191990] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-91f5bc9a-1927-468d-945b-ba423763d4b7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2536.197230] env[62824]: DEBUG oslo_vmware.api [None req-f0b48ff0-09f4-4baf-bd37-eafcaf373e83 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2536.197230] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]529c8418-eb7f-d40b-1aa0-b46e22a688b7" [ 2536.197230] env[62824]: _type = "Task" [ 2536.197230] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2536.204656] env[62824]: DEBUG oslo_vmware.api [None req-f0b48ff0-09f4-4baf-bd37-eafcaf373e83 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]529c8418-eb7f-d40b-1aa0-b46e22a688b7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2536.457550] env[62824]: INFO nova.compute.manager [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Swapping old allocation on dict_keys(['bf95157b-a274-42de-9ccf-9851128a44a1']) held by migration 2f209b10-6d6f-4fc7-bc13-333520807ce6 for instance [ 2536.479102] env[62824]: DEBUG nova.scheduler.client.report [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Overwriting current allocation {'allocations': {'bf95157b-a274-42de-9ccf-9851128a44a1': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 167}}, 'project_id': '7fcdbfc29df64a1ba8d982bdcc667b64', 'user_id': '5a5891a75fee4caf92d4bba152254ad5', 'consumer_generation': 1} on consumer 80c57382-8f6f-4537-b210-70a902b02cb4 {{(pid=62824) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2036}} [ 2536.554409] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquiring lock "refresh_cache-80c57382-8f6f-4537-b210-70a902b02cb4" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2536.554603] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquired lock "refresh_cache-80c57382-8f6f-4537-b210-70a902b02cb4" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2536.554785] env[62824]: DEBUG nova.network.neutron [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2536.707681] env[62824]: DEBUG oslo_vmware.api [None req-f0b48ff0-09f4-4baf-bd37-eafcaf373e83 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]529c8418-eb7f-d40b-1aa0-b46e22a688b7, 'name': SearchDatastore_Task, 'duration_secs': 0.007475} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2536.707953] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f0b48ff0-09f4-4baf-bd37-eafcaf373e83 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2536.708191] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f0b48ff0-09f4-4baf-bd37-eafcaf373e83 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2537.269941] env[62824]: DEBUG nova.network.neutron [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Updating instance_info_cache with network_info: [{"id": "ae875573-c95e-4613-b343-e3e5e3829964", "address": "fa:16:3e:ac:dd:e4", "network": {"id": "bba9b98c-c7eb-4890-8de5-eafacfdebd7d", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1928394728-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.160", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7fcdbfc29df64a1ba8d982bdcc667b64", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b49c5024-2ced-42ca-90cc-6066766d43e6", "external-id": "nsx-vlan-transportzone-239", "segmentation_id": 239, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapae875573-c9", "ovs_interfaceid": "ae875573-c95e-4613-b343-e3e5e3829964", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2537.300427] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d6508e6-0295-4c4b-8790-2126f449ab0d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2537.307839] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49ff29bc-5c3f-48d7-94fe-a0a2da0e8e74 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2537.338721] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a6f89ce-6650-4b70-98de-a5b72ed1d118 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2537.345518] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1749dea5-a219-4271-89fa-7722f826df36 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2537.358326] env[62824]: DEBUG nova.compute.provider_tree [None req-f0b48ff0-09f4-4baf-bd37-eafcaf373e83 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2537.772649] env[62824]: DEBUG oslo_concurrency.lockutils [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Releasing lock "refresh_cache-80c57382-8f6f-4537-b210-70a902b02cb4" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2537.773110] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2537.773533] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bef359b9-6df6-42a7-95e8-18d6e2a97280 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2537.781977] env[62824]: DEBUG oslo_vmware.api [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2537.781977] env[62824]: value = "task-2147095" [ 2537.781977] env[62824]: _type = "Task" [ 2537.781977] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2537.789948] env[62824]: DEBUG oslo_vmware.api [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2147095, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2537.861240] env[62824]: DEBUG nova.scheduler.client.report [None req-f0b48ff0-09f4-4baf-bd37-eafcaf373e83 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2538.292512] env[62824]: DEBUG oslo_vmware.api [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2147095, 'name': PowerOffVM_Task, 'duration_secs': 0.216949} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2538.292883] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2538.293544] env[62824]: DEBUG nova.virt.hardware [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2538.293782] env[62824]: DEBUG nova.virt.hardware [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2538.293945] env[62824]: DEBUG nova.virt.hardware [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2538.294158] env[62824]: DEBUG nova.virt.hardware [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2538.294311] env[62824]: DEBUG nova.virt.hardware [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2538.294461] env[62824]: DEBUG nova.virt.hardware [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2538.294662] env[62824]: DEBUG nova.virt.hardware [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2538.294823] env[62824]: DEBUG nova.virt.hardware [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2538.295034] env[62824]: DEBUG nova.virt.hardware [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2538.295226] env[62824]: DEBUG nova.virt.hardware [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2538.295406] env[62824]: DEBUG nova.virt.hardware [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2538.300391] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6052fa0d-6a3d-455b-b1de-262ac55dc6af {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2538.315367] env[62824]: DEBUG oslo_vmware.api [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2538.315367] env[62824]: value = "task-2147096" [ 2538.315367] env[62824]: _type = "Task" [ 2538.315367] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2538.322780] env[62824]: DEBUG oslo_vmware.api [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2147096, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2538.826105] env[62824]: DEBUG oslo_vmware.api [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2147096, 'name': ReconfigVM_Task, 'duration_secs': 0.127994} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2538.826892] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13cc37c1-2270-4935-93ed-00543303dc34 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2538.844335] env[62824]: DEBUG nova.virt.hardware [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2538.844577] env[62824]: DEBUG nova.virt.hardware [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2538.844740] env[62824]: DEBUG nova.virt.hardware [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2538.844929] env[62824]: DEBUG nova.virt.hardware [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2538.845095] env[62824]: DEBUG nova.virt.hardware [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2538.845253] env[62824]: DEBUG nova.virt.hardware [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2538.845466] env[62824]: DEBUG nova.virt.hardware [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2538.845632] env[62824]: DEBUG nova.virt.hardware [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2538.845800] env[62824]: DEBUG nova.virt.hardware [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2538.845963] env[62824]: DEBUG nova.virt.hardware [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2538.846155] env[62824]: DEBUG nova.virt.hardware [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2538.846895] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cd6de0f9-c059-4c73-9fea-1b6c5c78d910 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2538.851649] env[62824]: DEBUG oslo_vmware.api [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2538.851649] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]521f0077-bb90-9a2b-c514-baa2d9e07a94" [ 2538.851649] env[62824]: _type = "Task" [ 2538.851649] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2538.859219] env[62824]: DEBUG oslo_vmware.api [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]521f0077-bb90-9a2b-c514-baa2d9e07a94, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2538.870508] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f0b48ff0-09f4-4baf-bd37-eafcaf373e83 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.162s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2538.870711] env[62824]: DEBUG nova.compute.manager [None req-f0b48ff0-09f4-4baf-bd37-eafcaf373e83 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Resized/migrated instance is powered off. Setting vm_state to 'stopped'. {{(pid=62824) _confirm_resize /opt/stack/nova/nova/compute/manager.py:5376}} [ 2539.362793] env[62824]: DEBUG oslo_vmware.api [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]521f0077-bb90-9a2b-c514-baa2d9e07a94, 'name': SearchDatastore_Task, 'duration_secs': 0.009557} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2539.367634] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Reconfiguring VM instance instance-00000074 to detach disk 2000 {{(pid=62824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 2539.367920] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4c34db14-19b3-4664-b906-167ec42eabf7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2539.387664] env[62824]: DEBUG oslo_vmware.api [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2539.387664] env[62824]: value = "task-2147097" [ 2539.387664] env[62824]: _type = "Task" [ 2539.387664] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2539.396287] env[62824]: DEBUG oslo_vmware.api [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2147097, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2539.436748] env[62824]: INFO nova.scheduler.client.report [None req-f0b48ff0-09f4-4baf-bd37-eafcaf373e83 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Deleted allocation for migration f2cec4b7-a19c-4954-8ad9-934ba57984e8 [ 2539.898166] env[62824]: DEBUG oslo_vmware.api [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2147097, 'name': ReconfigVM_Task, 'duration_secs': 0.186589} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2539.898452] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Reconfigured VM instance instance-00000074 to detach disk 2000 {{(pid=62824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 2539.899246] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1075fb5-a713-46e7-8a05-b4d4d7ec7bb9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2539.921048] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Reconfiguring VM instance instance-00000074 to attach disk [datastore1] 80c57382-8f6f-4537-b210-70a902b02cb4/80c57382-8f6f-4537-b210-70a902b02cb4.vmdk or device None with type thin {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2539.921348] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b0c153a9-63ae-4b06-9aea-f6a4f95606dc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2539.939968] env[62824]: DEBUG oslo_vmware.api [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2539.939968] env[62824]: value = "task-2147098" [ 2539.939968] env[62824]: _type = "Task" [ 2539.939968] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2539.943669] env[62824]: DEBUG nova.objects.instance [None req-66583751-af8f-4919-9ed1-3f9bbe7e13c0 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lazy-loading 'flavor' on Instance uuid 7c47bb92-a58e-4537-a68a-289e2a5fc19a {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2539.945905] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f0b48ff0-09f4-4baf-bd37-eafcaf373e83 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lock "7c47bb92-a58e-4537-a68a-289e2a5fc19a" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.864s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2539.952364] env[62824]: DEBUG oslo_vmware.api [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2147098, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2540.452669] env[62824]: DEBUG oslo_vmware.api [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2147098, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2540.453493] env[62824]: DEBUG oslo_concurrency.lockutils [None req-66583751-af8f-4919-9ed1-3f9bbe7e13c0 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquiring lock "refresh_cache-7c47bb92-a58e-4537-a68a-289e2a5fc19a" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2540.453653] env[62824]: DEBUG oslo_concurrency.lockutils [None req-66583751-af8f-4919-9ed1-3f9bbe7e13c0 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquired lock "refresh_cache-7c47bb92-a58e-4537-a68a-289e2a5fc19a" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2540.453827] env[62824]: DEBUG nova.network.neutron [None req-66583751-af8f-4919-9ed1-3f9bbe7e13c0 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2540.454012] env[62824]: DEBUG nova.objects.instance [None req-66583751-af8f-4919-9ed1-3f9bbe7e13c0 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lazy-loading 'info_cache' on Instance uuid 7c47bb92-a58e-4537-a68a-289e2a5fc19a {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2540.950882] env[62824]: DEBUG oslo_vmware.api [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2147098, 'name': ReconfigVM_Task, 'duration_secs': 0.833384} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2540.951197] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Reconfigured VM instance instance-00000074 to attach disk [datastore1] 80c57382-8f6f-4537-b210-70a902b02cb4/80c57382-8f6f-4537-b210-70a902b02cb4.vmdk or device None with type thin {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2540.952093] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74cfad85-2d94-4517-b252-8744cbf41fa3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2540.957428] env[62824]: DEBUG nova.objects.base [None req-66583751-af8f-4919-9ed1-3f9bbe7e13c0 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Object Instance<7c47bb92-a58e-4537-a68a-289e2a5fc19a> lazy-loaded attributes: flavor,info_cache {{(pid=62824) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 2540.975959] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8419f2b7-a929-4ed8-b0ef-775b95462e83 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2540.996258] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe76ca4a-5fa9-4afe-90a3-92ef2e83a4a1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2541.016876] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3901680-e134-4a7e-97a9-26c51a8f56cf {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2541.024122] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2541.024257] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8708d4ff-4298-4954-b43c-ffc6b970ccfc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2541.030028] env[62824]: DEBUG oslo_vmware.api [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2541.030028] env[62824]: value = "task-2147099" [ 2541.030028] env[62824]: _type = "Task" [ 2541.030028] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2541.037725] env[62824]: DEBUG oslo_vmware.api [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2147099, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2541.204332] env[62824]: DEBUG nova.network.neutron [None req-66583751-af8f-4919-9ed1-3f9bbe7e13c0 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Updating instance_info_cache with network_info: [{"id": "821db814-549f-4b95-befb-cdb81b9bcb70", "address": "fa:16:3e:29:1d:69", "network": {"id": "5233a794-3514-4912-a857-7e76a66b1e8f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-910685188-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.129", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0eb41836305847fca773afebdafa01db", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd98a25d-a7a9-4fb5-8fef-e8df4dbbbf11", "external-id": "nsx-vlan-transportzone-707", "segmentation_id": 707, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap821db814-54", "ovs_interfaceid": "821db814-549f-4b95-befb-cdb81b9bcb70", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2541.539678] env[62824]: DEBUG oslo_vmware.api [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2147099, 'name': PowerOnVM_Task, 'duration_secs': 0.362551} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2541.540097] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2541.707381] env[62824]: DEBUG oslo_concurrency.lockutils [None req-66583751-af8f-4919-9ed1-3f9bbe7e13c0 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Releasing lock "refresh_cache-7c47bb92-a58e-4537-a68a-289e2a5fc19a" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2542.551073] env[62824]: INFO nova.compute.manager [None req-3439a84e-626c-4a97-a634-e62310e7e85e tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Updating instance to original state: 'active' [ 2542.713679] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-66583751-af8f-4919-9ed1-3f9bbe7e13c0 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2542.714013] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b1054fad-e1a0-4058-a608-63ed1610284b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2542.724183] env[62824]: DEBUG oslo_vmware.api [None req-66583751-af8f-4919-9ed1-3f9bbe7e13c0 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2542.724183] env[62824]: value = "task-2147100" [ 2542.724183] env[62824]: _type = "Task" [ 2542.724183] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2542.731692] env[62824]: DEBUG oslo_vmware.api [None req-66583751-af8f-4919-9ed1-3f9bbe7e13c0 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147100, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2543.234941] env[62824]: DEBUG oslo_vmware.api [None req-66583751-af8f-4919-9ed1-3f9bbe7e13c0 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147100, 'name': PowerOnVM_Task, 'duration_secs': 0.480607} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2543.235274] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-66583751-af8f-4919-9ed1-3f9bbe7e13c0 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2543.235484] env[62824]: DEBUG nova.compute.manager [None req-66583751-af8f-4919-9ed1-3f9bbe7e13c0 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2543.236289] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c73c51e-f131-4076-a7b2-cedfe8b47b90 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2544.000547] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0ddedebe-e92c-4ef1-8b0f-b28af2a6c9e8 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquiring lock "80c57382-8f6f-4537-b210-70a902b02cb4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2544.000891] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0ddedebe-e92c-4ef1-8b0f-b28af2a6c9e8 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lock "80c57382-8f6f-4537-b210-70a902b02cb4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2544.001033] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0ddedebe-e92c-4ef1-8b0f-b28af2a6c9e8 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquiring lock "80c57382-8f6f-4537-b210-70a902b02cb4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2544.001513] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0ddedebe-e92c-4ef1-8b0f-b28af2a6c9e8 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lock "80c57382-8f6f-4537-b210-70a902b02cb4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2544.001513] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0ddedebe-e92c-4ef1-8b0f-b28af2a6c9e8 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lock "80c57382-8f6f-4537-b210-70a902b02cb4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2544.003509] env[62824]: INFO nova.compute.manager [None req-0ddedebe-e92c-4ef1-8b0f-b28af2a6c9e8 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Terminating instance [ 2544.508220] env[62824]: DEBUG nova.compute.manager [None req-0ddedebe-e92c-4ef1-8b0f-b28af2a6c9e8 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2544.508610] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-0ddedebe-e92c-4ef1-8b0f-b28af2a6c9e8 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2544.509986] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e345eb7-bf52-4ce6-93b3-7368685eea2f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2544.520709] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ddedebe-e92c-4ef1-8b0f-b28af2a6c9e8 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2544.521205] env[62824]: DEBUG oslo_concurrency.lockutils [None req-eab809bd-65a3-460a-a578-09415b9e2371 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquiring lock "7c47bb92-a58e-4537-a68a-289e2a5fc19a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2544.521527] env[62824]: DEBUG oslo_concurrency.lockutils [None req-eab809bd-65a3-460a-a578-09415b9e2371 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lock "7c47bb92-a58e-4537-a68a-289e2a5fc19a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2544.521833] env[62824]: DEBUG oslo_concurrency.lockutils [None req-eab809bd-65a3-460a-a578-09415b9e2371 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquiring lock "7c47bb92-a58e-4537-a68a-289e2a5fc19a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2544.522148] env[62824]: DEBUG oslo_concurrency.lockutils [None req-eab809bd-65a3-460a-a578-09415b9e2371 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lock "7c47bb92-a58e-4537-a68a-289e2a5fc19a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2544.522430] env[62824]: DEBUG oslo_concurrency.lockutils [None req-eab809bd-65a3-460a-a578-09415b9e2371 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lock "7c47bb92-a58e-4537-a68a-289e2a5fc19a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2544.525080] env[62824]: INFO nova.compute.manager [None req-eab809bd-65a3-460a-a578-09415b9e2371 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Terminating instance [ 2544.528067] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-da0f0b02-c891-47ce-acb3-64b004224c68 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2544.535838] env[62824]: DEBUG oslo_vmware.api [None req-0ddedebe-e92c-4ef1-8b0f-b28af2a6c9e8 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2544.535838] env[62824]: value = "task-2147101" [ 2544.535838] env[62824]: _type = "Task" [ 2544.535838] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2544.547938] env[62824]: DEBUG oslo_vmware.api [None req-0ddedebe-e92c-4ef1-8b0f-b28af2a6c9e8 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2147101, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2545.032808] env[62824]: DEBUG nova.compute.manager [None req-eab809bd-65a3-460a-a578-09415b9e2371 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2545.033212] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-eab809bd-65a3-460a-a578-09415b9e2371 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2545.033973] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a673c83c-17eb-4db6-8a52-467535546f7b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2545.046594] env[62824]: DEBUG oslo_vmware.api [None req-0ddedebe-e92c-4ef1-8b0f-b28af2a6c9e8 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2147101, 'name': PowerOffVM_Task, 'duration_secs': 0.182471} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2545.048625] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ddedebe-e92c-4ef1-8b0f-b28af2a6c9e8 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2545.048819] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-0ddedebe-e92c-4ef1-8b0f-b28af2a6c9e8 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2545.049096] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-eab809bd-65a3-460a-a578-09415b9e2371 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2545.049310] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-271b1159-a62e-49de-99e2-6b0b34af7eb3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2545.050615] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3e556c40-b221-4a1b-a67b-1573e84847b8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2545.057084] env[62824]: DEBUG oslo_vmware.api [None req-eab809bd-65a3-460a-a578-09415b9e2371 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2545.057084] env[62824]: value = "task-2147103" [ 2545.057084] env[62824]: _type = "Task" [ 2545.057084] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2545.064394] env[62824]: DEBUG oslo_vmware.api [None req-eab809bd-65a3-460a-a578-09415b9e2371 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147103, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2545.203503] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-0ddedebe-e92c-4ef1-8b0f-b28af2a6c9e8 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2545.203750] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-0ddedebe-e92c-4ef1-8b0f-b28af2a6c9e8 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Deleting contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2545.203902] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-0ddedebe-e92c-4ef1-8b0f-b28af2a6c9e8 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Deleting the datastore file [datastore1] 80c57382-8f6f-4537-b210-70a902b02cb4 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2545.204237] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-caa4604b-9268-46e1-ab82-4c60a3bbca42 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2545.211769] env[62824]: DEBUG oslo_vmware.api [None req-0ddedebe-e92c-4ef1-8b0f-b28af2a6c9e8 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2545.211769] env[62824]: value = "task-2147104" [ 2545.211769] env[62824]: _type = "Task" [ 2545.211769] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2545.220646] env[62824]: DEBUG oslo_vmware.api [None req-0ddedebe-e92c-4ef1-8b0f-b28af2a6c9e8 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2147104, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2545.567077] env[62824]: DEBUG oslo_vmware.api [None req-eab809bd-65a3-460a-a578-09415b9e2371 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147103, 'name': PowerOffVM_Task, 'duration_secs': 0.196921} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2545.567355] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-eab809bd-65a3-460a-a578-09415b9e2371 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2545.567530] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-eab809bd-65a3-460a-a578-09415b9e2371 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2545.567774] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c7c23c5f-ba4e-468b-9d49-9ba280ec9b3d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2545.721583] env[62824]: DEBUG oslo_vmware.api [None req-0ddedebe-e92c-4ef1-8b0f-b28af2a6c9e8 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2147104, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.148132} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2545.721842] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-0ddedebe-e92c-4ef1-8b0f-b28af2a6c9e8 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2545.722040] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-0ddedebe-e92c-4ef1-8b0f-b28af2a6c9e8 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Deleted contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2545.722231] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-0ddedebe-e92c-4ef1-8b0f-b28af2a6c9e8 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2545.722402] env[62824]: INFO nova.compute.manager [None req-0ddedebe-e92c-4ef1-8b0f-b28af2a6c9e8 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Took 1.21 seconds to destroy the instance on the hypervisor. [ 2545.722643] env[62824]: DEBUG oslo.service.loopingcall [None req-0ddedebe-e92c-4ef1-8b0f-b28af2a6c9e8 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2545.722833] env[62824]: DEBUG nova.compute.manager [-] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2545.722930] env[62824]: DEBUG nova.network.neutron [-] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2545.814081] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-eab809bd-65a3-460a-a578-09415b9e2371 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2545.814355] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-eab809bd-65a3-460a-a578-09415b9e2371 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Deleting contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2545.814560] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-eab809bd-65a3-460a-a578-09415b9e2371 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Deleting the datastore file [datastore2] 7c47bb92-a58e-4537-a68a-289e2a5fc19a {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2545.814883] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e2edd429-3c28-4094-a9e5-ba4a91f679d1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2545.821388] env[62824]: DEBUG oslo_vmware.api [None req-eab809bd-65a3-460a-a578-09415b9e2371 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2545.821388] env[62824]: value = "task-2147106" [ 2545.821388] env[62824]: _type = "Task" [ 2545.821388] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2545.829457] env[62824]: DEBUG oslo_vmware.api [None req-eab809bd-65a3-460a-a578-09415b9e2371 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147106, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2546.186388] env[62824]: DEBUG nova.compute.manager [req-eebb08c2-a68f-4abd-a91b-0fe8436a10a8 req-f4916db1-0a8e-43dc-865e-79c81f1890a0 service nova] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Received event network-vif-deleted-ae875573-c95e-4613-b343-e3e5e3829964 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2546.186634] env[62824]: INFO nova.compute.manager [req-eebb08c2-a68f-4abd-a91b-0fe8436a10a8 req-f4916db1-0a8e-43dc-865e-79c81f1890a0 service nova] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Neutron deleted interface ae875573-c95e-4613-b343-e3e5e3829964; detaching it from the instance and deleting it from the info cache [ 2546.186744] env[62824]: DEBUG nova.network.neutron [req-eebb08c2-a68f-4abd-a91b-0fe8436a10a8 req-f4916db1-0a8e-43dc-865e-79c81f1890a0 service nova] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2546.331618] env[62824]: DEBUG oslo_vmware.api [None req-eab809bd-65a3-460a-a578-09415b9e2371 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147106, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.132642} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2546.334851] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-eab809bd-65a3-460a-a578-09415b9e2371 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2546.334851] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-eab809bd-65a3-460a-a578-09415b9e2371 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Deleted contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2546.334851] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-eab809bd-65a3-460a-a578-09415b9e2371 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2546.334851] env[62824]: INFO nova.compute.manager [None req-eab809bd-65a3-460a-a578-09415b9e2371 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Took 1.30 seconds to destroy the instance on the hypervisor. [ 2546.334851] env[62824]: DEBUG oslo.service.loopingcall [None req-eab809bd-65a3-460a-a578-09415b9e2371 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2546.334851] env[62824]: DEBUG nova.compute.manager [-] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2546.334851] env[62824]: DEBUG nova.network.neutron [-] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2546.659171] env[62824]: DEBUG nova.network.neutron [-] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2546.688888] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3917955b-fbd2-4835-8352-72d5f8c1b1cd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2546.698907] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b0109cf-e1c4-4bb3-bc25-2609447c42b6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2546.724987] env[62824]: DEBUG nova.compute.manager [req-eebb08c2-a68f-4abd-a91b-0fe8436a10a8 req-f4916db1-0a8e-43dc-865e-79c81f1890a0 service nova] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Detach interface failed, port_id=ae875573-c95e-4613-b343-e3e5e3829964, reason: Instance 80c57382-8f6f-4537-b210-70a902b02cb4 could not be found. {{(pid=62824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11482}} [ 2547.163096] env[62824]: INFO nova.compute.manager [-] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Took 1.44 seconds to deallocate network for instance. [ 2547.248171] env[62824]: DEBUG nova.network.neutron [-] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2547.668912] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0ddedebe-e92c-4ef1-8b0f-b28af2a6c9e8 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2547.669210] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0ddedebe-e92c-4ef1-8b0f-b28af2a6c9e8 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2547.669469] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0ddedebe-e92c-4ef1-8b0f-b28af2a6c9e8 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2547.691696] env[62824]: INFO nova.scheduler.client.report [None req-0ddedebe-e92c-4ef1-8b0f-b28af2a6c9e8 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Deleted allocations for instance 80c57382-8f6f-4537-b210-70a902b02cb4 [ 2547.750522] env[62824]: INFO nova.compute.manager [-] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Took 1.42 seconds to deallocate network for instance. [ 2548.199364] env[62824]: DEBUG oslo_concurrency.lockutils [None req-0ddedebe-e92c-4ef1-8b0f-b28af2a6c9e8 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lock "80c57382-8f6f-4537-b210-70a902b02cb4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.198s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2548.216209] env[62824]: DEBUG nova.compute.manager [req-d791def2-d952-4cf6-9d8c-d3afd80bf818 req-78e79767-3009-4db6-87af-aa52f7bffcba service nova] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Received event network-vif-deleted-821db814-549f-4b95-befb-cdb81b9bcb70 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2548.256797] env[62824]: DEBUG oslo_concurrency.lockutils [None req-eab809bd-65a3-460a-a578-09415b9e2371 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2548.257214] env[62824]: DEBUG oslo_concurrency.lockutils [None req-eab809bd-65a3-460a-a578-09415b9e2371 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2548.257264] env[62824]: DEBUG oslo_concurrency.lockutils [None req-eab809bd-65a3-460a-a578-09415b9e2371 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2548.279643] env[62824]: INFO nova.scheduler.client.report [None req-eab809bd-65a3-460a-a578-09415b9e2371 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Deleted allocations for instance 7c47bb92-a58e-4537-a68a-289e2a5fc19a [ 2548.787832] env[62824]: DEBUG oslo_concurrency.lockutils [None req-eab809bd-65a3-460a-a578-09415b9e2371 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lock "7c47bb92-a58e-4537-a68a-289e2a5fc19a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.266s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2548.922313] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquiring lock "6e066235-99c3-4bde-972a-345faccf9fa6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2548.922558] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lock "6e066235-99c3-4bde-972a-345faccf9fa6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2549.425450] env[62824]: DEBUG nova.compute.manager [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2549.454525] env[62824]: DEBUG oslo_concurrency.lockutils [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquiring lock "eb0a5669-7b2a-4d2f-9596-6486e8b78289" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2549.454779] env[62824]: DEBUG oslo_concurrency.lockutils [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lock "eb0a5669-7b2a-4d2f-9596-6486e8b78289" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2549.946999] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2549.947320] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2549.948831] env[62824]: INFO nova.compute.claims [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2549.957567] env[62824]: DEBUG nova.compute.manager [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2550.476729] env[62824]: DEBUG oslo_concurrency.lockutils [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2551.038573] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50a426db-c120-4455-a8dd-323a804888e2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2551.047628] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ac82943-9c83-45f6-b35a-2fab0fae07ac {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2551.080191] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b0842f5-9d64-4a47-ae5f-fb986e067fef {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2551.087912] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8946e794-d7e0-4d90-8b75-8fe2e15700c9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2551.101197] env[62824]: DEBUG nova.compute.provider_tree [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2551.605028] env[62824]: DEBUG nova.scheduler.client.report [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2552.110735] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.163s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2552.111309] env[62824]: DEBUG nova.compute.manager [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2552.115766] env[62824]: DEBUG oslo_concurrency.lockutils [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.639s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2552.117744] env[62824]: INFO nova.compute.claims [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2552.624750] env[62824]: DEBUG nova.compute.utils [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2552.626413] env[62824]: DEBUG nova.compute.manager [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2552.626657] env[62824]: DEBUG nova.network.neutron [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2552.673494] env[62824]: DEBUG nova.policy [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5a5891a75fee4caf92d4bba152254ad5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7fcdbfc29df64a1ba8d982bdcc667b64', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2552.923052] env[62824]: DEBUG nova.network.neutron [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Successfully created port: d8d8df40-7696-495a-9dca-6e76f780e1a7 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2553.130306] env[62824]: DEBUG nova.compute.manager [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2553.215582] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73cc882e-b53d-492b-975b-96be6be27201 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2553.223315] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d47392b-576a-4ac8-b204-f92530931589 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2553.252905] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1c45855-82de-45a8-bfc4-36bae0b84231 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2553.259822] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8da71886-039b-4120-9414-9080a179b546 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2553.273059] env[62824]: DEBUG nova.compute.provider_tree [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2553.775858] env[62824]: DEBUG nova.scheduler.client.report [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2553.952146] env[62824]: DEBUG oslo_concurrency.lockutils [None req-be39a5ce-9908-47e3-9cb2-5e3f101b676a tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquiring lock "e524ce78-6a63-4950-b727-ff72f8df4424" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2553.952442] env[62824]: DEBUG oslo_concurrency.lockutils [None req-be39a5ce-9908-47e3-9cb2-5e3f101b676a tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "e524ce78-6a63-4950-b727-ff72f8df4424" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2553.952666] env[62824]: DEBUG oslo_concurrency.lockutils [None req-be39a5ce-9908-47e3-9cb2-5e3f101b676a tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquiring lock "e524ce78-6a63-4950-b727-ff72f8df4424-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2553.952857] env[62824]: DEBUG oslo_concurrency.lockutils [None req-be39a5ce-9908-47e3-9cb2-5e3f101b676a tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "e524ce78-6a63-4950-b727-ff72f8df4424-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2553.953051] env[62824]: DEBUG oslo_concurrency.lockutils [None req-be39a5ce-9908-47e3-9cb2-5e3f101b676a tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "e524ce78-6a63-4950-b727-ff72f8df4424-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2553.955484] env[62824]: INFO nova.compute.manager [None req-be39a5ce-9908-47e3-9cb2-5e3f101b676a tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Terminating instance [ 2554.141060] env[62824]: DEBUG nova.compute.manager [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2554.173664] env[62824]: DEBUG nova.virt.hardware [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2554.173927] env[62824]: DEBUG nova.virt.hardware [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2554.174101] env[62824]: DEBUG nova.virt.hardware [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2554.174288] env[62824]: DEBUG nova.virt.hardware [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2554.174582] env[62824]: DEBUG nova.virt.hardware [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2554.174771] env[62824]: DEBUG nova.virt.hardware [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2554.174989] env[62824]: DEBUG nova.virt.hardware [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2554.175171] env[62824]: DEBUG nova.virt.hardware [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2554.175405] env[62824]: DEBUG nova.virt.hardware [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2554.175678] env[62824]: DEBUG nova.virt.hardware [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2554.175890] env[62824]: DEBUG nova.virt.hardware [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2554.176767] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd183685-be3b-41b2-ba1c-fea08ffa54cf {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2554.185351] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bebeb50-b4ac-4cf0-8735-0dccbdd82d18 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2554.280431] env[62824]: DEBUG oslo_concurrency.lockutils [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.164s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2554.280894] env[62824]: DEBUG nova.compute.manager [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2554.459290] env[62824]: DEBUG nova.compute.manager [None req-be39a5ce-9908-47e3-9cb2-5e3f101b676a tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2554.459539] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-be39a5ce-9908-47e3-9cb2-5e3f101b676a tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2554.459828] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-32573713-be11-4bf2-8330-e3c0660d3dd8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2554.467486] env[62824]: DEBUG oslo_vmware.api [None req-be39a5ce-9908-47e3-9cb2-5e3f101b676a tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2554.467486] env[62824]: value = "task-2147107" [ 2554.467486] env[62824]: _type = "Task" [ 2554.467486] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2554.474873] env[62824]: DEBUG oslo_vmware.api [None req-be39a5ce-9908-47e3-9cb2-5e3f101b676a tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2147107, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2554.785897] env[62824]: DEBUG nova.compute.utils [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2554.787331] env[62824]: DEBUG nova.compute.manager [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2554.787501] env[62824]: DEBUG nova.network.neutron [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2554.835769] env[62824]: DEBUG nova.policy [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e6ecc84f6587474f89419598f5c44dbb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0eb41836305847fca773afebdafa01db', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2554.978619] env[62824]: DEBUG oslo_vmware.api [None req-be39a5ce-9908-47e3-9cb2-5e3f101b676a tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2147107, 'name': PowerOffVM_Task, 'duration_secs': 0.21408} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2554.978884] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-be39a5ce-9908-47e3-9cb2-5e3f101b676a tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2554.979305] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-be39a5ce-9908-47e3-9cb2-5e3f101b676a tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Volume detach. Driver type: vmdk {{(pid=62824) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 2554.979598] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-be39a5ce-9908-47e3-9cb2-5e3f101b676a tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-438783', 'volume_id': '08d8bf83-f546-44c3-8c92-1efd79d0f46e', 'name': 'volume-08d8bf83-f546-44c3-8c92-1efd79d0f46e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attaching', 'instance': 'e524ce78-6a63-4950-b727-ff72f8df4424', 'attached_at': '2025-01-16T15:02:00.000000', 'detached_at': '', 'volume_id': '08d8bf83-f546-44c3-8c92-1efd79d0f46e', 'serial': '08d8bf83-f546-44c3-8c92-1efd79d0f46e'} {{(pid=62824) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 2554.980530] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a35e642-1963-4b83-86d9-78243c832910 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2554.998845] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e687529b-5820-4455-b545-12b167a3ff35 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2555.007474] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da020c20-fb14-4035-a05d-53dce3001516 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2555.025138] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0e9ac7f-3763-44a3-987f-5384b4a63593 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2555.040031] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-be39a5ce-9908-47e3-9cb2-5e3f101b676a tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] The volume has not been displaced from its original location: [datastore2] volume-08d8bf83-f546-44c3-8c92-1efd79d0f46e/volume-08d8bf83-f546-44c3-8c92-1efd79d0f46e.vmdk. No consolidation needed. {{(pid=62824) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 2555.045338] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-be39a5ce-9908-47e3-9cb2-5e3f101b676a tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Reconfiguring VM instance instance-00000075 to detach disk 2000 {{(pid=62824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 2555.045648] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c964847e-3b7f-4ef9-8c76-0e4591cc4da3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2555.065222] env[62824]: DEBUG oslo_vmware.api [None req-be39a5ce-9908-47e3-9cb2-5e3f101b676a tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2555.065222] env[62824]: value = "task-2147108" [ 2555.065222] env[62824]: _type = "Task" [ 2555.065222] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2555.075375] env[62824]: DEBUG oslo_vmware.api [None req-be39a5ce-9908-47e3-9cb2-5e3f101b676a tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2147108, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2555.088840] env[62824]: DEBUG nova.network.neutron [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Successfully created port: 8be13d2b-3033-431e-92aa-3e05a180498f {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2555.291220] env[62824]: DEBUG nova.compute.manager [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2555.574549] env[62824]: DEBUG oslo_vmware.api [None req-be39a5ce-9908-47e3-9cb2-5e3f101b676a tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2147108, 'name': ReconfigVM_Task, 'duration_secs': 0.155295} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2555.574925] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-be39a5ce-9908-47e3-9cb2-5e3f101b676a tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Reconfigured VM instance instance-00000075 to detach disk 2000 {{(pid=62824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 2555.579542] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7314fb21-589c-4019-9a47-71d158da816e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2555.594183] env[62824]: DEBUG oslo_vmware.api [None req-be39a5ce-9908-47e3-9cb2-5e3f101b676a tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2555.594183] env[62824]: value = "task-2147109" [ 2555.594183] env[62824]: _type = "Task" [ 2555.594183] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2555.601670] env[62824]: DEBUG oslo_vmware.api [None req-be39a5ce-9908-47e3-9cb2-5e3f101b676a tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2147109, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2556.103664] env[62824]: DEBUG oslo_vmware.api [None req-be39a5ce-9908-47e3-9cb2-5e3f101b676a tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2147109, 'name': ReconfigVM_Task, 'duration_secs': 0.142562} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2556.103664] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-be39a5ce-9908-47e3-9cb2-5e3f101b676a tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-438783', 'volume_id': '08d8bf83-f546-44c3-8c92-1efd79d0f46e', 'name': 'volume-08d8bf83-f546-44c3-8c92-1efd79d0f46e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attaching', 'instance': 'e524ce78-6a63-4950-b727-ff72f8df4424', 'attached_at': '2025-01-16T15:02:00.000000', 'detached_at': '', 'volume_id': '08d8bf83-f546-44c3-8c92-1efd79d0f46e', 'serial': '08d8bf83-f546-44c3-8c92-1efd79d0f46e'} {{(pid=62824) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 2556.104176] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-be39a5ce-9908-47e3-9cb2-5e3f101b676a tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2556.104476] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16045be0-9de5-4c65-8905-f1043bcb54d3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2556.110552] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-be39a5ce-9908-47e3-9cb2-5e3f101b676a tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2556.110760] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-60d389c2-b817-4afc-8549-a2b3ddc75539 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2556.301782] env[62824]: DEBUG nova.compute.manager [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2556.331034] env[62824]: DEBUG nova.virt.hardware [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2556.331293] env[62824]: DEBUG nova.virt.hardware [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2556.331451] env[62824]: DEBUG nova.virt.hardware [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2556.331633] env[62824]: DEBUG nova.virt.hardware [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2556.331778] env[62824]: DEBUG nova.virt.hardware [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2556.331924] env[62824]: DEBUG nova.virt.hardware [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2556.332148] env[62824]: DEBUG nova.virt.hardware [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2556.332308] env[62824]: DEBUG nova.virt.hardware [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2556.332475] env[62824]: DEBUG nova.virt.hardware [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2556.332635] env[62824]: DEBUG nova.virt.hardware [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2556.332804] env[62824]: DEBUG nova.virt.hardware [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2556.333686] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30015f07-0f73-4424-9af5-ff5f83e7b9c7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2556.341220] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1534fd94-dc7d-4ff8-8cb2-22309a84fc22 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2557.769228] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2557.772803] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager.update_available_resource {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2558.275676] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2558.275867] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2558.276046] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2558.276211] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62824) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 2558.277097] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2ab446c-b95f-4858-aa47-dffaf70c9bf0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2558.285361] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-612b72b3-69c3-4949-b2c0-6e654e933555 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2558.299071] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78a30064-3a62-47b8-acf6-592275142f2f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2558.305238] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da75e474-05f2-47d1-b2fa-df8b7871becd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2558.334850] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180680MB free_disk=175GB free_vcpus=48 pci_devices=None {{(pid=62824) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 2558.334850] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2558.334850] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2559.362699] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance ca42fef0-1e90-4ab6-9d60-8ef7e4997884 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2559.362942] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 4dff2a76-b0a2-438c-b52f-d6495948d041 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2559.362990] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance e524ce78-6a63-4950-b727-ff72f8df4424 actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2559.363124] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2559.363243] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 6e066235-99c3-4bde-972a-345faccf9fa6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2559.363361] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance eb0a5669-7b2a-4d2f-9596-6486e8b78289 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2559.363569] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Total usable vcpus: 48, total allocated vcpus: 6 {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 2559.363712] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1728MB phys_disk=200GB used_disk=5GB total_vcpus=48 used_vcpus=6 pci_stats=[] {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 2559.435018] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca2b31fe-c368-4e3d-972c-1196c82f28c1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2559.442481] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f435262-e903-40a6-bb16-7be412c4cd17 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2559.472903] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6bfb147-3cad-418e-9518-9afaa8b2961c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2559.480011] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-825d4761-a85f-4f53-9951-283ece7e9a9d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2559.492695] env[62824]: DEBUG nova.compute.provider_tree [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2559.585888] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-be39a5ce-9908-47e3-9cb2-5e3f101b676a tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2559.586177] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-be39a5ce-9908-47e3-9cb2-5e3f101b676a tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Deleting contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2559.586372] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-be39a5ce-9908-47e3-9cb2-5e3f101b676a tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Deleting the datastore file [datastore2] e524ce78-6a63-4950-b727-ff72f8df4424 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2559.586644] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-33012db2-8ad1-4065-9bae-8c82771d994e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2559.593246] env[62824]: DEBUG oslo_vmware.api [None req-be39a5ce-9908-47e3-9cb2-5e3f101b676a tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2559.593246] env[62824]: value = "task-2147111" [ 2559.593246] env[62824]: _type = "Task" [ 2559.593246] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2559.601608] env[62824]: DEBUG oslo_vmware.api [None req-be39a5ce-9908-47e3-9cb2-5e3f101b676a tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2147111, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2559.996360] env[62824]: DEBUG nova.scheduler.client.report [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2560.103278] env[62824]: DEBUG oslo_vmware.api [None req-be39a5ce-9908-47e3-9cb2-5e3f101b676a tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2147111, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.092454} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2560.103911] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-be39a5ce-9908-47e3-9cb2-5e3f101b676a tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2560.103911] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-be39a5ce-9908-47e3-9cb2-5e3f101b676a tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Deleted contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2560.103911] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-be39a5ce-9908-47e3-9cb2-5e3f101b676a tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2560.104131] env[62824]: INFO nova.compute.manager [None req-be39a5ce-9908-47e3-9cb2-5e3f101b676a tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Took 5.64 seconds to destroy the instance on the hypervisor. [ 2560.104390] env[62824]: DEBUG oslo.service.loopingcall [None req-be39a5ce-9908-47e3-9cb2-5e3f101b676a tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2560.104587] env[62824]: DEBUG nova.compute.manager [-] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2560.104678] env[62824]: DEBUG nova.network.neutron [-] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2560.500996] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62824) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 2560.501251] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.167s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2560.566970] env[62824]: DEBUG nova.compute.manager [req-2967b52d-7679-4daa-9684-cf520c53c8dd req-576a32d6-959a-4830-ba36-e62091d4962d service nova] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Received event network-vif-plugged-8be13d2b-3033-431e-92aa-3e05a180498f {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2560.567216] env[62824]: DEBUG oslo_concurrency.lockutils [req-2967b52d-7679-4daa-9684-cf520c53c8dd req-576a32d6-959a-4830-ba36-e62091d4962d service nova] Acquiring lock "eb0a5669-7b2a-4d2f-9596-6486e8b78289-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2560.567427] env[62824]: DEBUG oslo_concurrency.lockutils [req-2967b52d-7679-4daa-9684-cf520c53c8dd req-576a32d6-959a-4830-ba36-e62091d4962d service nova] Lock "eb0a5669-7b2a-4d2f-9596-6486e8b78289-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2560.567594] env[62824]: DEBUG oslo_concurrency.lockutils [req-2967b52d-7679-4daa-9684-cf520c53c8dd req-576a32d6-959a-4830-ba36-e62091d4962d service nova] Lock "eb0a5669-7b2a-4d2f-9596-6486e8b78289-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2560.567761] env[62824]: DEBUG nova.compute.manager [req-2967b52d-7679-4daa-9684-cf520c53c8dd req-576a32d6-959a-4830-ba36-e62091d4962d service nova] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] No waiting events found dispatching network-vif-plugged-8be13d2b-3033-431e-92aa-3e05a180498f {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2560.567956] env[62824]: WARNING nova.compute.manager [req-2967b52d-7679-4daa-9684-cf520c53c8dd req-576a32d6-959a-4830-ba36-e62091d4962d service nova] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Received unexpected event network-vif-plugged-8be13d2b-3033-431e-92aa-3e05a180498f for instance with vm_state building and task_state spawning. [ 2560.777700] env[62824]: DEBUG nova.network.neutron [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Successfully updated port: 8be13d2b-3033-431e-92aa-3e05a180498f {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2560.947696] env[62824]: DEBUG nova.network.neutron [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Successfully updated port: d8d8df40-7696-495a-9dca-6e76f780e1a7 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2561.215287] env[62824]: DEBUG nova.network.neutron [-] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2561.280642] env[62824]: DEBUG oslo_concurrency.lockutils [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquiring lock "refresh_cache-eb0a5669-7b2a-4d2f-9596-6486e8b78289" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2561.280642] env[62824]: DEBUG oslo_concurrency.lockutils [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquired lock "refresh_cache-eb0a5669-7b2a-4d2f-9596-6486e8b78289" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2561.280642] env[62824]: DEBUG nova.network.neutron [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2561.449876] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquiring lock "refresh_cache-6e066235-99c3-4bde-972a-345faccf9fa6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2561.450040] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquired lock "refresh_cache-6e066235-99c3-4bde-972a-345faccf9fa6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2561.450197] env[62824]: DEBUG nova.network.neutron [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2561.719058] env[62824]: INFO nova.compute.manager [-] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Took 1.61 seconds to deallocate network for instance. [ 2561.812587] env[62824]: DEBUG nova.network.neutron [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2561.941022] env[62824]: DEBUG nova.network.neutron [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Updating instance_info_cache with network_info: [{"id": "8be13d2b-3033-431e-92aa-3e05a180498f", "address": "fa:16:3e:9f:4d:c4", "network": {"id": "5233a794-3514-4912-a857-7e76a66b1e8f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-910685188-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0eb41836305847fca773afebdafa01db", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd98a25d-a7a9-4fb5-8fef-e8df4dbbbf11", "external-id": "nsx-vlan-transportzone-707", "segmentation_id": 707, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8be13d2b-30", "ovs_interfaceid": "8be13d2b-3033-431e-92aa-3e05a180498f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2561.979542] env[62824]: DEBUG nova.network.neutron [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2562.102074] env[62824]: DEBUG nova.network.neutron [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Updating instance_info_cache with network_info: [{"id": "d8d8df40-7696-495a-9dca-6e76f780e1a7", "address": "fa:16:3e:4d:47:a1", "network": {"id": "bba9b98c-c7eb-4890-8de5-eafacfdebd7d", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1928394728-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7fcdbfc29df64a1ba8d982bdcc667b64", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b49c5024-2ced-42ca-90cc-6066766d43e6", "external-id": "nsx-vlan-transportzone-239", "segmentation_id": 239, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd8d8df40-76", "ovs_interfaceid": "d8d8df40-7696-495a-9dca-6e76f780e1a7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2562.260153] env[62824]: INFO nova.compute.manager [None req-be39a5ce-9908-47e3-9cb2-5e3f101b676a tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Took 0.54 seconds to detach 1 volumes for instance. [ 2562.262379] env[62824]: DEBUG nova.compute.manager [None req-be39a5ce-9908-47e3-9cb2-5e3f101b676a tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Deleting volume: 08d8bf83-f546-44c3-8c92-1efd79d0f46e {{(pid=62824) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3282}} [ 2562.443849] env[62824]: DEBUG oslo_concurrency.lockutils [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Releasing lock "refresh_cache-eb0a5669-7b2a-4d2f-9596-6486e8b78289" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2562.444211] env[62824]: DEBUG nova.compute.manager [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Instance network_info: |[{"id": "8be13d2b-3033-431e-92aa-3e05a180498f", "address": "fa:16:3e:9f:4d:c4", "network": {"id": "5233a794-3514-4912-a857-7e76a66b1e8f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-910685188-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0eb41836305847fca773afebdafa01db", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd98a25d-a7a9-4fb5-8fef-e8df4dbbbf11", "external-id": "nsx-vlan-transportzone-707", "segmentation_id": 707, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8be13d2b-30", "ovs_interfaceid": "8be13d2b-3033-431e-92aa-3e05a180498f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2562.444713] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9f:4d:c4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dd98a25d-a7a9-4fb5-8fef-e8df4dbbbf11', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8be13d2b-3033-431e-92aa-3e05a180498f', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2562.452271] env[62824]: DEBUG oslo.service.loopingcall [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2562.452504] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2562.452735] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b7aa5df6-a3e6-4831-9d02-23197f8768a6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2562.473478] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2562.473478] env[62824]: value = "task-2147113" [ 2562.473478] env[62824]: _type = "Task" [ 2562.473478] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2562.481263] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2147113, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2562.600661] env[62824]: DEBUG nova.compute.manager [req-2a587fe1-7a3d-42e8-93b1-2163f2e1a3fa req-3bb27909-8e75-41b8-be47-d1d912e0db2c service nova] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Received event network-vif-deleted-62aa7735-5027-4b47-917b-7b769eac4cee {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2562.600940] env[62824]: DEBUG nova.compute.manager [req-2a587fe1-7a3d-42e8-93b1-2163f2e1a3fa req-3bb27909-8e75-41b8-be47-d1d912e0db2c service nova] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Received event network-changed-8be13d2b-3033-431e-92aa-3e05a180498f {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2562.601184] env[62824]: DEBUG nova.compute.manager [req-2a587fe1-7a3d-42e8-93b1-2163f2e1a3fa req-3bb27909-8e75-41b8-be47-d1d912e0db2c service nova] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Refreshing instance network info cache due to event network-changed-8be13d2b-3033-431e-92aa-3e05a180498f. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2562.601454] env[62824]: DEBUG oslo_concurrency.lockutils [req-2a587fe1-7a3d-42e8-93b1-2163f2e1a3fa req-3bb27909-8e75-41b8-be47-d1d912e0db2c service nova] Acquiring lock "refresh_cache-eb0a5669-7b2a-4d2f-9596-6486e8b78289" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2562.601773] env[62824]: DEBUG oslo_concurrency.lockutils [req-2a587fe1-7a3d-42e8-93b1-2163f2e1a3fa req-3bb27909-8e75-41b8-be47-d1d912e0db2c service nova] Acquired lock "refresh_cache-eb0a5669-7b2a-4d2f-9596-6486e8b78289" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2562.601971] env[62824]: DEBUG nova.network.neutron [req-2a587fe1-7a3d-42e8-93b1-2163f2e1a3fa req-3bb27909-8e75-41b8-be47-d1d912e0db2c service nova] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Refreshing network info cache for port 8be13d2b-3033-431e-92aa-3e05a180498f {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2562.604666] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Releasing lock "refresh_cache-6e066235-99c3-4bde-972a-345faccf9fa6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2562.604935] env[62824]: DEBUG nova.compute.manager [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Instance network_info: |[{"id": "d8d8df40-7696-495a-9dca-6e76f780e1a7", "address": "fa:16:3e:4d:47:a1", "network": {"id": "bba9b98c-c7eb-4890-8de5-eafacfdebd7d", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1928394728-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7fcdbfc29df64a1ba8d982bdcc667b64", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b49c5024-2ced-42ca-90cc-6066766d43e6", "external-id": "nsx-vlan-transportzone-239", "segmentation_id": 239, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd8d8df40-76", "ovs_interfaceid": "d8d8df40-7696-495a-9dca-6e76f780e1a7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2562.605325] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4d:47:a1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b49c5024-2ced-42ca-90cc-6066766d43e6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd8d8df40-7696-495a-9dca-6e76f780e1a7', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2562.613190] env[62824]: DEBUG oslo.service.loopingcall [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2562.614104] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2562.614340] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-16aabf59-ae97-49a6-b50b-b4a2456582b4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2562.633909] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2562.633909] env[62824]: value = "task-2147114" [ 2562.633909] env[62824]: _type = "Task" [ 2562.633909] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2562.642049] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2147114, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2562.798982] env[62824]: DEBUG oslo_concurrency.lockutils [None req-be39a5ce-9908-47e3-9cb2-5e3f101b676a tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2562.799372] env[62824]: DEBUG oslo_concurrency.lockutils [None req-be39a5ce-9908-47e3-9cb2-5e3f101b676a tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2562.799515] env[62824]: DEBUG nova.objects.instance [None req-be39a5ce-9908-47e3-9cb2-5e3f101b676a tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lazy-loading 'resources' on Instance uuid e524ce78-6a63-4950-b727-ff72f8df4424 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2562.983468] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2147113, 'name': CreateVM_Task, 'duration_secs': 0.305631} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2562.983646] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2562.984373] env[62824]: DEBUG oslo_concurrency.lockutils [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2562.984581] env[62824]: DEBUG oslo_concurrency.lockutils [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2562.984915] env[62824]: DEBUG oslo_concurrency.lockutils [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2562.985189] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-83624892-ca26-4bc7-8c69-2d7ca382a350 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2562.989717] env[62824]: DEBUG oslo_vmware.api [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2562.989717] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5236176a-bca6-6564-9355-a96c0669de19" [ 2562.989717] env[62824]: _type = "Task" [ 2562.989717] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2562.998520] env[62824]: DEBUG oslo_vmware.api [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5236176a-bca6-6564-9355-a96c0669de19, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2563.146868] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2147114, 'name': CreateVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2563.296389] env[62824]: DEBUG nova.network.neutron [req-2a587fe1-7a3d-42e8-93b1-2163f2e1a3fa req-3bb27909-8e75-41b8-be47-d1d912e0db2c service nova] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Updated VIF entry in instance network info cache for port 8be13d2b-3033-431e-92aa-3e05a180498f. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2563.296742] env[62824]: DEBUG nova.network.neutron [req-2a587fe1-7a3d-42e8-93b1-2163f2e1a3fa req-3bb27909-8e75-41b8-be47-d1d912e0db2c service nova] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Updating instance_info_cache with network_info: [{"id": "8be13d2b-3033-431e-92aa-3e05a180498f", "address": "fa:16:3e:9f:4d:c4", "network": {"id": "5233a794-3514-4912-a857-7e76a66b1e8f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-910685188-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0eb41836305847fca773afebdafa01db", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd98a25d-a7a9-4fb5-8fef-e8df4dbbbf11", "external-id": "nsx-vlan-transportzone-707", "segmentation_id": 707, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8be13d2b-30", "ovs_interfaceid": "8be13d2b-3033-431e-92aa-3e05a180498f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2563.387128] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17619676-9446-419e-96cc-0dde441a33ff {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2563.394997] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95800475-4dd9-4efb-8539-444dd2a7a678 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2563.424830] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08febb9a-e371-47fd-a929-b308006650df {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2563.432113] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-204bea41-5392-4b83-b7f9-4265891018ec {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2563.445848] env[62824]: DEBUG nova.compute.provider_tree [None req-be39a5ce-9908-47e3-9cb2-5e3f101b676a tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2563.498799] env[62824]: DEBUG oslo_vmware.api [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5236176a-bca6-6564-9355-a96c0669de19, 'name': SearchDatastore_Task, 'duration_secs': 0.00958} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2563.499107] env[62824]: DEBUG oslo_concurrency.lockutils [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2563.499339] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2563.499594] env[62824]: DEBUG oslo_concurrency.lockutils [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2563.499760] env[62824]: DEBUG oslo_concurrency.lockutils [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2563.499942] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2563.500200] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f803dddb-5d9a-4800-94c8-5491b876459b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2563.502109] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2563.502256] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Starting heal instance info cache {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10445}} [ 2563.510294] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2563.510508] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2563.511225] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6a551ae0-6a50-4769-add0-ac5b2d99bdcd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2563.516819] env[62824]: DEBUG oslo_vmware.api [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2563.516819] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5295d02e-b80f-3182-f3ab-dd3f362ebcd6" [ 2563.516819] env[62824]: _type = "Task" [ 2563.516819] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2563.524290] env[62824]: DEBUG oslo_vmware.api [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5295d02e-b80f-3182-f3ab-dd3f362ebcd6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2563.646424] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2147114, 'name': CreateVM_Task, 'duration_secs': 0.570739} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2563.646671] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2563.647409] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2563.647629] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2563.647963] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2563.648235] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2759f68a-f35e-447f-8d6e-eece3cdb39e1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2563.652882] env[62824]: DEBUG oslo_vmware.api [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2563.652882] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52ed9b7d-fcd0-5c88-487f-550217973078" [ 2563.652882] env[62824]: _type = "Task" [ 2563.652882] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2563.662263] env[62824]: DEBUG oslo_vmware.api [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52ed9b7d-fcd0-5c88-487f-550217973078, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2563.799601] env[62824]: DEBUG oslo_concurrency.lockutils [req-2a587fe1-7a3d-42e8-93b1-2163f2e1a3fa req-3bb27909-8e75-41b8-be47-d1d912e0db2c service nova] Releasing lock "refresh_cache-eb0a5669-7b2a-4d2f-9596-6486e8b78289" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2563.799910] env[62824]: DEBUG nova.compute.manager [req-2a587fe1-7a3d-42e8-93b1-2163f2e1a3fa req-3bb27909-8e75-41b8-be47-d1d912e0db2c service nova] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Received event network-vif-plugged-d8d8df40-7696-495a-9dca-6e76f780e1a7 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2563.800097] env[62824]: DEBUG oslo_concurrency.lockutils [req-2a587fe1-7a3d-42e8-93b1-2163f2e1a3fa req-3bb27909-8e75-41b8-be47-d1d912e0db2c service nova] Acquiring lock "6e066235-99c3-4bde-972a-345faccf9fa6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2563.800306] env[62824]: DEBUG oslo_concurrency.lockutils [req-2a587fe1-7a3d-42e8-93b1-2163f2e1a3fa req-3bb27909-8e75-41b8-be47-d1d912e0db2c service nova] Lock "6e066235-99c3-4bde-972a-345faccf9fa6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2563.800468] env[62824]: DEBUG oslo_concurrency.lockutils [req-2a587fe1-7a3d-42e8-93b1-2163f2e1a3fa req-3bb27909-8e75-41b8-be47-d1d912e0db2c service nova] Lock "6e066235-99c3-4bde-972a-345faccf9fa6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2563.800633] env[62824]: DEBUG nova.compute.manager [req-2a587fe1-7a3d-42e8-93b1-2163f2e1a3fa req-3bb27909-8e75-41b8-be47-d1d912e0db2c service nova] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] No waiting events found dispatching network-vif-plugged-d8d8df40-7696-495a-9dca-6e76f780e1a7 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2563.800796] env[62824]: WARNING nova.compute.manager [req-2a587fe1-7a3d-42e8-93b1-2163f2e1a3fa req-3bb27909-8e75-41b8-be47-d1d912e0db2c service nova] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Received unexpected event network-vif-plugged-d8d8df40-7696-495a-9dca-6e76f780e1a7 for instance with vm_state building and task_state spawning. [ 2563.800956] env[62824]: DEBUG nova.compute.manager [req-2a587fe1-7a3d-42e8-93b1-2163f2e1a3fa req-3bb27909-8e75-41b8-be47-d1d912e0db2c service nova] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Received event network-changed-d8d8df40-7696-495a-9dca-6e76f780e1a7 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2563.801129] env[62824]: DEBUG nova.compute.manager [req-2a587fe1-7a3d-42e8-93b1-2163f2e1a3fa req-3bb27909-8e75-41b8-be47-d1d912e0db2c service nova] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Refreshing instance network info cache due to event network-changed-d8d8df40-7696-495a-9dca-6e76f780e1a7. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2563.801309] env[62824]: DEBUG oslo_concurrency.lockutils [req-2a587fe1-7a3d-42e8-93b1-2163f2e1a3fa req-3bb27909-8e75-41b8-be47-d1d912e0db2c service nova] Acquiring lock "refresh_cache-6e066235-99c3-4bde-972a-345faccf9fa6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2563.801590] env[62824]: DEBUG oslo_concurrency.lockutils [req-2a587fe1-7a3d-42e8-93b1-2163f2e1a3fa req-3bb27909-8e75-41b8-be47-d1d912e0db2c service nova] Acquired lock "refresh_cache-6e066235-99c3-4bde-972a-345faccf9fa6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2563.801836] env[62824]: DEBUG nova.network.neutron [req-2a587fe1-7a3d-42e8-93b1-2163f2e1a3fa req-3bb27909-8e75-41b8-be47-d1d912e0db2c service nova] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Refreshing network info cache for port d8d8df40-7696-495a-9dca-6e76f780e1a7 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2563.949015] env[62824]: DEBUG nova.scheduler.client.report [None req-be39a5ce-9908-47e3-9cb2-5e3f101b676a tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2564.028117] env[62824]: DEBUG oslo_vmware.api [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5295d02e-b80f-3182-f3ab-dd3f362ebcd6, 'name': SearchDatastore_Task, 'duration_secs': 0.008067} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2564.028883] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-55b4a113-5884-4d47-9cf1-144c4f2108a2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2564.034330] env[62824]: DEBUG oslo_vmware.api [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2564.034330] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]523d3b4e-65d3-6e6d-df93-ad7677459e1b" [ 2564.034330] env[62824]: _type = "Task" [ 2564.034330] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2564.041859] env[62824]: DEBUG oslo_vmware.api [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]523d3b4e-65d3-6e6d-df93-ad7677459e1b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2564.164045] env[62824]: DEBUG oslo_vmware.api [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52ed9b7d-fcd0-5c88-487f-550217973078, 'name': SearchDatastore_Task, 'duration_secs': 0.008806} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2564.164357] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2564.164623] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2564.164847] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2564.454310] env[62824]: DEBUG oslo_concurrency.lockutils [None req-be39a5ce-9908-47e3-9cb2-5e3f101b676a tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.655s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2564.472528] env[62824]: INFO nova.scheduler.client.report [None req-be39a5ce-9908-47e3-9cb2-5e3f101b676a tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Deleted allocations for instance e524ce78-6a63-4950-b727-ff72f8df4424 [ 2564.511940] env[62824]: DEBUG nova.network.neutron [req-2a587fe1-7a3d-42e8-93b1-2163f2e1a3fa req-3bb27909-8e75-41b8-be47-d1d912e0db2c service nova] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Updated VIF entry in instance network info cache for port d8d8df40-7696-495a-9dca-6e76f780e1a7. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2564.512321] env[62824]: DEBUG nova.network.neutron [req-2a587fe1-7a3d-42e8-93b1-2163f2e1a3fa req-3bb27909-8e75-41b8-be47-d1d912e0db2c service nova] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Updating instance_info_cache with network_info: [{"id": "d8d8df40-7696-495a-9dca-6e76f780e1a7", "address": "fa:16:3e:4d:47:a1", "network": {"id": "bba9b98c-c7eb-4890-8de5-eafacfdebd7d", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1928394728-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7fcdbfc29df64a1ba8d982bdcc667b64", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b49c5024-2ced-42ca-90cc-6066766d43e6", "external-id": "nsx-vlan-transportzone-239", "segmentation_id": 239, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd8d8df40-76", "ovs_interfaceid": "d8d8df40-7696-495a-9dca-6e76f780e1a7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2564.534374] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "refresh_cache-4dff2a76-b0a2-438c-b52f-d6495948d041" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2564.534534] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquired lock "refresh_cache-4dff2a76-b0a2-438c-b52f-d6495948d041" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2564.534721] env[62824]: DEBUG nova.network.neutron [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 4dff2a76-b0a2-438c-b52f-d6495948d041] Forcefully refreshing network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 2564.545788] env[62824]: DEBUG oslo_vmware.api [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]523d3b4e-65d3-6e6d-df93-ad7677459e1b, 'name': SearchDatastore_Task, 'duration_secs': 0.00853} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2564.546054] env[62824]: DEBUG oslo_concurrency.lockutils [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2564.546325] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] eb0a5669-7b2a-4d2f-9596-6486e8b78289/eb0a5669-7b2a-4d2f-9596-6486e8b78289.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2564.546653] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2564.546853] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2564.547093] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-283adbce-84c0-43eb-92bb-35b929b09bc8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2564.549011] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5631abc6-2170-4bef-a976-e1a02249ab53 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2564.556581] env[62824]: DEBUG oslo_vmware.api [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2564.556581] env[62824]: value = "task-2147115" [ 2564.556581] env[62824]: _type = "Task" [ 2564.556581] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2564.560622] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2564.560833] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2564.561875] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c65c4849-59e5-4ffb-989e-9915777bc86c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2564.567197] env[62824]: DEBUG oslo_vmware.api [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147115, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2564.570312] env[62824]: DEBUG oslo_vmware.api [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2564.570312] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52e29f73-a505-a634-8d97-f871f70888f3" [ 2564.570312] env[62824]: _type = "Task" [ 2564.570312] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2564.577210] env[62824]: DEBUG oslo_vmware.api [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52e29f73-a505-a634-8d97-f871f70888f3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2564.982964] env[62824]: DEBUG oslo_concurrency.lockutils [None req-be39a5ce-9908-47e3-9cb2-5e3f101b676a tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "e524ce78-6a63-4950-b727-ff72f8df4424" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.030s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2565.018231] env[62824]: DEBUG oslo_concurrency.lockutils [req-2a587fe1-7a3d-42e8-93b1-2163f2e1a3fa req-3bb27909-8e75-41b8-be47-d1d912e0db2c service nova] Releasing lock "refresh_cache-6e066235-99c3-4bde-972a-345faccf9fa6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2565.067041] env[62824]: DEBUG oslo_vmware.api [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147115, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2565.079981] env[62824]: DEBUG oslo_vmware.api [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52e29f73-a505-a634-8d97-f871f70888f3, 'name': SearchDatastore_Task, 'duration_secs': 0.009236} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2565.080779] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3d76729a-6b6b-4bbd-9ed4-4dc9ef96ce7d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2565.085851] env[62824]: DEBUG oslo_vmware.api [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2565.085851] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52345cf3-648f-ef22-b1bb-fdbadecc4605" [ 2565.085851] env[62824]: _type = "Task" [ 2565.085851] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2565.094670] env[62824]: DEBUG oslo_vmware.api [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52345cf3-648f-ef22-b1bb-fdbadecc4605, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2565.568172] env[62824]: DEBUG oslo_vmware.api [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147115, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.513515} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2565.568458] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] eb0a5669-7b2a-4d2f-9596-6486e8b78289/eb0a5669-7b2a-4d2f-9596-6486e8b78289.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2565.568688] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2565.568942] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2833f21c-3a87-4766-a5e8-ae661982101f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2565.577696] env[62824]: DEBUG oslo_vmware.api [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2565.577696] env[62824]: value = "task-2147116" [ 2565.577696] env[62824]: _type = "Task" [ 2565.577696] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2565.589662] env[62824]: DEBUG oslo_vmware.api [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147116, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2565.600865] env[62824]: DEBUG oslo_vmware.api [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52345cf3-648f-ef22-b1bb-fdbadecc4605, 'name': SearchDatastore_Task, 'duration_secs': 0.008253} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2565.601218] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2565.601533] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 6e066235-99c3-4bde-972a-345faccf9fa6/6e066235-99c3-4bde-972a-345faccf9fa6.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2565.601838] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ebd8131a-1efe-4f2a-91d3-1e23da79a62e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2565.608350] env[62824]: DEBUG oslo_vmware.api [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2565.608350] env[62824]: value = "task-2147117" [ 2565.608350] env[62824]: _type = "Task" [ 2565.608350] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2565.615978] env[62824]: DEBUG oslo_vmware.api [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2147117, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2565.753755] env[62824]: DEBUG nova.network.neutron [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 4dff2a76-b0a2-438c-b52f-d6495948d041] Updating instance_info_cache with network_info: [{"id": "2e0a71bd-c461-4fe9-a810-fe72908cf629", "address": "fa:16:3e:29:2e:10", "network": {"id": "b4ed285c-26a7-461b-8a71-d18e07582ab8", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-997150974-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "815ecf9340934733a83d4dcb26612fd4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69f65356-c85e-4b7f-ad28-7c7b5e8cf50c", "external-id": "nsx-vlan-transportzone-281", "segmentation_id": 281, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2e0a71bd-c4", "ovs_interfaceid": "2e0a71bd-c461-4fe9-a810-fe72908cf629", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2566.087347] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1616716c-7cbe-467d-a410-94340128d160 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquiring lock "4dff2a76-b0a2-438c-b52f-d6495948d041" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2566.087758] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1616716c-7cbe-467d-a410-94340128d160 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "4dff2a76-b0a2-438c-b52f-d6495948d041" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2566.087812] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1616716c-7cbe-467d-a410-94340128d160 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquiring lock "4dff2a76-b0a2-438c-b52f-d6495948d041-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2566.087965] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1616716c-7cbe-467d-a410-94340128d160 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "4dff2a76-b0a2-438c-b52f-d6495948d041-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2566.088154] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1616716c-7cbe-467d-a410-94340128d160 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "4dff2a76-b0a2-438c-b52f-d6495948d041-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2566.089777] env[62824]: DEBUG oslo_vmware.api [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147116, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075698} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2566.090227] env[62824]: INFO nova.compute.manager [None req-1616716c-7cbe-467d-a410-94340128d160 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 4dff2a76-b0a2-438c-b52f-d6495948d041] Terminating instance [ 2566.091623] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2566.093423] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0db4a0ad-2c92-4cf3-b802-7b8ab4ada0ac {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2566.116987] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Reconfiguring VM instance instance-00000078 to attach disk [datastore2] eb0a5669-7b2a-4d2f-9596-6486e8b78289/eb0a5669-7b2a-4d2f-9596-6486e8b78289.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2566.117942] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-574b3d30-e3c4-4504-b5ec-085c6789b581 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2566.140159] env[62824]: DEBUG oslo_vmware.api [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2147117, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.481818} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2566.141310] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 6e066235-99c3-4bde-972a-345faccf9fa6/6e066235-99c3-4bde-972a-345faccf9fa6.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2566.141535] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2566.141828] env[62824]: DEBUG oslo_vmware.api [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2566.141828] env[62824]: value = "task-2147118" [ 2566.141828] env[62824]: _type = "Task" [ 2566.141828] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2566.142023] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f34a78e9-f1d4-4f92-b1d2-b899cd5ae18e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2566.152146] env[62824]: DEBUG oslo_vmware.api [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147118, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2566.153307] env[62824]: DEBUG oslo_vmware.api [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2566.153307] env[62824]: value = "task-2147119" [ 2566.153307] env[62824]: _type = "Task" [ 2566.153307] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2566.160805] env[62824]: DEBUG oslo_vmware.api [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2147119, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2566.256684] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Releasing lock "refresh_cache-4dff2a76-b0a2-438c-b52f-d6495948d041" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2566.256896] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 4dff2a76-b0a2-438c-b52f-d6495948d041] Updated the network info_cache for instance {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10516}} [ 2566.257146] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2566.257331] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2566.257490] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2566.257652] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2566.597289] env[62824]: DEBUG nova.compute.manager [None req-1616716c-7cbe-467d-a410-94340128d160 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 4dff2a76-b0a2-438c-b52f-d6495948d041] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2566.597479] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-1616716c-7cbe-467d-a410-94340128d160 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 4dff2a76-b0a2-438c-b52f-d6495948d041] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2566.598430] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f654bca-5865-4898-9619-7d4d4676c251 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2566.607608] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-1616716c-7cbe-467d-a410-94340128d160 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 4dff2a76-b0a2-438c-b52f-d6495948d041] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2566.607833] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f3e27af4-7a22-44fd-a023-4ce4582f8e8b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2566.614151] env[62824]: DEBUG oslo_vmware.api [None req-1616716c-7cbe-467d-a410-94340128d160 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2566.614151] env[62824]: value = "task-2147120" [ 2566.614151] env[62824]: _type = "Task" [ 2566.614151] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2566.621494] env[62824]: DEBUG oslo_vmware.api [None req-1616716c-7cbe-467d-a410-94340128d160 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2147120, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2566.653723] env[62824]: DEBUG oslo_vmware.api [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147118, 'name': ReconfigVM_Task, 'duration_secs': 0.271653} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2566.654157] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Reconfigured VM instance instance-00000078 to attach disk [datastore2] eb0a5669-7b2a-4d2f-9596-6486e8b78289/eb0a5669-7b2a-4d2f-9596-6486e8b78289.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2566.657545] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-98de00e2-c12f-4a84-96c4-88dd4880172a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2566.664155] env[62824]: DEBUG oslo_vmware.api [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2147119, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067741} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2566.665240] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2566.665553] env[62824]: DEBUG oslo_vmware.api [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2566.665553] env[62824]: value = "task-2147121" [ 2566.665553] env[62824]: _type = "Task" [ 2566.665553] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2566.666217] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7ee989f-141f-4178-9888-ee50f1244456 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2566.690481] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Reconfiguring VM instance instance-00000077 to attach disk [datastore2] 6e066235-99c3-4bde-972a-345faccf9fa6/6e066235-99c3-4bde-972a-345faccf9fa6.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2566.693844] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-881aea02-7ad0-4685-8eeb-96f5cfe4ad99 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2566.707743] env[62824]: DEBUG oslo_vmware.api [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147121, 'name': Rename_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2566.713235] env[62824]: DEBUG oslo_vmware.api [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2566.713235] env[62824]: value = "task-2147122" [ 2566.713235] env[62824]: _type = "Task" [ 2566.713235] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2566.721210] env[62824]: DEBUG oslo_vmware.api [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2147122, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2566.773220] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2567.123905] env[62824]: DEBUG oslo_vmware.api [None req-1616716c-7cbe-467d-a410-94340128d160 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2147120, 'name': PowerOffVM_Task, 'duration_secs': 0.199812} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2567.124291] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-1616716c-7cbe-467d-a410-94340128d160 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 4dff2a76-b0a2-438c-b52f-d6495948d041] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2567.124291] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-1616716c-7cbe-467d-a410-94340128d160 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 4dff2a76-b0a2-438c-b52f-d6495948d041] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2567.124518] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4250ba19-f733-4e01-b54e-015217df171b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2567.177619] env[62824]: DEBUG oslo_vmware.api [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147121, 'name': Rename_Task, 'duration_secs': 0.142495} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2567.177885] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2567.178147] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8fe39ac3-c497-402b-9845-213e116d8bd1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2567.184149] env[62824]: DEBUG oslo_vmware.api [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2567.184149] env[62824]: value = "task-2147124" [ 2567.184149] env[62824]: _type = "Task" [ 2567.184149] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2567.191249] env[62824]: DEBUG oslo_vmware.api [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147124, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2567.222288] env[62824]: DEBUG oslo_vmware.api [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2147122, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2567.237703] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-1616716c-7cbe-467d-a410-94340128d160 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 4dff2a76-b0a2-438c-b52f-d6495948d041] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2567.237965] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-1616716c-7cbe-467d-a410-94340128d160 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 4dff2a76-b0a2-438c-b52f-d6495948d041] Deleting contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2567.238233] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-1616716c-7cbe-467d-a410-94340128d160 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Deleting the datastore file [datastore2] 4dff2a76-b0a2-438c-b52f-d6495948d041 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2567.238562] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-82094774-3982-484f-9171-d3c646d0f93c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2567.245534] env[62824]: DEBUG oslo_vmware.api [None req-1616716c-7cbe-467d-a410-94340128d160 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2567.245534] env[62824]: value = "task-2147125" [ 2567.245534] env[62824]: _type = "Task" [ 2567.245534] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2567.252821] env[62824]: DEBUG oslo_vmware.api [None req-1616716c-7cbe-467d-a410-94340128d160 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2147125, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2567.694646] env[62824]: DEBUG oslo_vmware.api [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147124, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2567.723728] env[62824]: DEBUG oslo_vmware.api [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2147122, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2567.754308] env[62824]: DEBUG oslo_vmware.api [None req-1616716c-7cbe-467d-a410-94340128d160 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2147125, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.240097} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2567.754556] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-1616716c-7cbe-467d-a410-94340128d160 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2567.754759] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-1616716c-7cbe-467d-a410-94340128d160 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 4dff2a76-b0a2-438c-b52f-d6495948d041] Deleted contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2567.754916] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-1616716c-7cbe-467d-a410-94340128d160 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 4dff2a76-b0a2-438c-b52f-d6495948d041] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2567.755105] env[62824]: INFO nova.compute.manager [None req-1616716c-7cbe-467d-a410-94340128d160 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: 4dff2a76-b0a2-438c-b52f-d6495948d041] Took 1.16 seconds to destroy the instance on the hypervisor. [ 2567.755344] env[62824]: DEBUG oslo.service.loopingcall [None req-1616716c-7cbe-467d-a410-94340128d160 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2567.755525] env[62824]: DEBUG nova.compute.manager [-] [instance: 4dff2a76-b0a2-438c-b52f-d6495948d041] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2567.755624] env[62824]: DEBUG nova.network.neutron [-] [instance: 4dff2a76-b0a2-438c-b52f-d6495948d041] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2568.012722] env[62824]: DEBUG nova.compute.manager [req-c78e9096-a94f-4de5-a79d-582cdb6cf9b3 req-ce4ec93f-707f-4015-9173-547f6ebd7910 service nova] [instance: 4dff2a76-b0a2-438c-b52f-d6495948d041] Received event network-vif-deleted-2e0a71bd-c461-4fe9-a810-fe72908cf629 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2568.012965] env[62824]: INFO nova.compute.manager [req-c78e9096-a94f-4de5-a79d-582cdb6cf9b3 req-ce4ec93f-707f-4015-9173-547f6ebd7910 service nova] [instance: 4dff2a76-b0a2-438c-b52f-d6495948d041] Neutron deleted interface 2e0a71bd-c461-4fe9-a810-fe72908cf629; detaching it from the instance and deleting it from the info cache [ 2568.014083] env[62824]: DEBUG nova.network.neutron [req-c78e9096-a94f-4de5-a79d-582cdb6cf9b3 req-ce4ec93f-707f-4015-9173-547f6ebd7910 service nova] [instance: 4dff2a76-b0a2-438c-b52f-d6495948d041] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2568.195332] env[62824]: DEBUG oslo_vmware.api [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147124, 'name': PowerOnVM_Task, 'duration_secs': 0.654329} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2568.195710] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2568.195789] env[62824]: INFO nova.compute.manager [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Took 11.89 seconds to spawn the instance on the hypervisor. [ 2568.196019] env[62824]: DEBUG nova.compute.manager [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2568.196786] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e69932ef-5c27-46dd-a306-5d2e34fe6398 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2568.224244] env[62824]: DEBUG oslo_vmware.api [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2147122, 'name': ReconfigVM_Task, 'duration_secs': 1.351} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2568.224530] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Reconfigured VM instance instance-00000077 to attach disk [datastore2] 6e066235-99c3-4bde-972a-345faccf9fa6/6e066235-99c3-4bde-972a-345faccf9fa6.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2568.225166] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-aa2e63de-06ab-4122-9f6b-91997f2a11c0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2568.230560] env[62824]: DEBUG oslo_vmware.api [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2568.230560] env[62824]: value = "task-2147126" [ 2568.230560] env[62824]: _type = "Task" [ 2568.230560] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2568.237968] env[62824]: DEBUG oslo_vmware.api [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2147126, 'name': Rename_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2568.490746] env[62824]: DEBUG nova.network.neutron [-] [instance: 4dff2a76-b0a2-438c-b52f-d6495948d041] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2568.515148] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-190f4891-9a00-4330-bbf3-f7d3c8c2b5ee {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2568.524234] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea21de85-5770-4524-ae96-f50f2115bcb6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2568.550597] env[62824]: DEBUG nova.compute.manager [req-c78e9096-a94f-4de5-a79d-582cdb6cf9b3 req-ce4ec93f-707f-4015-9173-547f6ebd7910 service nova] [instance: 4dff2a76-b0a2-438c-b52f-d6495948d041] Detach interface failed, port_id=2e0a71bd-c461-4fe9-a810-fe72908cf629, reason: Instance 4dff2a76-b0a2-438c-b52f-d6495948d041 could not be found. {{(pid=62824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11482}} [ 2568.713520] env[62824]: INFO nova.compute.manager [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Took 18.25 seconds to build instance. [ 2568.740561] env[62824]: DEBUG oslo_vmware.api [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2147126, 'name': Rename_Task, 'duration_secs': 0.143993} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2568.740848] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2568.741076] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6caa2caa-1192-429e-89a6-968f42deb45c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2568.747463] env[62824]: DEBUG oslo_vmware.api [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2568.747463] env[62824]: value = "task-2147127" [ 2568.747463] env[62824]: _type = "Task" [ 2568.747463] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2568.755193] env[62824]: DEBUG oslo_vmware.api [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2147127, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2568.773177] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2568.773369] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62824) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11064}} [ 2568.993797] env[62824]: INFO nova.compute.manager [-] [instance: 4dff2a76-b0a2-438c-b52f-d6495948d041] Took 1.24 seconds to deallocate network for instance. [ 2569.058070] env[62824]: DEBUG nova.compute.manager [req-223a5060-276a-4f44-9d0b-7e9993ffb6d7 req-93170050-f70d-4d5f-95f3-4bb9397d1752 service nova] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Received event network-changed-8be13d2b-3033-431e-92aa-3e05a180498f {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2569.058273] env[62824]: DEBUG nova.compute.manager [req-223a5060-276a-4f44-9d0b-7e9993ffb6d7 req-93170050-f70d-4d5f-95f3-4bb9397d1752 service nova] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Refreshing instance network info cache due to event network-changed-8be13d2b-3033-431e-92aa-3e05a180498f. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2569.058493] env[62824]: DEBUG oslo_concurrency.lockutils [req-223a5060-276a-4f44-9d0b-7e9993ffb6d7 req-93170050-f70d-4d5f-95f3-4bb9397d1752 service nova] Acquiring lock "refresh_cache-eb0a5669-7b2a-4d2f-9596-6486e8b78289" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2569.058638] env[62824]: DEBUG oslo_concurrency.lockutils [req-223a5060-276a-4f44-9d0b-7e9993ffb6d7 req-93170050-f70d-4d5f-95f3-4bb9397d1752 service nova] Acquired lock "refresh_cache-eb0a5669-7b2a-4d2f-9596-6486e8b78289" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2569.058800] env[62824]: DEBUG nova.network.neutron [req-223a5060-276a-4f44-9d0b-7e9993ffb6d7 req-93170050-f70d-4d5f-95f3-4bb9397d1752 service nova] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Refreshing network info cache for port 8be13d2b-3033-431e-92aa-3e05a180498f {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2569.214762] env[62824]: DEBUG oslo_concurrency.lockutils [None req-de7f3711-0189-4b6d-a193-752213350b0d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lock "eb0a5669-7b2a-4d2f-9596-6486e8b78289" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.760s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2569.259283] env[62824]: DEBUG oslo_vmware.api [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2147127, 'name': PowerOnVM_Task, 'duration_secs': 0.473035} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2569.259283] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2569.259283] env[62824]: INFO nova.compute.manager [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Took 15.12 seconds to spawn the instance on the hypervisor. [ 2569.259543] env[62824]: DEBUG nova.compute.manager [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2569.260601] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee3b7abb-3114-4709-a710-f605a36de361 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2569.500270] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1616716c-7cbe-467d-a410-94340128d160 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2569.500539] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1616716c-7cbe-467d-a410-94340128d160 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2569.500768] env[62824]: DEBUG nova.objects.instance [None req-1616716c-7cbe-467d-a410-94340128d160 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lazy-loading 'resources' on Instance uuid 4dff2a76-b0a2-438c-b52f-d6495948d041 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2569.767383] env[62824]: DEBUG nova.network.neutron [req-223a5060-276a-4f44-9d0b-7e9993ffb6d7 req-93170050-f70d-4d5f-95f3-4bb9397d1752 service nova] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Updated VIF entry in instance network info cache for port 8be13d2b-3033-431e-92aa-3e05a180498f. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2569.767746] env[62824]: DEBUG nova.network.neutron [req-223a5060-276a-4f44-9d0b-7e9993ffb6d7 req-93170050-f70d-4d5f-95f3-4bb9397d1752 service nova] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Updating instance_info_cache with network_info: [{"id": "8be13d2b-3033-431e-92aa-3e05a180498f", "address": "fa:16:3e:9f:4d:c4", "network": {"id": "5233a794-3514-4912-a857-7e76a66b1e8f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-910685188-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.129", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0eb41836305847fca773afebdafa01db", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd98a25d-a7a9-4fb5-8fef-e8df4dbbbf11", "external-id": "nsx-vlan-transportzone-707", "segmentation_id": 707, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8be13d2b-30", "ovs_interfaceid": "8be13d2b-3033-431e-92aa-3e05a180498f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2569.780334] env[62824]: INFO nova.compute.manager [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Took 19.85 seconds to build instance. [ 2570.083984] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d76b9db-ad6d-4617-9cf2-7422d1f17f0d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2570.091820] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f1b81f7-cead-49a6-a19b-d69a48f3862f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2570.121454] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c5521ab-064b-41ee-8300-f8d7bfb722f0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2570.129140] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dedd9af-10cf-46d2-aad6-40838a634597 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2570.142353] env[62824]: DEBUG nova.compute.provider_tree [None req-1616716c-7cbe-467d-a410-94340128d160 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2570.270652] env[62824]: DEBUG oslo_concurrency.lockutils [req-223a5060-276a-4f44-9d0b-7e9993ffb6d7 req-93170050-f70d-4d5f-95f3-4bb9397d1752 service nova] Releasing lock "refresh_cache-eb0a5669-7b2a-4d2f-9596-6486e8b78289" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2570.282263] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9781fbf7-3fd2-442a-82bb-122198c723e0 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lock "6e066235-99c3-4bde-972a-345faccf9fa6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.359s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2570.645618] env[62824]: DEBUG nova.scheduler.client.report [None req-1616716c-7cbe-467d-a410-94340128d160 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2571.082511] env[62824]: DEBUG nova.compute.manager [req-dd772301-1648-45d8-9993-e7bcb1a4d90e req-cfe18996-9779-442c-9cae-985caef06308 service nova] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Received event network-changed-d8d8df40-7696-495a-9dca-6e76f780e1a7 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2571.082916] env[62824]: DEBUG nova.compute.manager [req-dd772301-1648-45d8-9993-e7bcb1a4d90e req-cfe18996-9779-442c-9cae-985caef06308 service nova] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Refreshing instance network info cache due to event network-changed-d8d8df40-7696-495a-9dca-6e76f780e1a7. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2571.083015] env[62824]: DEBUG oslo_concurrency.lockutils [req-dd772301-1648-45d8-9993-e7bcb1a4d90e req-cfe18996-9779-442c-9cae-985caef06308 service nova] Acquiring lock "refresh_cache-6e066235-99c3-4bde-972a-345faccf9fa6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2571.083126] env[62824]: DEBUG oslo_concurrency.lockutils [req-dd772301-1648-45d8-9993-e7bcb1a4d90e req-cfe18996-9779-442c-9cae-985caef06308 service nova] Acquired lock "refresh_cache-6e066235-99c3-4bde-972a-345faccf9fa6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2571.083381] env[62824]: DEBUG nova.network.neutron [req-dd772301-1648-45d8-9993-e7bcb1a4d90e req-cfe18996-9779-442c-9cae-985caef06308 service nova] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Refreshing network info cache for port d8d8df40-7696-495a-9dca-6e76f780e1a7 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2571.150291] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1616716c-7cbe-467d-a410-94340128d160 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.650s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2571.171508] env[62824]: INFO nova.scheduler.client.report [None req-1616716c-7cbe-467d-a410-94340128d160 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Deleted allocations for instance 4dff2a76-b0a2-438c-b52f-d6495948d041 [ 2571.458017] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dc10c81d-bb7f-4bf4-b8ed-b2fe3680eeb9 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Acquiring lock "8bdfe1c4-1c23-4360-9bb0-90dfc9123d79" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2571.458340] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dc10c81d-bb7f-4bf4-b8ed-b2fe3680eeb9 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Lock "8bdfe1c4-1c23-4360-9bb0-90dfc9123d79" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2571.678403] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1616716c-7cbe-467d-a410-94340128d160 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "4dff2a76-b0a2-438c-b52f-d6495948d041" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.591s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2571.800125] env[62824]: DEBUG nova.network.neutron [req-dd772301-1648-45d8-9993-e7bcb1a4d90e req-cfe18996-9779-442c-9cae-985caef06308 service nova] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Updated VIF entry in instance network info cache for port d8d8df40-7696-495a-9dca-6e76f780e1a7. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2571.800533] env[62824]: DEBUG nova.network.neutron [req-dd772301-1648-45d8-9993-e7bcb1a4d90e req-cfe18996-9779-442c-9cae-985caef06308 service nova] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Updating instance_info_cache with network_info: [{"id": "d8d8df40-7696-495a-9dca-6e76f780e1a7", "address": "fa:16:3e:4d:47:a1", "network": {"id": "bba9b98c-c7eb-4890-8de5-eafacfdebd7d", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1928394728-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.160", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7fcdbfc29df64a1ba8d982bdcc667b64", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b49c5024-2ced-42ca-90cc-6066766d43e6", "external-id": "nsx-vlan-transportzone-239", "segmentation_id": 239, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd8d8df40-76", "ovs_interfaceid": "d8d8df40-7696-495a-9dca-6e76f780e1a7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2571.961165] env[62824]: DEBUG nova.compute.utils [None req-dc10c81d-bb7f-4bf4-b8ed-b2fe3680eeb9 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2572.303462] env[62824]: DEBUG oslo_concurrency.lockutils [req-dd772301-1648-45d8-9993-e7bcb1a4d90e req-cfe18996-9779-442c-9cae-985caef06308 service nova] Releasing lock "refresh_cache-6e066235-99c3-4bde-972a-345faccf9fa6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2572.464741] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dc10c81d-bb7f-4bf4-b8ed-b2fe3680eeb9 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Lock "8bdfe1c4-1c23-4360-9bb0-90dfc9123d79" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2573.521573] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dc10c81d-bb7f-4bf4-b8ed-b2fe3680eeb9 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Acquiring lock "8bdfe1c4-1c23-4360-9bb0-90dfc9123d79" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2573.521964] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dc10c81d-bb7f-4bf4-b8ed-b2fe3680eeb9 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Lock "8bdfe1c4-1c23-4360-9bb0-90dfc9123d79" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2573.522262] env[62824]: INFO nova.compute.manager [None req-dc10c81d-bb7f-4bf4-b8ed-b2fe3680eeb9 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] Attaching volume e0e85130-8172-4714-adf0-34a99efe3451 to /dev/sdb [ 2573.552603] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84566292-28b8-4a27-8a31-a92c36391da2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2573.559913] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9cd8e68-3771-47fe-a2a4-74a23a15f1a0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2573.574080] env[62824]: DEBUG nova.virt.block_device [None req-dc10c81d-bb7f-4bf4-b8ed-b2fe3680eeb9 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] Updating existing volume attachment record: 2f732f92-96d4-4738-a013-b32899a40146 {{(pid=62824) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 2573.768741] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2575.782183] env[62824]: DEBUG oslo_concurrency.lockutils [None req-62c7d0e2-d3fb-4592-a7a0-980d85605652 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquiring lock "ca42fef0-1e90-4ab6-9d60-8ef7e4997884" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2575.782183] env[62824]: DEBUG oslo_concurrency.lockutils [None req-62c7d0e2-d3fb-4592-a7a0-980d85605652 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "ca42fef0-1e90-4ab6-9d60-8ef7e4997884" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2575.782509] env[62824]: DEBUG oslo_concurrency.lockutils [None req-62c7d0e2-d3fb-4592-a7a0-980d85605652 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquiring lock "ca42fef0-1e90-4ab6-9d60-8ef7e4997884-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2575.782605] env[62824]: DEBUG oslo_concurrency.lockutils [None req-62c7d0e2-d3fb-4592-a7a0-980d85605652 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "ca42fef0-1e90-4ab6-9d60-8ef7e4997884-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2575.782808] env[62824]: DEBUG oslo_concurrency.lockutils [None req-62c7d0e2-d3fb-4592-a7a0-980d85605652 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "ca42fef0-1e90-4ab6-9d60-8ef7e4997884-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2575.785144] env[62824]: INFO nova.compute.manager [None req-62c7d0e2-d3fb-4592-a7a0-980d85605652 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] Terminating instance [ 2576.290122] env[62824]: DEBUG nova.compute.manager [None req-62c7d0e2-d3fb-4592-a7a0-980d85605652 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2576.290122] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-62c7d0e2-d3fb-4592-a7a0-980d85605652 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2576.290362] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39572aac-4861-4e85-82d9-d5ccc94f4eaf {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2576.298217] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-62c7d0e2-d3fb-4592-a7a0-980d85605652 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2576.298471] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-769297ee-ffce-45e1-8a03-5a6b34d4b6f9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2576.304150] env[62824]: DEBUG oslo_vmware.api [None req-62c7d0e2-d3fb-4592-a7a0-980d85605652 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2576.304150] env[62824]: value = "task-2147131" [ 2576.304150] env[62824]: _type = "Task" [ 2576.304150] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2576.312012] env[62824]: DEBUG oslo_vmware.api [None req-62c7d0e2-d3fb-4592-a7a0-980d85605652 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2147131, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2576.814255] env[62824]: DEBUG oslo_vmware.api [None req-62c7d0e2-d3fb-4592-a7a0-980d85605652 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2147131, 'name': PowerOffVM_Task, 'duration_secs': 0.243711} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2576.814744] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-62c7d0e2-d3fb-4592-a7a0-980d85605652 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2576.814744] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-62c7d0e2-d3fb-4592-a7a0-980d85605652 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2576.814987] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-21c63133-1c16-49eb-8fb4-fee3560a19df {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2576.982412] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-62c7d0e2-d3fb-4592-a7a0-980d85605652 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2576.982732] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-62c7d0e2-d3fb-4592-a7a0-980d85605652 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] Deleting contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2576.982868] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-62c7d0e2-d3fb-4592-a7a0-980d85605652 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Deleting the datastore file [datastore1] ca42fef0-1e90-4ab6-9d60-8ef7e4997884 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2576.983148] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-51c91b89-8b67-4547-9723-21ede04bb8c8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2576.989602] env[62824]: DEBUG oslo_vmware.api [None req-62c7d0e2-d3fb-4592-a7a0-980d85605652 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for the task: (returnval){ [ 2576.989602] env[62824]: value = "task-2147133" [ 2576.989602] env[62824]: _type = "Task" [ 2576.989602] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2576.998595] env[62824]: DEBUG oslo_vmware.api [None req-62c7d0e2-d3fb-4592-a7a0-980d85605652 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2147133, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2577.499295] env[62824]: DEBUG oslo_vmware.api [None req-62c7d0e2-d3fb-4592-a7a0-980d85605652 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Task: {'id': task-2147133, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.421983} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2577.499568] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-62c7d0e2-d3fb-4592-a7a0-980d85605652 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2577.499728] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-62c7d0e2-d3fb-4592-a7a0-980d85605652 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] Deleted contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2577.499924] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-62c7d0e2-d3fb-4592-a7a0-980d85605652 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2577.500113] env[62824]: INFO nova.compute.manager [None req-62c7d0e2-d3fb-4592-a7a0-980d85605652 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] Took 1.21 seconds to destroy the instance on the hypervisor. [ 2577.500361] env[62824]: DEBUG oslo.service.loopingcall [None req-62c7d0e2-d3fb-4592-a7a0-980d85605652 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2577.500555] env[62824]: DEBUG nova.compute.manager [-] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2577.500667] env[62824]: DEBUG nova.network.neutron [-] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2577.780256] env[62824]: DEBUG nova.compute.manager [req-bfd57b55-729a-4756-aee9-abc57647f450 req-f1ff1379-e41c-41d1-bd63-723d46564832 service nova] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] Received event network-vif-deleted-3eb965bf-7ceb-4a51-bc6f-00803bc6b99e {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2577.780467] env[62824]: INFO nova.compute.manager [req-bfd57b55-729a-4756-aee9-abc57647f450 req-f1ff1379-e41c-41d1-bd63-723d46564832 service nova] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] Neutron deleted interface 3eb965bf-7ceb-4a51-bc6f-00803bc6b99e; detaching it from the instance and deleting it from the info cache [ 2577.780642] env[62824]: DEBUG nova.network.neutron [req-bfd57b55-729a-4756-aee9-abc57647f450 req-f1ff1379-e41c-41d1-bd63-723d46564832 service nova] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2578.118202] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc10c81d-bb7f-4bf4-b8ed-b2fe3680eeb9 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] Volume attach. Driver type: vmdk {{(pid=62824) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 2578.118638] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc10c81d-bb7f-4bf4-b8ed-b2fe3680eeb9 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-438792', 'volume_id': 'e0e85130-8172-4714-adf0-34a99efe3451', 'name': 'volume-e0e85130-8172-4714-adf0-34a99efe3451', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '8bdfe1c4-1c23-4360-9bb0-90dfc9123d79', 'attached_at': '', 'detached_at': '', 'volume_id': 'e0e85130-8172-4714-adf0-34a99efe3451', 'serial': 'e0e85130-8172-4714-adf0-34a99efe3451'} {{(pid=62824) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 2578.120110] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9447d42-a7f8-40ec-bf41-b0f62dc82de9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2578.146898] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe56470c-71f5-4b5b-8644-2bbe8e0b6eff {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2578.190388] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc10c81d-bb7f-4bf4-b8ed-b2fe3680eeb9 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] Reconfiguring VM instance instance-00000076 to attach disk [datastore2] volume-e0e85130-8172-4714-adf0-34a99efe3451/volume-e0e85130-8172-4714-adf0-34a99efe3451.vmdk or device None with type thin {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2578.190855] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-01d0c588-c639-4589-97c2-369e4233cb8f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2578.221091] env[62824]: DEBUG oslo_vmware.api [None req-dc10c81d-bb7f-4bf4-b8ed-b2fe3680eeb9 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Waiting for the task: (returnval){ [ 2578.221091] env[62824]: value = "task-2147134" [ 2578.221091] env[62824]: _type = "Task" [ 2578.221091] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2578.234096] env[62824]: DEBUG oslo_vmware.api [None req-dc10c81d-bb7f-4bf4-b8ed-b2fe3680eeb9 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147134, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2578.258717] env[62824]: DEBUG nova.network.neutron [-] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2578.283563] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4d074559-688c-49ef-8485-d9c425f14fb8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2578.294303] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-851d7117-6fff-40ba-89a5-d5180e9a4457 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2578.321086] env[62824]: DEBUG nova.compute.manager [req-bfd57b55-729a-4756-aee9-abc57647f450 req-f1ff1379-e41c-41d1-bd63-723d46564832 service nova] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] Detach interface failed, port_id=3eb965bf-7ceb-4a51-bc6f-00803bc6b99e, reason: Instance ca42fef0-1e90-4ab6-9d60-8ef7e4997884 could not be found. {{(pid=62824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11482}} [ 2578.731316] env[62824]: DEBUG oslo_vmware.api [None req-dc10c81d-bb7f-4bf4-b8ed-b2fe3680eeb9 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147134, 'name': ReconfigVM_Task, 'duration_secs': 0.408881} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2578.731484] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc10c81d-bb7f-4bf4-b8ed-b2fe3680eeb9 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] Reconfigured VM instance instance-00000076 to attach disk [datastore2] volume-e0e85130-8172-4714-adf0-34a99efe3451/volume-e0e85130-8172-4714-adf0-34a99efe3451.vmdk or device None with type thin {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2578.736347] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c9eef913-9216-4c7e-a8f4-c4a396ce019c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2578.752583] env[62824]: DEBUG oslo_vmware.api [None req-dc10c81d-bb7f-4bf4-b8ed-b2fe3680eeb9 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Waiting for the task: (returnval){ [ 2578.752583] env[62824]: value = "task-2147135" [ 2578.752583] env[62824]: _type = "Task" [ 2578.752583] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2578.760803] env[62824]: DEBUG oslo_vmware.api [None req-dc10c81d-bb7f-4bf4-b8ed-b2fe3680eeb9 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147135, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2578.761317] env[62824]: INFO nova.compute.manager [-] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] Took 1.26 seconds to deallocate network for instance. [ 2579.263075] env[62824]: DEBUG oslo_vmware.api [None req-dc10c81d-bb7f-4bf4-b8ed-b2fe3680eeb9 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147135, 'name': ReconfigVM_Task, 'duration_secs': 0.139522} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2579.263075] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc10c81d-bb7f-4bf4-b8ed-b2fe3680eeb9 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-438792', 'volume_id': 'e0e85130-8172-4714-adf0-34a99efe3451', 'name': 'volume-e0e85130-8172-4714-adf0-34a99efe3451', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '8bdfe1c4-1c23-4360-9bb0-90dfc9123d79', 'attached_at': '', 'detached_at': '', 'volume_id': 'e0e85130-8172-4714-adf0-34a99efe3451', 'serial': 'e0e85130-8172-4714-adf0-34a99efe3451'} {{(pid=62824) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 2579.267539] env[62824]: DEBUG oslo_concurrency.lockutils [None req-62c7d0e2-d3fb-4592-a7a0-980d85605652 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2579.267776] env[62824]: DEBUG oslo_concurrency.lockutils [None req-62c7d0e2-d3fb-4592-a7a0-980d85605652 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2579.267990] env[62824]: DEBUG nova.objects.instance [None req-62c7d0e2-d3fb-4592-a7a0-980d85605652 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lazy-loading 'resources' on Instance uuid ca42fef0-1e90-4ab6-9d60-8ef7e4997884 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2579.833304] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c972e33-521c-42ab-8fa8-2cb0752b1cb1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2579.840483] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c4267a3-b177-4972-bf34-fdf434eb4aef {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2579.875567] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d57bf78-336d-42ed-b400-ce75b902cbb9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2579.882409] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b498e753-d9ab-4b7d-bebb-75d765b82dc6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2579.895362] env[62824]: DEBUG nova.compute.provider_tree [None req-62c7d0e2-d3fb-4592-a7a0-980d85605652 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2580.300907] env[62824]: DEBUG nova.objects.instance [None req-dc10c81d-bb7f-4bf4-b8ed-b2fe3680eeb9 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Lazy-loading 'flavor' on Instance uuid 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2580.398883] env[62824]: DEBUG nova.scheduler.client.report [None req-62c7d0e2-d3fb-4592-a7a0-980d85605652 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2580.806887] env[62824]: DEBUG oslo_concurrency.lockutils [None req-dc10c81d-bb7f-4bf4-b8ed-b2fe3680eeb9 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Lock "8bdfe1c4-1c23-4360-9bb0-90dfc9123d79" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.285s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2580.903252] env[62824]: DEBUG oslo_concurrency.lockutils [None req-62c7d0e2-d3fb-4592-a7a0-980d85605652 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.635s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2580.925015] env[62824]: INFO nova.scheduler.client.report [None req-62c7d0e2-d3fb-4592-a7a0-980d85605652 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Deleted allocations for instance ca42fef0-1e90-4ab6-9d60-8ef7e4997884 [ 2580.980367] env[62824]: DEBUG oslo_concurrency.lockutils [None req-faf0117d-aa3c-4dbd-b97f-8e4c333dff30 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Acquiring lock "8bdfe1c4-1c23-4360-9bb0-90dfc9123d79" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2580.980610] env[62824]: DEBUG oslo_concurrency.lockutils [None req-faf0117d-aa3c-4dbd-b97f-8e4c333dff30 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Lock "8bdfe1c4-1c23-4360-9bb0-90dfc9123d79" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2581.432054] env[62824]: DEBUG oslo_concurrency.lockutils [None req-62c7d0e2-d3fb-4592-a7a0-980d85605652 tempest-ServerActionsTestOtherA-292288889 tempest-ServerActionsTestOtherA-292288889-project-member] Lock "ca42fef0-1e90-4ab6-9d60-8ef7e4997884" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.650s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2581.482948] env[62824]: INFO nova.compute.manager [None req-faf0117d-aa3c-4dbd-b97f-8e4c333dff30 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] Detaching volume e0e85130-8172-4714-adf0-34a99efe3451 [ 2581.518675] env[62824]: INFO nova.virt.block_device [None req-faf0117d-aa3c-4dbd-b97f-8e4c333dff30 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] Attempting to driver detach volume e0e85130-8172-4714-adf0-34a99efe3451 from mountpoint /dev/sdb [ 2581.518931] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-faf0117d-aa3c-4dbd-b97f-8e4c333dff30 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] Volume detach. Driver type: vmdk {{(pid=62824) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 2581.519169] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-faf0117d-aa3c-4dbd-b97f-8e4c333dff30 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-438792', 'volume_id': 'e0e85130-8172-4714-adf0-34a99efe3451', 'name': 'volume-e0e85130-8172-4714-adf0-34a99efe3451', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '8bdfe1c4-1c23-4360-9bb0-90dfc9123d79', 'attached_at': '', 'detached_at': '', 'volume_id': 'e0e85130-8172-4714-adf0-34a99efe3451', 'serial': 'e0e85130-8172-4714-adf0-34a99efe3451'} {{(pid=62824) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 2581.520022] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfa396ea-6181-4658-a4b1-b6f8badb0228 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2581.541741] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b46cc29b-3bc3-44ba-b2e8-96c58b6b5d1e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2581.548697] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08c6bd93-61b5-494d-9994-97f8d85bb668 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2581.567900] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e6fa13e-249e-4a82-a8ee-75dd42001c49 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2581.582222] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-faf0117d-aa3c-4dbd-b97f-8e4c333dff30 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] The volume has not been displaced from its original location: [datastore2] volume-e0e85130-8172-4714-adf0-34a99efe3451/volume-e0e85130-8172-4714-adf0-34a99efe3451.vmdk. No consolidation needed. {{(pid=62824) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 2581.587358] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-faf0117d-aa3c-4dbd-b97f-8e4c333dff30 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] Reconfiguring VM instance instance-00000076 to detach disk 2001 {{(pid=62824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 2581.587616] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4a905dd8-68f7-4787-9545-bb99b1ef765b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2581.604313] env[62824]: DEBUG oslo_vmware.api [None req-faf0117d-aa3c-4dbd-b97f-8e4c333dff30 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Waiting for the task: (returnval){ [ 2581.604313] env[62824]: value = "task-2147136" [ 2581.604313] env[62824]: _type = "Task" [ 2581.604313] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2581.611549] env[62824]: DEBUG oslo_vmware.api [None req-faf0117d-aa3c-4dbd-b97f-8e4c333dff30 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147136, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2582.115051] env[62824]: DEBUG oslo_vmware.api [None req-faf0117d-aa3c-4dbd-b97f-8e4c333dff30 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147136, 'name': ReconfigVM_Task, 'duration_secs': 0.20465} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2582.115391] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-faf0117d-aa3c-4dbd-b97f-8e4c333dff30 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] Reconfigured VM instance instance-00000076 to detach disk 2001 {{(pid=62824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 2582.120641] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ef17f259-e3dd-4328-84d1-57e0ccdcb967 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2582.136791] env[62824]: DEBUG oslo_vmware.api [None req-faf0117d-aa3c-4dbd-b97f-8e4c333dff30 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Waiting for the task: (returnval){ [ 2582.136791] env[62824]: value = "task-2147137" [ 2582.136791] env[62824]: _type = "Task" [ 2582.136791] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2582.145400] env[62824]: DEBUG oslo_vmware.api [None req-faf0117d-aa3c-4dbd-b97f-8e4c333dff30 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147137, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2582.646615] env[62824]: DEBUG oslo_vmware.api [None req-faf0117d-aa3c-4dbd-b97f-8e4c333dff30 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147137, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2583.148819] env[62824]: DEBUG oslo_vmware.api [None req-faf0117d-aa3c-4dbd-b97f-8e4c333dff30 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147137, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2583.649197] env[62824]: DEBUG oslo_vmware.api [None req-faf0117d-aa3c-4dbd-b97f-8e4c333dff30 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147137, 'name': ReconfigVM_Task, 'duration_secs': 1.150506} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2583.649504] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-faf0117d-aa3c-4dbd-b97f-8e4c333dff30 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-438792', 'volume_id': 'e0e85130-8172-4714-adf0-34a99efe3451', 'name': 'volume-e0e85130-8172-4714-adf0-34a99efe3451', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '8bdfe1c4-1c23-4360-9bb0-90dfc9123d79', 'attached_at': '', 'detached_at': '', 'volume_id': 'e0e85130-8172-4714-adf0-34a99efe3451', 'serial': 'e0e85130-8172-4714-adf0-34a99efe3451'} {{(pid=62824) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 2584.194031] env[62824]: DEBUG nova.objects.instance [None req-faf0117d-aa3c-4dbd-b97f-8e4c333dff30 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Lazy-loading 'flavor' on Instance uuid 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2585.201506] env[62824]: DEBUG oslo_concurrency.lockutils [None req-faf0117d-aa3c-4dbd-b97f-8e4c333dff30 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Lock "8bdfe1c4-1c23-4360-9bb0-90dfc9123d79" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 4.221s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2586.290253] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fbd418c9-a6fa-4c4f-a0db-b0bd4f22da21 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Acquiring lock "8bdfe1c4-1c23-4360-9bb0-90dfc9123d79" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2586.290565] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fbd418c9-a6fa-4c4f-a0db-b0bd4f22da21 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Lock "8bdfe1c4-1c23-4360-9bb0-90dfc9123d79" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2586.290712] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fbd418c9-a6fa-4c4f-a0db-b0bd4f22da21 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Acquiring lock "8bdfe1c4-1c23-4360-9bb0-90dfc9123d79-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2586.290893] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fbd418c9-a6fa-4c4f-a0db-b0bd4f22da21 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Lock "8bdfe1c4-1c23-4360-9bb0-90dfc9123d79-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2586.291081] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fbd418c9-a6fa-4c4f-a0db-b0bd4f22da21 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Lock "8bdfe1c4-1c23-4360-9bb0-90dfc9123d79-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2586.293172] env[62824]: INFO nova.compute.manager [None req-fbd418c9-a6fa-4c4f-a0db-b0bd4f22da21 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] Terminating instance [ 2586.797433] env[62824]: DEBUG nova.compute.manager [None req-fbd418c9-a6fa-4c4f-a0db-b0bd4f22da21 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2586.797718] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-fbd418c9-a6fa-4c4f-a0db-b0bd4f22da21 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2586.798658] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39983eb4-ec95-487d-baef-42f8ef2123e0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2586.806648] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-fbd418c9-a6fa-4c4f-a0db-b0bd4f22da21 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2586.806913] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cbdd80c5-c470-48ad-8076-9a4d496985fb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2586.812630] env[62824]: DEBUG oslo_vmware.api [None req-fbd418c9-a6fa-4c4f-a0db-b0bd4f22da21 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Waiting for the task: (returnval){ [ 2586.812630] env[62824]: value = "task-2147138" [ 2586.812630] env[62824]: _type = "Task" [ 2586.812630] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2586.820672] env[62824]: DEBUG oslo_vmware.api [None req-fbd418c9-a6fa-4c4f-a0db-b0bd4f22da21 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147138, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2587.322790] env[62824]: DEBUG oslo_vmware.api [None req-fbd418c9-a6fa-4c4f-a0db-b0bd4f22da21 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147138, 'name': PowerOffVM_Task, 'duration_secs': 0.171482} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2587.323199] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-fbd418c9-a6fa-4c4f-a0db-b0bd4f22da21 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2587.323245] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-fbd418c9-a6fa-4c4f-a0db-b0bd4f22da21 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2587.323469] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3e9c0e4b-f1ae-4cae-a728-de45ca8aff8d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2593.634471] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-fbd418c9-a6fa-4c4f-a0db-b0bd4f22da21 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2593.634899] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-fbd418c9-a6fa-4c4f-a0db-b0bd4f22da21 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] Deleting contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2593.634899] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-fbd418c9-a6fa-4c4f-a0db-b0bd4f22da21 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Deleting the datastore file [datastore1] 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2593.635168] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-04512978-0a2d-4043-9e5f-d908b1892c80 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2593.641467] env[62824]: DEBUG oslo_vmware.api [None req-fbd418c9-a6fa-4c4f-a0db-b0bd4f22da21 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Waiting for the task: (returnval){ [ 2593.641467] env[62824]: value = "task-2147140" [ 2593.641467] env[62824]: _type = "Task" [ 2593.641467] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2593.649351] env[62824]: DEBUG oslo_vmware.api [None req-fbd418c9-a6fa-4c4f-a0db-b0bd4f22da21 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147140, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2594.155578] env[62824]: DEBUG oslo_vmware.api [None req-fbd418c9-a6fa-4c4f-a0db-b0bd4f22da21 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147140, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.161853} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2594.155831] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-fbd418c9-a6fa-4c4f-a0db-b0bd4f22da21 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2594.155975] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-fbd418c9-a6fa-4c4f-a0db-b0bd4f22da21 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] Deleted contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2594.156176] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-fbd418c9-a6fa-4c4f-a0db-b0bd4f22da21 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2594.156384] env[62824]: INFO nova.compute.manager [None req-fbd418c9-a6fa-4c4f-a0db-b0bd4f22da21 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] Took 7.36 seconds to destroy the instance on the hypervisor. [ 2594.156621] env[62824]: DEBUG oslo.service.loopingcall [None req-fbd418c9-a6fa-4c4f-a0db-b0bd4f22da21 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2594.156803] env[62824]: DEBUG nova.compute.manager [-] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2594.156901] env[62824]: DEBUG nova.network.neutron [-] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2594.586248] env[62824]: DEBUG nova.compute.manager [req-6a7a9f96-546c-47b6-8c65-f81755c2892b req-a80094b7-9008-4e5e-9adf-ec1f58bfce52 service nova] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] Received event network-vif-deleted-1013ac0c-ff48-42e0-8fe5-07f23015ac08 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2594.586248] env[62824]: INFO nova.compute.manager [req-6a7a9f96-546c-47b6-8c65-f81755c2892b req-a80094b7-9008-4e5e-9adf-ec1f58bfce52 service nova] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] Neutron deleted interface 1013ac0c-ff48-42e0-8fe5-07f23015ac08; detaching it from the instance and deleting it from the info cache [ 2594.586248] env[62824]: DEBUG nova.network.neutron [req-6a7a9f96-546c-47b6-8c65-f81755c2892b req-a80094b7-9008-4e5e-9adf-ec1f58bfce52 service nova] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2595.067290] env[62824]: DEBUG nova.network.neutron [-] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2595.088123] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7dfe6e52-bd56-4277-9c44-e05ec27b1832 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2595.098619] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf0f2376-4ecc-4a46-a343-2bd3e5fe8e5d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2595.125119] env[62824]: DEBUG nova.compute.manager [req-6a7a9f96-546c-47b6-8c65-f81755c2892b req-a80094b7-9008-4e5e-9adf-ec1f58bfce52 service nova] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] Detach interface failed, port_id=1013ac0c-ff48-42e0-8fe5-07f23015ac08, reason: Instance 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79 could not be found. {{(pid=62824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11482}} [ 2595.569769] env[62824]: INFO nova.compute.manager [-] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] Took 1.41 seconds to deallocate network for instance. [ 2596.076219] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fbd418c9-a6fa-4c4f-a0db-b0bd4f22da21 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2596.076580] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fbd418c9-a6fa-4c4f-a0db-b0bd4f22da21 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2596.076731] env[62824]: DEBUG nova.objects.instance [None req-fbd418c9-a6fa-4c4f-a0db-b0bd4f22da21 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Lazy-loading 'resources' on Instance uuid 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2596.633424] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da1b9a25-0a07-47e6-a3dd-af1ffabf8d80 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2596.640931] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a47ec45-adba-4081-beb0-d3ce2da8eef6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2596.670886] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9daa30af-3ef5-4789-9e5e-470157f124f5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2596.677861] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e298b097-ef41-4539-85c6-66a5eed5f887 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2596.691722] env[62824]: DEBUG nova.compute.provider_tree [None req-fbd418c9-a6fa-4c4f-a0db-b0bd4f22da21 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2597.194791] env[62824]: DEBUG nova.scheduler.client.report [None req-fbd418c9-a6fa-4c4f-a0db-b0bd4f22da21 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2597.700575] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fbd418c9-a6fa-4c4f-a0db-b0bd4f22da21 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.624s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2597.720687] env[62824]: INFO nova.scheduler.client.report [None req-fbd418c9-a6fa-4c4f-a0db-b0bd4f22da21 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Deleted allocations for instance 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79 [ 2598.228469] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fbd418c9-a6fa-4c4f-a0db-b0bd4f22da21 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Lock "8bdfe1c4-1c23-4360-9bb0-90dfc9123d79" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.938s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2599.856720] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Acquiring lock "c41daca7-21fc-44fa-9957-d26767e06658" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2599.857042] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Lock "c41daca7-21fc-44fa-9957-d26767e06658" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2600.359297] env[62824]: DEBUG nova.compute.manager [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c41daca7-21fc-44fa-9957-d26767e06658] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2600.883742] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2600.883997] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2600.885540] env[62824]: INFO nova.compute.claims [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c41daca7-21fc-44fa-9957-d26767e06658] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2601.939083] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b11d13af-034a-425d-b3a5-af9a8aac0fd7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2601.946875] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52468496-d1db-4ac1-8cb1-3e1171563775 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2601.976605] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93c2168f-614a-4b12-9ec6-3a72ba9d55f6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2601.983173] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78cd8cc4-6dd1-4e2f-9992-d941d5d6525a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2601.995603] env[62824]: DEBUG nova.compute.provider_tree [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2602.499275] env[62824]: DEBUG nova.scheduler.client.report [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2603.004362] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.120s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2603.004880] env[62824]: DEBUG nova.compute.manager [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c41daca7-21fc-44fa-9957-d26767e06658] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2603.510291] env[62824]: DEBUG nova.compute.utils [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2603.511938] env[62824]: DEBUG nova.compute.manager [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c41daca7-21fc-44fa-9957-d26767e06658] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2603.511938] env[62824]: DEBUG nova.network.neutron [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c41daca7-21fc-44fa-9957-d26767e06658] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2603.561154] env[62824]: DEBUG nova.policy [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cb36c701db8a471bbd65d2b68ce3becf', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c95aa87980de4ec8a2170c2640423301', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2603.814134] env[62824]: DEBUG nova.network.neutron [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c41daca7-21fc-44fa-9957-d26767e06658] Successfully created port: 5dd607e5-efbc-43a8-ae77-a884eb5dbec4 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2604.014729] env[62824]: DEBUG nova.compute.manager [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c41daca7-21fc-44fa-9957-d26767e06658] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2605.024518] env[62824]: DEBUG nova.compute.manager [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c41daca7-21fc-44fa-9957-d26767e06658] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2605.049620] env[62824]: DEBUG nova.virt.hardware [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2605.049873] env[62824]: DEBUG nova.virt.hardware [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2605.050046] env[62824]: DEBUG nova.virt.hardware [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2605.050238] env[62824]: DEBUG nova.virt.hardware [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2605.050386] env[62824]: DEBUG nova.virt.hardware [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2605.050533] env[62824]: DEBUG nova.virt.hardware [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2605.050749] env[62824]: DEBUG nova.virt.hardware [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2605.050892] env[62824]: DEBUG nova.virt.hardware [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2605.051073] env[62824]: DEBUG nova.virt.hardware [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2605.051242] env[62824]: DEBUG nova.virt.hardware [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2605.051415] env[62824]: DEBUG nova.virt.hardware [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2605.052294] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-caf66bf9-7db4-4d1a-9b0c-d84b2c7fa234 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2605.060297] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d480aa47-9bec-4830-bc97-174cb263759f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2606.108529] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9046ec5a-0906-4699-8479-86db98319af0 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquiring lock "eb0a5669-7b2a-4d2f-9596-6486e8b78289" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2606.108836] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9046ec5a-0906-4699-8479-86db98319af0 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lock "eb0a5669-7b2a-4d2f-9596-6486e8b78289" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2606.611988] env[62824]: DEBUG nova.compute.utils [None req-9046ec5a-0906-4699-8479-86db98319af0 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2607.115505] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9046ec5a-0906-4699-8479-86db98319af0 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lock "eb0a5669-7b2a-4d2f-9596-6486e8b78289" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2607.760929] env[62824]: DEBUG oslo_concurrency.lockutils [None req-10d3af5f-45b0-46a1-801c-9d1f82f172cc tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquiring lock "6e066235-99c3-4bde-972a-345faccf9fa6" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2607.761236] env[62824]: DEBUG oslo_concurrency.lockutils [None req-10d3af5f-45b0-46a1-801c-9d1f82f172cc tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lock "6e066235-99c3-4bde-972a-345faccf9fa6" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2607.761428] env[62824]: DEBUG nova.compute.manager [None req-10d3af5f-45b0-46a1-801c-9d1f82f172cc tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2607.762357] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-548b5f30-030c-4a09-b714-835f645de3a1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2607.769442] env[62824]: DEBUG nova.compute.manager [None req-10d3af5f-45b0-46a1-801c-9d1f82f172cc tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62824) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3404}} [ 2607.769985] env[62824]: DEBUG nova.objects.instance [None req-10d3af5f-45b0-46a1-801c-9d1f82f172cc tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lazy-loading 'flavor' on Instance uuid 6e066235-99c3-4bde-972a-345faccf9fa6 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2608.171700] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9046ec5a-0906-4699-8479-86db98319af0 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquiring lock "eb0a5669-7b2a-4d2f-9596-6486e8b78289" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2608.172099] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9046ec5a-0906-4699-8479-86db98319af0 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lock "eb0a5669-7b2a-4d2f-9596-6486e8b78289" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2608.172197] env[62824]: INFO nova.compute.manager [None req-9046ec5a-0906-4699-8479-86db98319af0 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Attaching volume a0dcd1b7-ffe3-46f6-adc9-597e1c5487d6 to /dev/sdb [ 2608.202114] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b139868a-6e0c-4591-bf08-68d5a070557b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2608.209178] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cc5eef2-4de4-40e9-ae03-1aea2a5c72ef {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2608.222099] env[62824]: DEBUG nova.virt.block_device [None req-9046ec5a-0906-4699-8479-86db98319af0 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Updating existing volume attachment record: 8ec53224-4c4b-415b-bb6c-bfd0a7954973 {{(pid=62824) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 2608.686974] env[62824]: DEBUG nova.compute.manager [req-0e86c048-5120-4b31-b5bb-fb9f3812c1fa req-2e2eb5e9-fe8e-4338-b203-2f10ff6e5394 service nova] [instance: c41daca7-21fc-44fa-9957-d26767e06658] Received event network-vif-plugged-5dd607e5-efbc-43a8-ae77-a884eb5dbec4 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2608.687214] env[62824]: DEBUG oslo_concurrency.lockutils [req-0e86c048-5120-4b31-b5bb-fb9f3812c1fa req-2e2eb5e9-fe8e-4338-b203-2f10ff6e5394 service nova] Acquiring lock "c41daca7-21fc-44fa-9957-d26767e06658-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2608.689455] env[62824]: DEBUG oslo_concurrency.lockutils [req-0e86c048-5120-4b31-b5bb-fb9f3812c1fa req-2e2eb5e9-fe8e-4338-b203-2f10ff6e5394 service nova] Lock "c41daca7-21fc-44fa-9957-d26767e06658-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.002s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2608.689726] env[62824]: DEBUG oslo_concurrency.lockutils [req-0e86c048-5120-4b31-b5bb-fb9f3812c1fa req-2e2eb5e9-fe8e-4338-b203-2f10ff6e5394 service nova] Lock "c41daca7-21fc-44fa-9957-d26767e06658-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2608.689960] env[62824]: DEBUG nova.compute.manager [req-0e86c048-5120-4b31-b5bb-fb9f3812c1fa req-2e2eb5e9-fe8e-4338-b203-2f10ff6e5394 service nova] [instance: c41daca7-21fc-44fa-9957-d26767e06658] No waiting events found dispatching network-vif-plugged-5dd607e5-efbc-43a8-ae77-a884eb5dbec4 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2608.690208] env[62824]: WARNING nova.compute.manager [req-0e86c048-5120-4b31-b5bb-fb9f3812c1fa req-2e2eb5e9-fe8e-4338-b203-2f10ff6e5394 service nova] [instance: c41daca7-21fc-44fa-9957-d26767e06658] Received unexpected event network-vif-plugged-5dd607e5-efbc-43a8-ae77-a884eb5dbec4 for instance with vm_state building and task_state spawning. [ 2608.776597] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-10d3af5f-45b0-46a1-801c-9d1f82f172cc tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2608.777495] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1a6ba6e4-68b2-4f6d-b7d9-d004eda66581 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2608.781455] env[62824]: DEBUG nova.network.neutron [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c41daca7-21fc-44fa-9957-d26767e06658] Successfully updated port: 5dd607e5-efbc-43a8-ae77-a884eb5dbec4 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2608.786850] env[62824]: DEBUG oslo_vmware.api [None req-10d3af5f-45b0-46a1-801c-9d1f82f172cc tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2608.786850] env[62824]: value = "task-2147144" [ 2608.786850] env[62824]: _type = "Task" [ 2608.786850] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2608.795367] env[62824]: DEBUG oslo_vmware.api [None req-10d3af5f-45b0-46a1-801c-9d1f82f172cc tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2147144, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2609.284378] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Acquiring lock "refresh_cache-c41daca7-21fc-44fa-9957-d26767e06658" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2609.284673] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Acquired lock "refresh_cache-c41daca7-21fc-44fa-9957-d26767e06658" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2609.284874] env[62824]: DEBUG nova.network.neutron [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c41daca7-21fc-44fa-9957-d26767e06658] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2609.294419] env[62824]: DEBUG oslo_vmware.api [None req-10d3af5f-45b0-46a1-801c-9d1f82f172cc tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2147144, 'name': PowerOffVM_Task, 'duration_secs': 0.17637} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2609.295294] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-10d3af5f-45b0-46a1-801c-9d1f82f172cc tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2609.295500] env[62824]: DEBUG nova.compute.manager [None req-10d3af5f-45b0-46a1-801c-9d1f82f172cc tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2609.296320] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbc48924-1fbe-42ea-9773-fff844967589 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2609.808550] env[62824]: DEBUG oslo_concurrency.lockutils [None req-10d3af5f-45b0-46a1-801c-9d1f82f172cc tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lock "6e066235-99c3-4bde-972a-345faccf9fa6" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.047s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2609.818018] env[62824]: DEBUG nova.network.neutron [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c41daca7-21fc-44fa-9957-d26767e06658] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2609.955379] env[62824]: DEBUG nova.network.neutron [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c41daca7-21fc-44fa-9957-d26767e06658] Updating instance_info_cache with network_info: [{"id": "5dd607e5-efbc-43a8-ae77-a884eb5dbec4", "address": "fa:16:3e:14:2d:96", "network": {"id": "f7936d74-eeef-4c41-8ae7-9427c43663fa", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1643883257-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c95aa87980de4ec8a2170c2640423301", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6eaa481-1f92-4851-b98e-09ed0daad7cb", "external-id": "nsx-vlan-transportzone-636", "segmentation_id": 636, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5dd607e5-ef", "ovs_interfaceid": "5dd607e5-efbc-43a8-ae77-a884eb5dbec4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2610.121702] env[62824]: DEBUG nova.objects.instance [None req-1522588c-221c-46f2-a977-373de278cd65 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lazy-loading 'flavor' on Instance uuid 6e066235-99c3-4bde-972a-345faccf9fa6 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2610.458413] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Releasing lock "refresh_cache-c41daca7-21fc-44fa-9957-d26767e06658" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2610.458770] env[62824]: DEBUG nova.compute.manager [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c41daca7-21fc-44fa-9957-d26767e06658] Instance network_info: |[{"id": "5dd607e5-efbc-43a8-ae77-a884eb5dbec4", "address": "fa:16:3e:14:2d:96", "network": {"id": "f7936d74-eeef-4c41-8ae7-9427c43663fa", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1643883257-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c95aa87980de4ec8a2170c2640423301", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6eaa481-1f92-4851-b98e-09ed0daad7cb", "external-id": "nsx-vlan-transportzone-636", "segmentation_id": 636, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5dd607e5-ef", "ovs_interfaceid": "5dd607e5-efbc-43a8-ae77-a884eb5dbec4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2610.459234] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c41daca7-21fc-44fa-9957-d26767e06658] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:14:2d:96', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c6eaa481-1f92-4851-b98e-09ed0daad7cb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5dd607e5-efbc-43a8-ae77-a884eb5dbec4', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2610.466903] env[62824]: DEBUG oslo.service.loopingcall [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2610.467133] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c41daca7-21fc-44fa-9957-d26767e06658] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2610.467364] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0f5f4cd1-991a-446b-885f-443941cc0f6b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2610.486614] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2610.486614] env[62824]: value = "task-2147146" [ 2610.486614] env[62824]: _type = "Task" [ 2610.486614] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2610.495305] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2147146, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2610.627475] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1522588c-221c-46f2-a977-373de278cd65 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquiring lock "refresh_cache-6e066235-99c3-4bde-972a-345faccf9fa6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2610.627658] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1522588c-221c-46f2-a977-373de278cd65 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquired lock "refresh_cache-6e066235-99c3-4bde-972a-345faccf9fa6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2610.627813] env[62824]: DEBUG nova.network.neutron [None req-1522588c-221c-46f2-a977-373de278cd65 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2610.627990] env[62824]: DEBUG nova.objects.instance [None req-1522588c-221c-46f2-a977-373de278cd65 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lazy-loading 'info_cache' on Instance uuid 6e066235-99c3-4bde-972a-345faccf9fa6 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2610.712149] env[62824]: DEBUG nova.compute.manager [req-1e44ea4d-cee7-4eaf-aef7-3c6b25e51c45 req-8fbc12b6-62c7-4923-828e-31a06e987b59 service nova] [instance: c41daca7-21fc-44fa-9957-d26767e06658] Received event network-changed-5dd607e5-efbc-43a8-ae77-a884eb5dbec4 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2610.712376] env[62824]: DEBUG nova.compute.manager [req-1e44ea4d-cee7-4eaf-aef7-3c6b25e51c45 req-8fbc12b6-62c7-4923-828e-31a06e987b59 service nova] [instance: c41daca7-21fc-44fa-9957-d26767e06658] Refreshing instance network info cache due to event network-changed-5dd607e5-efbc-43a8-ae77-a884eb5dbec4. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2610.712587] env[62824]: DEBUG oslo_concurrency.lockutils [req-1e44ea4d-cee7-4eaf-aef7-3c6b25e51c45 req-8fbc12b6-62c7-4923-828e-31a06e987b59 service nova] Acquiring lock "refresh_cache-c41daca7-21fc-44fa-9957-d26767e06658" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2610.712732] env[62824]: DEBUG oslo_concurrency.lockutils [req-1e44ea4d-cee7-4eaf-aef7-3c6b25e51c45 req-8fbc12b6-62c7-4923-828e-31a06e987b59 service nova] Acquired lock "refresh_cache-c41daca7-21fc-44fa-9957-d26767e06658" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2610.712895] env[62824]: DEBUG nova.network.neutron [req-1e44ea4d-cee7-4eaf-aef7-3c6b25e51c45 req-8fbc12b6-62c7-4923-828e-31a06e987b59 service nova] [instance: c41daca7-21fc-44fa-9957-d26767e06658] Refreshing network info cache for port 5dd607e5-efbc-43a8-ae77-a884eb5dbec4 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2610.996795] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2147146, 'name': CreateVM_Task, 'duration_secs': 0.393309} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2610.996978] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c41daca7-21fc-44fa-9957-d26767e06658] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2610.997591] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2610.997757] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2610.998102] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2610.998349] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-37ddb466-88e0-4640-b8cf-9b8413b64a30 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2611.002666] env[62824]: DEBUG oslo_vmware.api [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Waiting for the task: (returnval){ [ 2611.002666] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52a97f84-52c7-7680-12dd-2d4f44bb05a3" [ 2611.002666] env[62824]: _type = "Task" [ 2611.002666] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2611.010192] env[62824]: DEBUG oslo_vmware.api [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52a97f84-52c7-7680-12dd-2d4f44bb05a3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2611.130806] env[62824]: DEBUG nova.objects.base [None req-1522588c-221c-46f2-a977-373de278cd65 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Object Instance<6e066235-99c3-4bde-972a-345faccf9fa6> lazy-loaded attributes: flavor,info_cache {{(pid=62824) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 2611.390871] env[62824]: DEBUG nova.network.neutron [req-1e44ea4d-cee7-4eaf-aef7-3c6b25e51c45 req-8fbc12b6-62c7-4923-828e-31a06e987b59 service nova] [instance: c41daca7-21fc-44fa-9957-d26767e06658] Updated VIF entry in instance network info cache for port 5dd607e5-efbc-43a8-ae77-a884eb5dbec4. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2611.391246] env[62824]: DEBUG nova.network.neutron [req-1e44ea4d-cee7-4eaf-aef7-3c6b25e51c45 req-8fbc12b6-62c7-4923-828e-31a06e987b59 service nova] [instance: c41daca7-21fc-44fa-9957-d26767e06658] Updating instance_info_cache with network_info: [{"id": "5dd607e5-efbc-43a8-ae77-a884eb5dbec4", "address": "fa:16:3e:14:2d:96", "network": {"id": "f7936d74-eeef-4c41-8ae7-9427c43663fa", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1643883257-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c95aa87980de4ec8a2170c2640423301", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6eaa481-1f92-4851-b98e-09ed0daad7cb", "external-id": "nsx-vlan-transportzone-636", "segmentation_id": 636, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5dd607e5-ef", "ovs_interfaceid": "5dd607e5-efbc-43a8-ae77-a884eb5dbec4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2611.513182] env[62824]: DEBUG oslo_vmware.api [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52a97f84-52c7-7680-12dd-2d4f44bb05a3, 'name': SearchDatastore_Task, 'duration_secs': 0.009635} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2611.513538] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2611.513670] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c41daca7-21fc-44fa-9957-d26767e06658] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2611.513899] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2611.514056] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2611.514265] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2611.514549] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6d7c45d5-61ef-4a8d-803d-bdbb67fb142a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2611.522279] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2611.522452] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2611.523124] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-30163014-85b2-4f8a-aa15-f42579085c03 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2611.527917] env[62824]: DEBUG oslo_vmware.api [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Waiting for the task: (returnval){ [ 2611.527917] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52e387a6-881d-c3a4-a287-aab5cbfd897e" [ 2611.527917] env[62824]: _type = "Task" [ 2611.527917] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2611.535303] env[62824]: DEBUG oslo_vmware.api [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52e387a6-881d-c3a4-a287-aab5cbfd897e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2611.829390] env[62824]: DEBUG nova.network.neutron [None req-1522588c-221c-46f2-a977-373de278cd65 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Updating instance_info_cache with network_info: [{"id": "d8d8df40-7696-495a-9dca-6e76f780e1a7", "address": "fa:16:3e:4d:47:a1", "network": {"id": "bba9b98c-c7eb-4890-8de5-eafacfdebd7d", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1928394728-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.160", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7fcdbfc29df64a1ba8d982bdcc667b64", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b49c5024-2ced-42ca-90cc-6066766d43e6", "external-id": "nsx-vlan-transportzone-239", "segmentation_id": 239, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd8d8df40-76", "ovs_interfaceid": "d8d8df40-7696-495a-9dca-6e76f780e1a7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2611.893685] env[62824]: DEBUG oslo_concurrency.lockutils [req-1e44ea4d-cee7-4eaf-aef7-3c6b25e51c45 req-8fbc12b6-62c7-4923-828e-31a06e987b59 service nova] Releasing lock "refresh_cache-c41daca7-21fc-44fa-9957-d26767e06658" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2612.038862] env[62824]: DEBUG oslo_vmware.api [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52e387a6-881d-c3a4-a287-aab5cbfd897e, 'name': SearchDatastore_Task, 'duration_secs': 0.008086} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2612.039593] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c289a5cc-bb05-44aa-b092-523fa32c115a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2612.044309] env[62824]: DEBUG oslo_vmware.api [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Waiting for the task: (returnval){ [ 2612.044309] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5269c2d5-52ff-5c2a-add2-0adf2c0c76d6" [ 2612.044309] env[62824]: _type = "Task" [ 2612.044309] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2612.051245] env[62824]: DEBUG oslo_vmware.api [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5269c2d5-52ff-5c2a-add2-0adf2c0c76d6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2612.330712] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1522588c-221c-46f2-a977-373de278cd65 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Releasing lock "refresh_cache-6e066235-99c3-4bde-972a-345faccf9fa6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2612.554184] env[62824]: DEBUG oslo_vmware.api [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5269c2d5-52ff-5c2a-add2-0adf2c0c76d6, 'name': SearchDatastore_Task, 'duration_secs': 0.009216} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2612.554530] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2612.554784] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] c41daca7-21fc-44fa-9957-d26767e06658/c41daca7-21fc-44fa-9957-d26767e06658.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2612.555050] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a3b1f8a3-9c39-4d45-aceb-f83de0162dcf {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2612.562222] env[62824]: DEBUG oslo_vmware.api [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Waiting for the task: (returnval){ [ 2612.562222] env[62824]: value = "task-2147147" [ 2612.562222] env[62824]: _type = "Task" [ 2612.562222] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2612.569449] env[62824]: DEBUG oslo_vmware.api [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147147, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2612.764653] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-9046ec5a-0906-4699-8479-86db98319af0 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Volume attach. Driver type: vmdk {{(pid=62824) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 2612.764903] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-9046ec5a-0906-4699-8479-86db98319af0 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-438794', 'volume_id': 'a0dcd1b7-ffe3-46f6-adc9-597e1c5487d6', 'name': 'volume-a0dcd1b7-ffe3-46f6-adc9-597e1c5487d6', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'eb0a5669-7b2a-4d2f-9596-6486e8b78289', 'attached_at': '', 'detached_at': '', 'volume_id': 'a0dcd1b7-ffe3-46f6-adc9-597e1c5487d6', 'serial': 'a0dcd1b7-ffe3-46f6-adc9-597e1c5487d6'} {{(pid=62824) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 2612.765907] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc483403-fdf0-461d-a051-6b3f6a37bf3e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2612.784805] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a236c628-443c-4832-9d35-b665a17d81cb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2612.811575] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-9046ec5a-0906-4699-8479-86db98319af0 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Reconfiguring VM instance instance-00000078 to attach disk [datastore1] volume-a0dcd1b7-ffe3-46f6-adc9-597e1c5487d6/volume-a0dcd1b7-ffe3-46f6-adc9-597e1c5487d6.vmdk or device None with type thin {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2612.811933] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2a05a6bf-ea2e-4b1a-ab32-b1c622667752 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2612.832941] env[62824]: DEBUG oslo_vmware.api [None req-9046ec5a-0906-4699-8479-86db98319af0 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2612.832941] env[62824]: value = "task-2147148" [ 2612.832941] env[62824]: _type = "Task" [ 2612.832941] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2612.844563] env[62824]: DEBUG oslo_vmware.api [None req-9046ec5a-0906-4699-8479-86db98319af0 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147148, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2613.072823] env[62824]: DEBUG oslo_vmware.api [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147147, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.419472} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2613.073079] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore1] c41daca7-21fc-44fa-9957-d26767e06658/c41daca7-21fc-44fa-9957-d26767e06658.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2613.073302] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c41daca7-21fc-44fa-9957-d26767e06658] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2613.073543] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fea1da50-2f7a-4231-b24a-1fd26ec28dfd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2613.080091] env[62824]: DEBUG oslo_vmware.api [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Waiting for the task: (returnval){ [ 2613.080091] env[62824]: value = "task-2147149" [ 2613.080091] env[62824]: _type = "Task" [ 2613.080091] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2613.087643] env[62824]: DEBUG oslo_vmware.api [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147149, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2613.339605] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-1522588c-221c-46f2-a977-373de278cd65 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2613.339792] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7201d442-9ee0-455b-86e5-dc31d99d3143 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2613.346802] env[62824]: DEBUG oslo_vmware.api [None req-9046ec5a-0906-4699-8479-86db98319af0 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147148, 'name': ReconfigVM_Task, 'duration_secs': 0.440466} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2613.347935] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-9046ec5a-0906-4699-8479-86db98319af0 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Reconfigured VM instance instance-00000078 to attach disk [datastore1] volume-a0dcd1b7-ffe3-46f6-adc9-597e1c5487d6/volume-a0dcd1b7-ffe3-46f6-adc9-597e1c5487d6.vmdk or device None with type thin {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2613.352684] env[62824]: DEBUG oslo_vmware.api [None req-1522588c-221c-46f2-a977-373de278cd65 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2613.352684] env[62824]: value = "task-2147150" [ 2613.352684] env[62824]: _type = "Task" [ 2613.352684] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2613.352905] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-085890ba-fc3f-4a2d-a007-9235099cb8ef {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2613.370268] env[62824]: DEBUG oslo_vmware.api [None req-1522588c-221c-46f2-a977-373de278cd65 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2147150, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2613.371397] env[62824]: DEBUG oslo_vmware.api [None req-9046ec5a-0906-4699-8479-86db98319af0 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2613.371397] env[62824]: value = "task-2147151" [ 2613.371397] env[62824]: _type = "Task" [ 2613.371397] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2613.379440] env[62824]: DEBUG oslo_vmware.api [None req-9046ec5a-0906-4699-8479-86db98319af0 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147151, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2613.590180] env[62824]: DEBUG oslo_vmware.api [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147149, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064756} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2613.590541] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c41daca7-21fc-44fa-9957-d26767e06658] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2613.591192] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccd89069-f94b-4dc3-84ef-7f6299dad8aa {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2613.612438] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c41daca7-21fc-44fa-9957-d26767e06658] Reconfiguring VM instance instance-00000079 to attach disk [datastore1] c41daca7-21fc-44fa-9957-d26767e06658/c41daca7-21fc-44fa-9957-d26767e06658.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2613.612692] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-260ec3c1-25a4-4e14-9765-0048a4594f8a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2613.632745] env[62824]: DEBUG oslo_vmware.api [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Waiting for the task: (returnval){ [ 2613.632745] env[62824]: value = "task-2147152" [ 2613.632745] env[62824]: _type = "Task" [ 2613.632745] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2613.640283] env[62824]: DEBUG oslo_vmware.api [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147152, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2613.873098] env[62824]: DEBUG oslo_vmware.api [None req-1522588c-221c-46f2-a977-373de278cd65 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2147150, 'name': PowerOnVM_Task, 'duration_secs': 0.398532} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2613.876359] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-1522588c-221c-46f2-a977-373de278cd65 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2613.876621] env[62824]: DEBUG nova.compute.manager [None req-1522588c-221c-46f2-a977-373de278cd65 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2613.877771] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc45bc7e-d42b-43f0-aee1-6318003b56e3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2613.887407] env[62824]: DEBUG oslo_vmware.api [None req-9046ec5a-0906-4699-8479-86db98319af0 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147151, 'name': ReconfigVM_Task, 'duration_secs': 0.14675} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2613.889142] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-9046ec5a-0906-4699-8479-86db98319af0 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-438794', 'volume_id': 'a0dcd1b7-ffe3-46f6-adc9-597e1c5487d6', 'name': 'volume-a0dcd1b7-ffe3-46f6-adc9-597e1c5487d6', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'eb0a5669-7b2a-4d2f-9596-6486e8b78289', 'attached_at': '', 'detached_at': '', 'volume_id': 'a0dcd1b7-ffe3-46f6-adc9-597e1c5487d6', 'serial': 'a0dcd1b7-ffe3-46f6-adc9-597e1c5487d6'} {{(pid=62824) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 2614.142710] env[62824]: DEBUG oslo_vmware.api [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147152, 'name': ReconfigVM_Task, 'duration_secs': 0.503038} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2614.143007] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c41daca7-21fc-44fa-9957-d26767e06658] Reconfigured VM instance instance-00000079 to attach disk [datastore1] c41daca7-21fc-44fa-9957-d26767e06658/c41daca7-21fc-44fa-9957-d26767e06658.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2614.143644] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6c79da87-7d01-47cd-a1b3-cbbffa0097b4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2614.151268] env[62824]: DEBUG oslo_vmware.api [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Waiting for the task: (returnval){ [ 2614.151268] env[62824]: value = "task-2147153" [ 2614.151268] env[62824]: _type = "Task" [ 2614.151268] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2614.162478] env[62824]: DEBUG oslo_vmware.api [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147153, 'name': Rename_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2614.660740] env[62824]: DEBUG oslo_vmware.api [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147153, 'name': Rename_Task, 'duration_secs': 0.127617} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2614.661113] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c41daca7-21fc-44fa-9957-d26767e06658] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2614.661225] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9cca520f-2b4c-4699-8a07-d8eaf2004db0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2614.667460] env[62824]: DEBUG oslo_vmware.api [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Waiting for the task: (returnval){ [ 2614.667460] env[62824]: value = "task-2147154" [ 2614.667460] env[62824]: _type = "Task" [ 2614.667460] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2614.675547] env[62824]: DEBUG oslo_vmware.api [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147154, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2614.890849] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e84bf54-4bfb-4dcb-88f5-310066bbe5e7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2614.897295] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-877d4b39-fa63-4162-8e52-624f7a03c37c tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Suspending the VM {{(pid=62824) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 2614.897520] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-6270b203-d1ca-4702-ab2c-3405596dffc0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2614.903168] env[62824]: DEBUG oslo_vmware.api [None req-877d4b39-fa63-4162-8e52-624f7a03c37c tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2614.903168] env[62824]: value = "task-2147155" [ 2614.903168] env[62824]: _type = "Task" [ 2614.903168] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2614.911264] env[62824]: DEBUG oslo_vmware.api [None req-877d4b39-fa63-4162-8e52-624f7a03c37c tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2147155, 'name': SuspendVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2614.930237] env[62824]: DEBUG nova.objects.instance [None req-9046ec5a-0906-4699-8479-86db98319af0 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lazy-loading 'flavor' on Instance uuid eb0a5669-7b2a-4d2f-9596-6486e8b78289 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2615.177991] env[62824]: DEBUG oslo_vmware.api [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147154, 'name': PowerOnVM_Task, 'duration_secs': 0.474985} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2615.178309] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c41daca7-21fc-44fa-9957-d26767e06658] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2615.178518] env[62824]: INFO nova.compute.manager [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c41daca7-21fc-44fa-9957-d26767e06658] Took 10.15 seconds to spawn the instance on the hypervisor. [ 2615.178699] env[62824]: DEBUG nova.compute.manager [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c41daca7-21fc-44fa-9957-d26767e06658] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2615.179546] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45fce7d6-068a-440b-9bf0-23e4a710380a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2615.413226] env[62824]: DEBUG oslo_vmware.api [None req-877d4b39-fa63-4162-8e52-624f7a03c37c tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2147155, 'name': SuspendVM_Task} progress is 62%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2615.435464] env[62824]: DEBUG oslo_concurrency.lockutils [None req-9046ec5a-0906-4699-8479-86db98319af0 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lock "eb0a5669-7b2a-4d2f-9596-6486e8b78289" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.263s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2615.697846] env[62824]: INFO nova.compute.manager [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c41daca7-21fc-44fa-9957-d26767e06658] Took 14.83 seconds to build instance. [ 2615.913893] env[62824]: DEBUG oslo_vmware.api [None req-877d4b39-fa63-4162-8e52-624f7a03c37c tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2147155, 'name': SuspendVM_Task, 'duration_secs': 0.595707} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2615.914298] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-877d4b39-fa63-4162-8e52-624f7a03c37c tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Suspended the VM {{(pid=62824) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 2615.914538] env[62824]: DEBUG nova.compute.manager [None req-877d4b39-fa63-4162-8e52-624f7a03c37c tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2615.915391] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e2c6bba-2567-43f8-bf2c-b0ba05bfacfc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2616.181140] env[62824]: DEBUG nova.compute.manager [req-1496d6da-4ccb-4ac7-afdf-07893270b48a req-6196a1cd-2b40-4f1a-adf8-50f11c957ee7 service nova] [instance: c41daca7-21fc-44fa-9957-d26767e06658] Received event network-changed-5dd607e5-efbc-43a8-ae77-a884eb5dbec4 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2616.181364] env[62824]: DEBUG nova.compute.manager [req-1496d6da-4ccb-4ac7-afdf-07893270b48a req-6196a1cd-2b40-4f1a-adf8-50f11c957ee7 service nova] [instance: c41daca7-21fc-44fa-9957-d26767e06658] Refreshing instance network info cache due to event network-changed-5dd607e5-efbc-43a8-ae77-a884eb5dbec4. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2616.181574] env[62824]: DEBUG oslo_concurrency.lockutils [req-1496d6da-4ccb-4ac7-afdf-07893270b48a req-6196a1cd-2b40-4f1a-adf8-50f11c957ee7 service nova] Acquiring lock "refresh_cache-c41daca7-21fc-44fa-9957-d26767e06658" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2616.181717] env[62824]: DEBUG oslo_concurrency.lockutils [req-1496d6da-4ccb-4ac7-afdf-07893270b48a req-6196a1cd-2b40-4f1a-adf8-50f11c957ee7 service nova] Acquired lock "refresh_cache-c41daca7-21fc-44fa-9957-d26767e06658" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2616.181877] env[62824]: DEBUG nova.network.neutron [req-1496d6da-4ccb-4ac7-afdf-07893270b48a req-6196a1cd-2b40-4f1a-adf8-50f11c957ee7 service nova] [instance: c41daca7-21fc-44fa-9957-d26767e06658] Refreshing network info cache for port 5dd607e5-efbc-43a8-ae77-a884eb5dbec4 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2616.199896] env[62824]: DEBUG oslo_concurrency.lockutils [None req-c45f3128-df8d-4ed8-8c14-c92ed0aecbd5 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Lock "c41daca7-21fc-44fa-9957-d26767e06658" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.343s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2616.279061] env[62824]: DEBUG nova.compute.manager [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Stashing vm_state: active {{(pid=62824) _prep_resize /opt/stack/nova/nova/compute/manager.py:6136}} [ 2616.803064] env[62824]: DEBUG oslo_concurrency.lockutils [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2616.803392] env[62824]: DEBUG oslo_concurrency.lockutils [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2616.895904] env[62824]: DEBUG nova.network.neutron [req-1496d6da-4ccb-4ac7-afdf-07893270b48a req-6196a1cd-2b40-4f1a-adf8-50f11c957ee7 service nova] [instance: c41daca7-21fc-44fa-9957-d26767e06658] Updated VIF entry in instance network info cache for port 5dd607e5-efbc-43a8-ae77-a884eb5dbec4. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2616.896279] env[62824]: DEBUG nova.network.neutron [req-1496d6da-4ccb-4ac7-afdf-07893270b48a req-6196a1cd-2b40-4f1a-adf8-50f11c957ee7 service nova] [instance: c41daca7-21fc-44fa-9957-d26767e06658] Updating instance_info_cache with network_info: [{"id": "5dd607e5-efbc-43a8-ae77-a884eb5dbec4", "address": "fa:16:3e:14:2d:96", "network": {"id": "f7936d74-eeef-4c41-8ae7-9427c43663fa", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1643883257-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.226", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c95aa87980de4ec8a2170c2640423301", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6eaa481-1f92-4851-b98e-09ed0daad7cb", "external-id": "nsx-vlan-transportzone-636", "segmentation_id": 636, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5dd607e5-ef", "ovs_interfaceid": "5dd607e5-efbc-43a8-ae77-a884eb5dbec4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2617.259918] env[62824]: INFO nova.compute.manager [None req-ef452665-25f7-4a5a-ac2d-6bbe04c9cf48 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Resuming [ 2617.260563] env[62824]: DEBUG nova.objects.instance [None req-ef452665-25f7-4a5a-ac2d-6bbe04c9cf48 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lazy-loading 'flavor' on Instance uuid 6e066235-99c3-4bde-972a-345faccf9fa6 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2617.308606] env[62824]: INFO nova.compute.claims [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2617.398998] env[62824]: DEBUG oslo_concurrency.lockutils [req-1496d6da-4ccb-4ac7-afdf-07893270b48a req-6196a1cd-2b40-4f1a-adf8-50f11c957ee7 service nova] Releasing lock "refresh_cache-c41daca7-21fc-44fa-9957-d26767e06658" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2617.773200] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2617.773478] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager.update_available_resource {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2617.815184] env[62824]: INFO nova.compute.resource_tracker [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Updating resource usage from migration 696bd587-625b-4376-993d-f1bc5f435717 [ 2617.883331] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2fb51ce-3153-4844-b4bb-8b457bdd9d85 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2617.891507] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1bb8951-f585-4da0-a2bc-fe5aed2fd6f9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2617.921943] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bade274-b51a-40ed-8a6f-2ff83bdd2a1c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2617.929825] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55e8e126-9a24-401a-aa07-fd73189cccd0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2617.943830] env[62824]: DEBUG nova.compute.provider_tree [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2618.277269] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2618.447967] env[62824]: DEBUG nova.scheduler.client.report [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2618.772664] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ef452665-25f7-4a5a-ac2d-6bbe04c9cf48 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquiring lock "refresh_cache-6e066235-99c3-4bde-972a-345faccf9fa6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2618.772849] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ef452665-25f7-4a5a-ac2d-6bbe04c9cf48 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquired lock "refresh_cache-6e066235-99c3-4bde-972a-345faccf9fa6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2618.772973] env[62824]: DEBUG nova.network.neutron [None req-ef452665-25f7-4a5a-ac2d-6bbe04c9cf48 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2618.953508] env[62824]: DEBUG oslo_concurrency.lockutils [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.150s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2618.953893] env[62824]: INFO nova.compute.manager [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Migrating [ 2618.964350] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.687s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2618.964540] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2618.964695] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62824) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 2618.969156] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41e57c61-6dad-44c5-8193-9c84d5344747 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2618.978412] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d69c0b0f-6e29-4f08-bc2a-93884ad294af {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2618.994029] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b80dc9fa-48df-4b72-9446-911df4a3186c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2619.000156] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83dba1fc-fe4b-4234-924c-81e0073e4d1c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2619.029379] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180535MB free_disk=175GB free_vcpus=48 pci_devices=None {{(pid=62824) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 2619.029529] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2619.029676] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2619.469483] env[62824]: DEBUG oslo_concurrency.lockutils [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquiring lock "refresh_cache-eb0a5669-7b2a-4d2f-9596-6486e8b78289" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2619.469649] env[62824]: DEBUG oslo_concurrency.lockutils [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquired lock "refresh_cache-eb0a5669-7b2a-4d2f-9596-6486e8b78289" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2619.469825] env[62824]: DEBUG nova.network.neutron [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2619.473588] env[62824]: DEBUG nova.network.neutron [None req-ef452665-25f7-4a5a-ac2d-6bbe04c9cf48 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Updating instance_info_cache with network_info: [{"id": "d8d8df40-7696-495a-9dca-6e76f780e1a7", "address": "fa:16:3e:4d:47:a1", "network": {"id": "bba9b98c-c7eb-4890-8de5-eafacfdebd7d", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1928394728-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.160", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7fcdbfc29df64a1ba8d982bdcc667b64", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b49c5024-2ced-42ca-90cc-6066766d43e6", "external-id": "nsx-vlan-transportzone-239", "segmentation_id": 239, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd8d8df40-76", "ovs_interfaceid": "d8d8df40-7696-495a-9dca-6e76f780e1a7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2619.975747] env[62824]: DEBUG oslo_concurrency.lockutils [None req-ef452665-25f7-4a5a-ac2d-6bbe04c9cf48 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Releasing lock "refresh_cache-6e066235-99c3-4bde-972a-345faccf9fa6" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2619.976432] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7f63ecb-85f8-445f-8b35-da8ccf512262 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2619.982925] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-ef452665-25f7-4a5a-ac2d-6bbe04c9cf48 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Resuming the VM {{(pid=62824) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1183}} [ 2619.983156] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d04ea33d-0eda-4e76-a8cb-fd16fbaccd32 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2619.989193] env[62824]: DEBUG oslo_vmware.api [None req-ef452665-25f7-4a5a-ac2d-6bbe04c9cf48 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2619.989193] env[62824]: value = "task-2147156" [ 2619.989193] env[62824]: _type = "Task" [ 2619.989193] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2619.996306] env[62824]: DEBUG oslo_vmware.api [None req-ef452665-25f7-4a5a-ac2d-6bbe04c9cf48 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2147156, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2620.036613] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Applying migration context for instance eb0a5669-7b2a-4d2f-9596-6486e8b78289 as it has an incoming, in-progress migration 696bd587-625b-4376-993d-f1bc5f435717. Migration status is pre-migrating {{(pid=62824) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 2620.037415] env[62824]: INFO nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Updating resource usage from migration 696bd587-625b-4376-993d-f1bc5f435717 [ 2620.053021] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 6e066235-99c3-4bde-972a-345faccf9fa6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2620.053021] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance c41daca7-21fc-44fa-9957-d26767e06658 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2620.053021] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Migration 696bd587-625b-4376-993d-f1bc5f435717 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 2620.053296] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance eb0a5669-7b2a-4d2f-9596-6486e8b78289 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2620.053296] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 2620.053384] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1344MB phys_disk=200GB used_disk=4GB total_vcpus=48 used_vcpus=4 pci_stats=[] {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 2620.110948] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7bb97fe-e70e-422c-b8a0-44d5f8396c36 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2620.120252] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27749e3a-7c06-4e5c-9f05-c43387ea20b4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2620.150507] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50613274-9809-4da5-ac9e-c06c096ae1c4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2620.159813] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8e48f70-d8f0-4ac8-b8d2-3927436e6621 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2620.172722] env[62824]: DEBUG nova.compute.provider_tree [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2620.189105] env[62824]: DEBUG nova.network.neutron [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Updating instance_info_cache with network_info: [{"id": "8be13d2b-3033-431e-92aa-3e05a180498f", "address": "fa:16:3e:9f:4d:c4", "network": {"id": "5233a794-3514-4912-a857-7e76a66b1e8f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-910685188-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.129", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0eb41836305847fca773afebdafa01db", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd98a25d-a7a9-4fb5-8fef-e8df4dbbbf11", "external-id": "nsx-vlan-transportzone-707", "segmentation_id": 707, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8be13d2b-30", "ovs_interfaceid": "8be13d2b-3033-431e-92aa-3e05a180498f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2620.501530] env[62824]: DEBUG oslo_vmware.api [None req-ef452665-25f7-4a5a-ac2d-6bbe04c9cf48 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2147156, 'name': PowerOnVM_Task, 'duration_secs': 0.48024} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2620.501923] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-ef452665-25f7-4a5a-ac2d-6bbe04c9cf48 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Resumed the VM {{(pid=62824) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1188}} [ 2620.502197] env[62824]: DEBUG nova.compute.manager [None req-ef452665-25f7-4a5a-ac2d-6bbe04c9cf48 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2620.503363] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b50f09fe-e5ff-4875-82ce-0d22724df6d6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2620.675788] env[62824]: DEBUG nova.scheduler.client.report [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2620.691181] env[62824]: DEBUG oslo_concurrency.lockutils [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Releasing lock "refresh_cache-eb0a5669-7b2a-4d2f-9596-6486e8b78289" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2621.181343] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62824) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 2621.181652] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.152s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2621.181760] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2621.181939] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Cleaning up deleted instances with incomplete migration {{(pid=62824) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11783}} [ 2621.946901] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fad9d49c-8bf2-4972-aa05-dbdb2aa7fbd6 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquiring lock "6e066235-99c3-4bde-972a-345faccf9fa6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2621.947190] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fad9d49c-8bf2-4972-aa05-dbdb2aa7fbd6 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lock "6e066235-99c3-4bde-972a-345faccf9fa6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2621.947409] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fad9d49c-8bf2-4972-aa05-dbdb2aa7fbd6 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquiring lock "6e066235-99c3-4bde-972a-345faccf9fa6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2621.947632] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fad9d49c-8bf2-4972-aa05-dbdb2aa7fbd6 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lock "6e066235-99c3-4bde-972a-345faccf9fa6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2621.947821] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fad9d49c-8bf2-4972-aa05-dbdb2aa7fbd6 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lock "6e066235-99c3-4bde-972a-345faccf9fa6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2621.950055] env[62824]: INFO nova.compute.manager [None req-fad9d49c-8bf2-4972-aa05-dbdb2aa7fbd6 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Terminating instance [ 2622.205919] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bfde59e-794d-4267-8f53-c9847d6fa036 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2622.227923] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Updating instance 'eb0a5669-7b2a-4d2f-9596-6486e8b78289' progress to 0 {{(pid=62824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 2622.453945] env[62824]: DEBUG nova.compute.manager [None req-fad9d49c-8bf2-4972-aa05-dbdb2aa7fbd6 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2622.454189] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-fad9d49c-8bf2-4972-aa05-dbdb2aa7fbd6 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2622.455053] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cf73eb1-4a37-4fea-8bd2-2a25e743b66b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2622.462729] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-fad9d49c-8bf2-4972-aa05-dbdb2aa7fbd6 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2622.462951] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8444fdff-701e-4a90-8f97-feb1461134eb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2622.469262] env[62824]: DEBUG oslo_vmware.api [None req-fad9d49c-8bf2-4972-aa05-dbdb2aa7fbd6 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2622.469262] env[62824]: value = "task-2147157" [ 2622.469262] env[62824]: _type = "Task" [ 2622.469262] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2622.478092] env[62824]: DEBUG oslo_vmware.api [None req-fad9d49c-8bf2-4972-aa05-dbdb2aa7fbd6 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2147157, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2622.733972] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2622.734292] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-47d589fe-bbf4-4dcc-b4f3-8dc7eb7af0e0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2622.740934] env[62824]: DEBUG oslo_vmware.api [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2622.740934] env[62824]: value = "task-2147158" [ 2622.740934] env[62824]: _type = "Task" [ 2622.740934] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2622.748459] env[62824]: DEBUG oslo_vmware.api [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147158, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2622.979476] env[62824]: DEBUG oslo_vmware.api [None req-fad9d49c-8bf2-4972-aa05-dbdb2aa7fbd6 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2147157, 'name': PowerOffVM_Task, 'duration_secs': 0.191762} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2622.979661] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-fad9d49c-8bf2-4972-aa05-dbdb2aa7fbd6 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2622.979805] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-fad9d49c-8bf2-4972-aa05-dbdb2aa7fbd6 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2622.980076] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1a50ee9c-c7f5-48d1-b174-35330c8c9bf6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2623.251126] env[62824]: DEBUG oslo_vmware.api [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147158, 'name': PowerOffVM_Task, 'duration_secs': 0.267851} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2623.251418] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2623.251583] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Updating instance 'eb0a5669-7b2a-4d2f-9596-6486e8b78289' progress to 17 {{(pid=62824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 2623.758658] env[62824]: DEBUG nova.virt.hardware [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:24Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2623.758926] env[62824]: DEBUG nova.virt.hardware [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2623.759099] env[62824]: DEBUG nova.virt.hardware [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2623.759289] env[62824]: DEBUG nova.virt.hardware [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2623.759440] env[62824]: DEBUG nova.virt.hardware [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2623.759585] env[62824]: DEBUG nova.virt.hardware [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2623.759789] env[62824]: DEBUG nova.virt.hardware [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2623.759949] env[62824]: DEBUG nova.virt.hardware [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2623.760140] env[62824]: DEBUG nova.virt.hardware [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2623.760307] env[62824]: DEBUG nova.virt.hardware [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2623.760484] env[62824]: DEBUG nova.virt.hardware [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2623.765544] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-db2a1e69-3364-4d17-9e9b-a5986cc3be46 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2623.781523] env[62824]: DEBUG oslo_vmware.api [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2623.781523] env[62824]: value = "task-2147160" [ 2623.781523] env[62824]: _type = "Task" [ 2623.781523] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2623.789274] env[62824]: DEBUG oslo_vmware.api [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147160, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2624.291407] env[62824]: DEBUG oslo_vmware.api [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147160, 'name': ReconfigVM_Task, 'duration_secs': 0.142818} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2624.291713] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Updating instance 'eb0a5669-7b2a-4d2f-9596-6486e8b78289' progress to 33 {{(pid=62824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 2624.798244] env[62824]: DEBUG nova.virt.hardware [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2624.798484] env[62824]: DEBUG nova.virt.hardware [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2624.798640] env[62824]: DEBUG nova.virt.hardware [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2624.798825] env[62824]: DEBUG nova.virt.hardware [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2624.798970] env[62824]: DEBUG nova.virt.hardware [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2624.799132] env[62824]: DEBUG nova.virt.hardware [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2624.799355] env[62824]: DEBUG nova.virt.hardware [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2624.799532] env[62824]: DEBUG nova.virt.hardware [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2624.799735] env[62824]: DEBUG nova.virt.hardware [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2624.799918] env[62824]: DEBUG nova.virt.hardware [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2624.800127] env[62824]: DEBUG nova.virt.hardware [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2624.805578] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Reconfiguring VM instance instance-00000078 to detach disk 2000 {{(pid=62824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 2624.805857] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bfc9f396-81b6-4482-a97e-65afe437cc75 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2624.832465] env[62824]: DEBUG oslo_vmware.api [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2624.832465] env[62824]: value = "task-2147161" [ 2624.832465] env[62824]: _type = "Task" [ 2624.832465] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2624.841537] env[62824]: DEBUG oslo_vmware.api [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147161, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2625.276549] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-fad9d49c-8bf2-4972-aa05-dbdb2aa7fbd6 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2625.276778] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-fad9d49c-8bf2-4972-aa05-dbdb2aa7fbd6 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Deleting contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2625.276960] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-fad9d49c-8bf2-4972-aa05-dbdb2aa7fbd6 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Deleting the datastore file [datastore2] 6e066235-99c3-4bde-972a-345faccf9fa6 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2625.277258] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-973d232e-4690-4042-9182-be5aa111a610 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2625.284077] env[62824]: DEBUG oslo_vmware.api [None req-fad9d49c-8bf2-4972-aa05-dbdb2aa7fbd6 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for the task: (returnval){ [ 2625.284077] env[62824]: value = "task-2147162" [ 2625.284077] env[62824]: _type = "Task" [ 2625.284077] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2625.291815] env[62824]: DEBUG oslo_vmware.api [None req-fad9d49c-8bf2-4972-aa05-dbdb2aa7fbd6 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2147162, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2625.341921] env[62824]: DEBUG oslo_vmware.api [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147161, 'name': ReconfigVM_Task, 'duration_secs': 0.189053} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2625.342246] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Reconfigured VM instance instance-00000078 to detach disk 2000 {{(pid=62824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 2625.343057] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96f9fa5b-d796-468f-b5f3-7e7b9bf58822 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2625.368778] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Reconfiguring VM instance instance-00000078 to attach disk [datastore2] eb0a5669-7b2a-4d2f-9596-6486e8b78289/eb0a5669-7b2a-4d2f-9596-6486e8b78289.vmdk or device None with type thin {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2625.369033] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cf2211f0-58d3-4e2a-8bc2-382bf07bf70c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2625.386677] env[62824]: DEBUG oslo_vmware.api [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2625.386677] env[62824]: value = "task-2147163" [ 2625.386677] env[62824]: _type = "Task" [ 2625.386677] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2625.394436] env[62824]: DEBUG oslo_vmware.api [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147163, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2625.683141] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2625.683389] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Starting heal instance info cache {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10445}} [ 2625.795073] env[62824]: DEBUG oslo_vmware.api [None req-fad9d49c-8bf2-4972-aa05-dbdb2aa7fbd6 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Task: {'id': task-2147162, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.142966} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2625.795342] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-fad9d49c-8bf2-4972-aa05-dbdb2aa7fbd6 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2625.795529] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-fad9d49c-8bf2-4972-aa05-dbdb2aa7fbd6 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Deleted contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2625.795700] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-fad9d49c-8bf2-4972-aa05-dbdb2aa7fbd6 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2625.795870] env[62824]: INFO nova.compute.manager [None req-fad9d49c-8bf2-4972-aa05-dbdb2aa7fbd6 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Took 3.34 seconds to destroy the instance on the hypervisor. [ 2625.796139] env[62824]: DEBUG oslo.service.loopingcall [None req-fad9d49c-8bf2-4972-aa05-dbdb2aa7fbd6 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2625.796309] env[62824]: DEBUG nova.compute.manager [-] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2625.796403] env[62824]: DEBUG nova.network.neutron [-] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2625.896581] env[62824]: DEBUG oslo_vmware.api [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147163, 'name': ReconfigVM_Task, 'duration_secs': 0.273781} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2625.896847] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Reconfigured VM instance instance-00000078 to attach disk [datastore2] eb0a5669-7b2a-4d2f-9596-6486e8b78289/eb0a5669-7b2a-4d2f-9596-6486e8b78289.vmdk or device None with type thin {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2625.897160] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Updating instance 'eb0a5669-7b2a-4d2f-9596-6486e8b78289' progress to 50 {{(pid=62824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 2626.240923] env[62824]: DEBUG nova.compute.manager [req-16049e55-3bc2-40d5-ac2d-fe239b7ce156 req-f77c39dc-c7a0-4c8d-95e5-7849daec3626 service nova] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Received event network-vif-deleted-d8d8df40-7696-495a-9dca-6e76f780e1a7 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2626.241139] env[62824]: INFO nova.compute.manager [req-16049e55-3bc2-40d5-ac2d-fe239b7ce156 req-f77c39dc-c7a0-4c8d-95e5-7849daec3626 service nova] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Neutron deleted interface d8d8df40-7696-495a-9dca-6e76f780e1a7; detaching it from the instance and deleting it from the info cache [ 2626.241275] env[62824]: DEBUG nova.network.neutron [req-16049e55-3bc2-40d5-ac2d-fe239b7ce156 req-f77c39dc-c7a0-4c8d-95e5-7849daec3626 service nova] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2626.403601] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e535ad1c-32ce-4fb2-acd8-6505c7b2cce2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2626.427251] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31065206-3856-4d35-abe9-437019b25003 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2626.447031] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Updating instance 'eb0a5669-7b2a-4d2f-9596-6486e8b78289' progress to 67 {{(pid=62824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 2626.689062] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Didn't find any instances for network info cache update. {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10531}} [ 2626.689062] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2626.689304] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2626.689430] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2626.689667] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2626.720754] env[62824]: DEBUG nova.network.neutron [-] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2626.745054] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b542b473-c1f8-4d2b-b7df-afb95ae515d3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2626.753781] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1b82274-e82a-4d96-91bc-41a3b9e40eb4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2626.772473] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2626.772682] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Cleaning up deleted instances {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11745}} [ 2626.779886] env[62824]: DEBUG nova.compute.manager [req-16049e55-3bc2-40d5-ac2d-fe239b7ce156 req-f77c39dc-c7a0-4c8d-95e5-7849daec3626 service nova] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Detach interface failed, port_id=d8d8df40-7696-495a-9dca-6e76f780e1a7, reason: Instance 6e066235-99c3-4bde-972a-345faccf9fa6 could not be found. {{(pid=62824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11482}} [ 2627.223950] env[62824]: INFO nova.compute.manager [-] [instance: 6e066235-99c3-4bde-972a-345faccf9fa6] Took 1.43 seconds to deallocate network for instance. [ 2627.281175] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] There are 26 instances to clean {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11754}} [ 2627.281329] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 8bdfe1c4-1c23-4360-9bb0-90dfc9123d79] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2627.730777] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fad9d49c-8bf2-4972-aa05-dbdb2aa7fbd6 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2627.731195] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fad9d49c-8bf2-4972-aa05-dbdb2aa7fbd6 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2627.731307] env[62824]: DEBUG nova.objects.instance [None req-fad9d49c-8bf2-4972-aa05-dbdb2aa7fbd6 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lazy-loading 'resources' on Instance uuid 6e066235-99c3-4bde-972a-345faccf9fa6 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2627.784627] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: e524ce78-6a63-4950-b727-ff72f8df4424] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2628.088710] env[62824]: DEBUG nova.network.neutron [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Port 8be13d2b-3033-431e-92aa-3e05a180498f binding to destination host cpu-1 is already ACTIVE {{(pid=62824) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 2628.287153] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 80c57382-8f6f-4537-b210-70a902b02cb4] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2628.298325] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a23f60bf-700f-4436-a231-ce5965c8667e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2628.305735] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c18b417-b59a-47de-af68-f6a50c53d477 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2628.337014] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bf431d9-9454-4d0d-9e4d-6689e583f6bc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2628.344034] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d9976fc-34b9-4a2d-87e9-6321faf5d0cb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2628.356996] env[62824]: DEBUG nova.compute.provider_tree [None req-fad9d49c-8bf2-4972-aa05-dbdb2aa7fbd6 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2628.790340] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 9227cc56-d55d-44ba-a619-63533dd77b66] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2628.859850] env[62824]: DEBUG nova.scheduler.client.report [None req-fad9d49c-8bf2-4972-aa05-dbdb2aa7fbd6 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2629.110308] env[62824]: DEBUG oslo_concurrency.lockutils [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquiring lock "eb0a5669-7b2a-4d2f-9596-6486e8b78289-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2629.110572] env[62824]: DEBUG oslo_concurrency.lockutils [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lock "eb0a5669-7b2a-4d2f-9596-6486e8b78289-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2629.110751] env[62824]: DEBUG oslo_concurrency.lockutils [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lock "eb0a5669-7b2a-4d2f-9596-6486e8b78289-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2629.293697] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 4dff2a76-b0a2-438c-b52f-d6495948d041] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2629.364227] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fad9d49c-8bf2-4972-aa05-dbdb2aa7fbd6 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.633s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2629.382481] env[62824]: INFO nova.scheduler.client.report [None req-fad9d49c-8bf2-4972-aa05-dbdb2aa7fbd6 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Deleted allocations for instance 6e066235-99c3-4bde-972a-345faccf9fa6 [ 2629.796608] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: ad9f01dd-d719-41a2-971b-b12cd3e8bdf6] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2629.889635] env[62824]: DEBUG oslo_concurrency.lockutils [None req-fad9d49c-8bf2-4972-aa05-dbdb2aa7fbd6 tempest-ServerActionsTestJSON-1451234720 tempest-ServerActionsTestJSON-1451234720-project-member] Lock "6e066235-99c3-4bde-972a-345faccf9fa6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.942s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2630.146489] env[62824]: DEBUG oslo_concurrency.lockutils [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquiring lock "refresh_cache-eb0a5669-7b2a-4d2f-9596-6486e8b78289" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2630.146713] env[62824]: DEBUG oslo_concurrency.lockutils [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquired lock "refresh_cache-eb0a5669-7b2a-4d2f-9596-6486e8b78289" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2630.146906] env[62824]: DEBUG nova.network.neutron [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2630.300293] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: df621c9c-8a68-4082-8f30-9bf3dda26fb6] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2630.803056] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 7c47bb92-a58e-4537-a68a-289e2a5fc19a] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2630.854743] env[62824]: DEBUG nova.network.neutron [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Updating instance_info_cache with network_info: [{"id": "8be13d2b-3033-431e-92aa-3e05a180498f", "address": "fa:16:3e:9f:4d:c4", "network": {"id": "5233a794-3514-4912-a857-7e76a66b1e8f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-910685188-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.129", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0eb41836305847fca773afebdafa01db", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd98a25d-a7a9-4fb5-8fef-e8df4dbbbf11", "external-id": "nsx-vlan-transportzone-707", "segmentation_id": 707, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8be13d2b-30", "ovs_interfaceid": "8be13d2b-3033-431e-92aa-3e05a180498f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2631.306842] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: c70880c9-a69a-49b9-b028-189c04276623] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2631.358261] env[62824]: DEBUG oslo_concurrency.lockutils [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Releasing lock "refresh_cache-eb0a5669-7b2a-4d2f-9596-6486e8b78289" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2631.809712] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 21a7eed6-c146-4911-9a0a-a1e1644e8e9d] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2631.866673] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d42bdfc6-17fd-4d1f-ae6b-97da1e7be17e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2631.874172] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb7e6086-d966-4dca-b1eb-86a11eeb9553 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2632.313578] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: fb9cfb53-285a-4586-acf0-cb08d56b93e7] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2632.816837] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: bf6e681d-b773-4787-886a-c90586bc81bf] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2632.979121] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2f9ac4c-d881-47bb-9e65-e80cb47377e8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2633.001372] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0abdd611-c926-4e1a-be8e-70bd331fb37f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2633.008822] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Updating instance 'eb0a5669-7b2a-4d2f-9596-6486e8b78289' progress to 83 {{(pid=62824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 2633.321151] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: c480f680-61fd-496b-8cf7-fa50f580b10d] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2633.515132] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2633.515428] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b391d121-acad-4696-a07b-41148a42506f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2633.523180] env[62824]: DEBUG oslo_vmware.api [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2633.523180] env[62824]: value = "task-2147164" [ 2633.523180] env[62824]: _type = "Task" [ 2633.523180] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2633.530904] env[62824]: DEBUG oslo_vmware.api [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147164, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2633.826292] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 511ef5bb-1dca-4eae-a0f1-19d1d7a17371] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2634.034528] env[62824]: DEBUG oslo_vmware.api [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147164, 'name': PowerOnVM_Task, 'duration_secs': 0.366198} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2634.034528] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2634.034528] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-81e1c8d9-9871-43ad-876b-0c94705324d6 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Updating instance 'eb0a5669-7b2a-4d2f-9596-6486e8b78289' progress to 100 {{(pid=62824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 2634.328082] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 1db9ff97-e59c-43f6-984e-6405d8e993a4] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2634.833996] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 68cc2c5d-77e3-479f-8a8c-b62ae2dd84e3] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2635.338038] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 3d294ee9-92f6-4115-95db-9d4b66562b56] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2635.400984] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Acquiring lock "ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2635.401275] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Lock "ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2635.841190] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: e1574266-e53b-4169-9db6-c66b84895edb] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2635.906552] env[62824]: DEBUG nova.compute.manager [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] [instance: ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2635.970965] env[62824]: DEBUG nova.network.neutron [None req-408db55b-4735-4b89-9ed8-b42c918d904f tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Port 8be13d2b-3033-431e-92aa-3e05a180498f binding to destination host cpu-1 is already ACTIVE {{(pid=62824) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 2635.971329] env[62824]: DEBUG oslo_concurrency.lockutils [None req-408db55b-4735-4b89-9ed8-b42c918d904f tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquiring lock "refresh_cache-eb0a5669-7b2a-4d2f-9596-6486e8b78289" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2635.971410] env[62824]: DEBUG oslo_concurrency.lockutils [None req-408db55b-4735-4b89-9ed8-b42c918d904f tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquired lock "refresh_cache-eb0a5669-7b2a-4d2f-9596-6486e8b78289" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2635.971551] env[62824]: DEBUG nova.network.neutron [None req-408db55b-4735-4b89-9ed8-b42c918d904f tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2636.344473] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 1805a6e7-48f2-47b4-b097-0644ba8217fa] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2636.431802] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2636.432071] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2636.433577] env[62824]: INFO nova.compute.claims [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] [instance: ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2636.704634] env[62824]: DEBUG nova.network.neutron [None req-408db55b-4735-4b89-9ed8-b42c918d904f tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Updating instance_info_cache with network_info: [{"id": "8be13d2b-3033-431e-92aa-3e05a180498f", "address": "fa:16:3e:9f:4d:c4", "network": {"id": "5233a794-3514-4912-a857-7e76a66b1e8f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-910685188-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.129", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0eb41836305847fca773afebdafa01db", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd98a25d-a7a9-4fb5-8fef-e8df4dbbbf11", "external-id": "nsx-vlan-transportzone-707", "segmentation_id": 707, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8be13d2b-30", "ovs_interfaceid": "8be13d2b-3033-431e-92aa-3e05a180498f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2636.847708] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 71e9733b-c84b-4501-8faf-a487f089b498] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2637.207272] env[62824]: DEBUG oslo_concurrency.lockutils [None req-408db55b-4735-4b89-9ed8-b42c918d904f tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Releasing lock "refresh_cache-eb0a5669-7b2a-4d2f-9596-6486e8b78289" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2637.350961] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: c1b3b2eb-3ba5-4523-94b6-0f9a64f2b412] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2637.512248] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e8396f2-9358-446e-b62b-195f0f8e227c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2637.520296] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-903f515b-ab46-4200-b558-e31702b1b407 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2637.552021] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a27d4575-5bd0-4921-bf9a-118f2dc56684 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2637.559783] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c87a8d8b-c70e-4a1e-966e-58059ececac7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2637.573647] env[62824]: DEBUG nova.compute.provider_tree [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2637.711552] env[62824]: DEBUG nova.compute.manager [None req-408db55b-4735-4b89-9ed8-b42c918d904f tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=62824) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:901}} [ 2637.854179] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 5b3df31f-15fe-473f-992c-ddb272661c53] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2638.077479] env[62824]: DEBUG nova.scheduler.client.report [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2638.358036] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: ca42fef0-1e90-4ab6-9d60-8ef7e4997884] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2638.584873] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.152s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2638.585494] env[62824]: DEBUG nova.compute.manager [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] [instance: ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2638.816188] env[62824]: DEBUG oslo_concurrency.lockutils [None req-408db55b-4735-4b89-9ed8-b42c918d904f tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2638.816480] env[62824]: DEBUG oslo_concurrency.lockutils [None req-408db55b-4735-4b89-9ed8-b42c918d904f tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2638.860848] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: a6cd2032-de60-4f78-bf1e-79801d049df0] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2639.090847] env[62824]: DEBUG nova.compute.utils [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2639.092297] env[62824]: DEBUG nova.compute.manager [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] [instance: ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2639.092496] env[62824]: DEBUG nova.network.neutron [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] [instance: ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2639.135538] env[62824]: DEBUG nova.policy [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '885ebc5fd0c044118fa432faaccf428a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3bf74635b57b411f81702b48643c23c0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2639.319723] env[62824]: DEBUG nova.objects.instance [None req-408db55b-4735-4b89-9ed8-b42c918d904f tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lazy-loading 'migration_context' on Instance uuid eb0a5669-7b2a-4d2f-9596-6486e8b78289 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2639.364144] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 6d80ec06-8559-4964-8577-a2512aa366ed] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2639.397348] env[62824]: DEBUG nova.network.neutron [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] [instance: ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea] Successfully created port: 8673757c-c371-4557-b6d6-cada5e787721 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2639.596137] env[62824]: DEBUG nova.compute.manager [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] [instance: ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2639.868785] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: b3003c4b-ae5a-48df-8c12-a915a76253f4] Instance has had 0 of 5 cleanup attempts {{(pid=62824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11758}} [ 2639.892203] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b8fac93-1229-4f82-9281-806e67e1944c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2639.901824] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-626b7793-aebe-4ad2-b105-f8b67939b5c6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2639.931378] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f1d7f02-9551-447b-b6ce-a6ffdbb31874 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2639.938509] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f502e454-396b-4e82-86dd-6fe4a84bca9c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2639.951307] env[62824]: DEBUG nova.compute.provider_tree [None req-408db55b-4735-4b89-9ed8-b42c918d904f tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2640.454701] env[62824]: DEBUG nova.scheduler.client.report [None req-408db55b-4735-4b89-9ed8-b42c918d904f tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2640.607199] env[62824]: DEBUG nova.compute.manager [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] [instance: ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2640.634607] env[62824]: DEBUG nova.virt.hardware [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2640.634786] env[62824]: DEBUG nova.virt.hardware [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2640.634945] env[62824]: DEBUG nova.virt.hardware [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2640.635145] env[62824]: DEBUG nova.virt.hardware [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2640.635295] env[62824]: DEBUG nova.virt.hardware [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2640.635448] env[62824]: DEBUG nova.virt.hardware [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2640.635680] env[62824]: DEBUG nova.virt.hardware [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2640.635844] env[62824]: DEBUG nova.virt.hardware [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2640.636023] env[62824]: DEBUG nova.virt.hardware [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2640.636190] env[62824]: DEBUG nova.virt.hardware [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2640.636365] env[62824]: DEBUG nova.virt.hardware [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2640.637241] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b520644b-e092-474e-bb7d-2cf807e3859c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2640.645389] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b3d3b5f-1d3f-43ad-8e02-ce33004db82c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2641.466068] env[62824]: DEBUG oslo_concurrency.lockutils [None req-408db55b-4735-4b89-9ed8-b42c918d904f tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.649s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2642.372840] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2642.373108] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2642.373255] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62824) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11064}} [ 2643.005623] env[62824]: INFO nova.compute.manager [None req-408db55b-4735-4b89-9ed8-b42c918d904f tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Swapping old allocation on dict_keys(['bf95157b-a274-42de-9ccf-9851128a44a1']) held by migration 696bd587-625b-4376-993d-f1bc5f435717 for instance [ 2643.027141] env[62824]: DEBUG nova.scheduler.client.report [None req-408db55b-4735-4b89-9ed8-b42c918d904f tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Overwriting current allocation {'allocations': {'bf95157b-a274-42de-9ccf-9851128a44a1': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 175}}, 'project_id': '0eb41836305847fca773afebdafa01db', 'user_id': 'e6ecc84f6587474f89419598f5c44dbb', 'consumer_generation': 1} on consumer eb0a5669-7b2a-4d2f-9596-6486e8b78289 {{(pid=62824) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2036}} [ 2643.101238] env[62824]: DEBUG oslo_concurrency.lockutils [None req-408db55b-4735-4b89-9ed8-b42c918d904f tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquiring lock "refresh_cache-eb0a5669-7b2a-4d2f-9596-6486e8b78289" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2643.101429] env[62824]: DEBUG oslo_concurrency.lockutils [None req-408db55b-4735-4b89-9ed8-b42c918d904f tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquired lock "refresh_cache-eb0a5669-7b2a-4d2f-9596-6486e8b78289" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2643.101609] env[62824]: DEBUG nova.network.neutron [None req-408db55b-4735-4b89-9ed8-b42c918d904f tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2643.908443] env[62824]: DEBUG nova.compute.manager [req-158ceada-c733-4db0-a4de-4fa02a1d3311 req-55583ef6-9112-459a-8e59-b864fc72aa27 service nova] [instance: ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea] Received event network-vif-plugged-8673757c-c371-4557-b6d6-cada5e787721 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2643.908798] env[62824]: DEBUG oslo_concurrency.lockutils [req-158ceada-c733-4db0-a4de-4fa02a1d3311 req-55583ef6-9112-459a-8e59-b864fc72aa27 service nova] Acquiring lock "ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2643.909149] env[62824]: DEBUG oslo_concurrency.lockutils [req-158ceada-c733-4db0-a4de-4fa02a1d3311 req-55583ef6-9112-459a-8e59-b864fc72aa27 service nova] Lock "ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2643.909542] env[62824]: DEBUG oslo_concurrency.lockutils [req-158ceada-c733-4db0-a4de-4fa02a1d3311 req-55583ef6-9112-459a-8e59-b864fc72aa27 service nova] Lock "ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2643.909612] env[62824]: DEBUG nova.compute.manager [req-158ceada-c733-4db0-a4de-4fa02a1d3311 req-55583ef6-9112-459a-8e59-b864fc72aa27 service nova] [instance: ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea] No waiting events found dispatching network-vif-plugged-8673757c-c371-4557-b6d6-cada5e787721 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2643.909749] env[62824]: WARNING nova.compute.manager [req-158ceada-c733-4db0-a4de-4fa02a1d3311 req-55583ef6-9112-459a-8e59-b864fc72aa27 service nova] [instance: ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea] Received unexpected event network-vif-plugged-8673757c-c371-4557-b6d6-cada5e787721 for instance with vm_state building and task_state spawning. [ 2643.942150] env[62824]: DEBUG nova.network.neutron [None req-408db55b-4735-4b89-9ed8-b42c918d904f tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Updating instance_info_cache with network_info: [{"id": "8be13d2b-3033-431e-92aa-3e05a180498f", "address": "fa:16:3e:9f:4d:c4", "network": {"id": "5233a794-3514-4912-a857-7e76a66b1e8f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-910685188-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.129", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0eb41836305847fca773afebdafa01db", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd98a25d-a7a9-4fb5-8fef-e8df4dbbbf11", "external-id": "nsx-vlan-transportzone-707", "segmentation_id": 707, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8be13d2b-30", "ovs_interfaceid": "8be13d2b-3033-431e-92aa-3e05a180498f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2644.016125] env[62824]: DEBUG nova.network.neutron [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] [instance: ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea] Successfully updated port: 8673757c-c371-4557-b6d6-cada5e787721 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2644.445958] env[62824]: DEBUG oslo_concurrency.lockutils [None req-408db55b-4735-4b89-9ed8-b42c918d904f tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Releasing lock "refresh_cache-eb0a5669-7b2a-4d2f-9596-6486e8b78289" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2644.447037] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ad678b8-7d79-40f7-ae0a-89a7801f9872 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2644.453833] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-befa54a5-b5f2-43ca-851b-d78d7728636e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2644.521047] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Acquiring lock "refresh_cache-ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2644.521212] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Acquired lock "refresh_cache-ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2644.521367] env[62824]: DEBUG nova.network.neutron [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] [instance: ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2644.773384] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2645.055681] env[62824]: DEBUG nova.network.neutron [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] [instance: ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2645.173456] env[62824]: DEBUG nova.network.neutron [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] [instance: ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea] Updating instance_info_cache with network_info: [{"id": "8673757c-c371-4557-b6d6-cada5e787721", "address": "fa:16:3e:9e:d8:9e", "network": {"id": "7b872f96-dd22-412c-85d0-3ffabd72afa9", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-1761582996-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3bf74635b57b411f81702b48643c23c0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "da623279-b6f6-4570-8b15-a332120b8b60", "external-id": "nsx-vlan-transportzone-733", "segmentation_id": 733, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8673757c-c3", "ovs_interfaceid": "8673757c-c371-4557-b6d6-cada5e787721", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2645.532411] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-408db55b-4735-4b89-9ed8-b42c918d904f tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2645.532650] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1c6cf162-948f-4c8c-bb7b-f1f46efe8055 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2645.540340] env[62824]: DEBUG oslo_vmware.api [None req-408db55b-4735-4b89-9ed8-b42c918d904f tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2645.540340] env[62824]: value = "task-2147165" [ 2645.540340] env[62824]: _type = "Task" [ 2645.540340] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2645.548316] env[62824]: DEBUG oslo_vmware.api [None req-408db55b-4735-4b89-9ed8-b42c918d904f tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147165, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2645.676330] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Releasing lock "refresh_cache-ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2645.676699] env[62824]: DEBUG nova.compute.manager [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] [instance: ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea] Instance network_info: |[{"id": "8673757c-c371-4557-b6d6-cada5e787721", "address": "fa:16:3e:9e:d8:9e", "network": {"id": "7b872f96-dd22-412c-85d0-3ffabd72afa9", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-1761582996-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3bf74635b57b411f81702b48643c23c0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "da623279-b6f6-4570-8b15-a332120b8b60", "external-id": "nsx-vlan-transportzone-733", "segmentation_id": 733, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8673757c-c3", "ovs_interfaceid": "8673757c-c371-4557-b6d6-cada5e787721", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2645.677158] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] [instance: ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9e:d8:9e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'da623279-b6f6-4570-8b15-a332120b8b60', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8673757c-c371-4557-b6d6-cada5e787721', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2645.684794] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Creating folder: Project (3bf74635b57b411f81702b48643c23c0). Parent ref: group-v438503. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 2645.685128] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9a491216-bcfd-4324-ab76-55d8e335a1d1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2645.696869] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Created folder: Project (3bf74635b57b411f81702b48643c23c0) in parent group-v438503. [ 2645.697071] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Creating folder: Instances. Parent ref: group-v438796. {{(pid=62824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 2645.697315] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3c5aaf7d-8f46-49c6-8653-db5b5214066e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2645.707480] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Created folder: Instances in parent group-v438796. [ 2645.707727] env[62824]: DEBUG oslo.service.loopingcall [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2645.708024] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2645.708221] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-710d3236-a313-48ad-a377-00dc9973b46b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2645.727482] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2645.727482] env[62824]: value = "task-2147168" [ 2645.727482] env[62824]: _type = "Task" [ 2645.727482] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2645.735251] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2147168, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2645.940445] env[62824]: DEBUG nova.compute.manager [req-6a549b00-54bb-4586-89b3-8cb1f62e149b req-dc42ba9b-bf77-4541-9bee-c2b3377de910 service nova] [instance: ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea] Received event network-changed-8673757c-c371-4557-b6d6-cada5e787721 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2645.940671] env[62824]: DEBUG nova.compute.manager [req-6a549b00-54bb-4586-89b3-8cb1f62e149b req-dc42ba9b-bf77-4541-9bee-c2b3377de910 service nova] [instance: ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea] Refreshing instance network info cache due to event network-changed-8673757c-c371-4557-b6d6-cada5e787721. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2645.940934] env[62824]: DEBUG oslo_concurrency.lockutils [req-6a549b00-54bb-4586-89b3-8cb1f62e149b req-dc42ba9b-bf77-4541-9bee-c2b3377de910 service nova] Acquiring lock "refresh_cache-ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2645.941093] env[62824]: DEBUG oslo_concurrency.lockutils [req-6a549b00-54bb-4586-89b3-8cb1f62e149b req-dc42ba9b-bf77-4541-9bee-c2b3377de910 service nova] Acquired lock "refresh_cache-ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2645.941302] env[62824]: DEBUG nova.network.neutron [req-6a549b00-54bb-4586-89b3-8cb1f62e149b req-dc42ba9b-bf77-4541-9bee-c2b3377de910 service nova] [instance: ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea] Refreshing network info cache for port 8673757c-c371-4557-b6d6-cada5e787721 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2646.052367] env[62824]: DEBUG oslo_vmware.api [None req-408db55b-4735-4b89-9ed8-b42c918d904f tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147165, 'name': PowerOffVM_Task, 'duration_secs': 0.211479} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2646.052560] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-408db55b-4735-4b89-9ed8-b42c918d904f tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2646.053167] env[62824]: DEBUG nova.virt.hardware [None req-408db55b-4735-4b89-9ed8-b42c918d904f tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2646.053384] env[62824]: DEBUG nova.virt.hardware [None req-408db55b-4735-4b89-9ed8-b42c918d904f tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2646.053531] env[62824]: DEBUG nova.virt.hardware [None req-408db55b-4735-4b89-9ed8-b42c918d904f tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2646.053713] env[62824]: DEBUG nova.virt.hardware [None req-408db55b-4735-4b89-9ed8-b42c918d904f tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2646.053880] env[62824]: DEBUG nova.virt.hardware [None req-408db55b-4735-4b89-9ed8-b42c918d904f tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2646.054094] env[62824]: DEBUG nova.virt.hardware [None req-408db55b-4735-4b89-9ed8-b42c918d904f tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2646.054224] env[62824]: DEBUG nova.virt.hardware [None req-408db55b-4735-4b89-9ed8-b42c918d904f tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2646.054384] env[62824]: DEBUG nova.virt.hardware [None req-408db55b-4735-4b89-9ed8-b42c918d904f tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2646.054560] env[62824]: DEBUG nova.virt.hardware [None req-408db55b-4735-4b89-9ed8-b42c918d904f tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2646.054755] env[62824]: DEBUG nova.virt.hardware [None req-408db55b-4735-4b89-9ed8-b42c918d904f tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2646.054938] env[62824]: DEBUG nova.virt.hardware [None req-408db55b-4735-4b89-9ed8-b42c918d904f tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2646.059906] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f1c41782-7ec8-4385-bcaa-68dda98089c9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2646.076219] env[62824]: DEBUG oslo_vmware.api [None req-408db55b-4735-4b89-9ed8-b42c918d904f tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2646.076219] env[62824]: value = "task-2147169" [ 2646.076219] env[62824]: _type = "Task" [ 2646.076219] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2646.084656] env[62824]: DEBUG oslo_vmware.api [None req-408db55b-4735-4b89-9ed8-b42c918d904f tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147169, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2646.236872] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2147168, 'name': CreateVM_Task, 'duration_secs': 0.421563} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2646.237055] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2646.237711] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2646.237877] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2646.238233] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2646.238521] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b17ef3ad-3928-4dd5-82da-38f1acd58c97 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2646.243067] env[62824]: DEBUG oslo_vmware.api [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Waiting for the task: (returnval){ [ 2646.243067] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]529b7e03-e24a-ad50-b369-a3b22bd9ef69" [ 2646.243067] env[62824]: _type = "Task" [ 2646.243067] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2646.250171] env[62824]: DEBUG oslo_vmware.api [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]529b7e03-e24a-ad50-b369-a3b22bd9ef69, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2646.586023] env[62824]: DEBUG oslo_vmware.api [None req-408db55b-4735-4b89-9ed8-b42c918d904f tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147169, 'name': ReconfigVM_Task, 'duration_secs': 0.174712} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2646.586812] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8a3af23-42bc-436b-9dc0-39c6e10bc0c2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2646.608444] env[62824]: DEBUG nova.virt.hardware [None req-408db55b-4735-4b89-9ed8-b42c918d904f tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2646.608669] env[62824]: DEBUG nova.virt.hardware [None req-408db55b-4735-4b89-9ed8-b42c918d904f tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2646.608827] env[62824]: DEBUG nova.virt.hardware [None req-408db55b-4735-4b89-9ed8-b42c918d904f tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2646.609071] env[62824]: DEBUG nova.virt.hardware [None req-408db55b-4735-4b89-9ed8-b42c918d904f tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2646.609173] env[62824]: DEBUG nova.virt.hardware [None req-408db55b-4735-4b89-9ed8-b42c918d904f tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2646.609323] env[62824]: DEBUG nova.virt.hardware [None req-408db55b-4735-4b89-9ed8-b42c918d904f tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2646.609523] env[62824]: DEBUG nova.virt.hardware [None req-408db55b-4735-4b89-9ed8-b42c918d904f tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2646.609677] env[62824]: DEBUG nova.virt.hardware [None req-408db55b-4735-4b89-9ed8-b42c918d904f tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2646.609841] env[62824]: DEBUG nova.virt.hardware [None req-408db55b-4735-4b89-9ed8-b42c918d904f tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2646.609997] env[62824]: DEBUG nova.virt.hardware [None req-408db55b-4735-4b89-9ed8-b42c918d904f tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2646.610184] env[62824]: DEBUG nova.virt.hardware [None req-408db55b-4735-4b89-9ed8-b42c918d904f tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2646.612955] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9f7aeee9-55bc-45a6-8ac2-bca8f65e49f0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2646.618780] env[62824]: DEBUG oslo_vmware.api [None req-408db55b-4735-4b89-9ed8-b42c918d904f tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2646.618780] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52c4d329-7684-3018-d3ea-082335b071a4" [ 2646.618780] env[62824]: _type = "Task" [ 2646.618780] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2646.626226] env[62824]: DEBUG oslo_vmware.api [None req-408db55b-4735-4b89-9ed8-b42c918d904f tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52c4d329-7684-3018-d3ea-082335b071a4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2646.647666] env[62824]: DEBUG nova.network.neutron [req-6a549b00-54bb-4586-89b3-8cb1f62e149b req-dc42ba9b-bf77-4541-9bee-c2b3377de910 service nova] [instance: ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea] Updated VIF entry in instance network info cache for port 8673757c-c371-4557-b6d6-cada5e787721. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2646.648054] env[62824]: DEBUG nova.network.neutron [req-6a549b00-54bb-4586-89b3-8cb1f62e149b req-dc42ba9b-bf77-4541-9bee-c2b3377de910 service nova] [instance: ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea] Updating instance_info_cache with network_info: [{"id": "8673757c-c371-4557-b6d6-cada5e787721", "address": "fa:16:3e:9e:d8:9e", "network": {"id": "7b872f96-dd22-412c-85d0-3ffabd72afa9", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-1761582996-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3bf74635b57b411f81702b48643c23c0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "da623279-b6f6-4570-8b15-a332120b8b60", "external-id": "nsx-vlan-transportzone-733", "segmentation_id": 733, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8673757c-c3", "ovs_interfaceid": "8673757c-c371-4557-b6d6-cada5e787721", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2646.753246] env[62824]: DEBUG oslo_vmware.api [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]529b7e03-e24a-ad50-b369-a3b22bd9ef69, 'name': SearchDatastore_Task, 'duration_secs': 0.016736} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2646.753531] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2646.753705] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] [instance: ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2646.753931] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2646.754086] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2646.754265] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2646.754533] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9ef8d68c-3ce6-4308-a29c-fc7620125716 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2646.762186] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2646.762354] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2646.763046] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-020a72e0-36a4-4f74-bffe-f604d56603a9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2646.767898] env[62824]: DEBUG oslo_vmware.api [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Waiting for the task: (returnval){ [ 2646.767898] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]522389af-ebaa-b688-9ac7-dfd7047a03a4" [ 2646.767898] env[62824]: _type = "Task" [ 2646.767898] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2646.775450] env[62824]: DEBUG oslo_vmware.api [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]522389af-ebaa-b688-9ac7-dfd7047a03a4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2647.129511] env[62824]: DEBUG oslo_vmware.api [None req-408db55b-4735-4b89-9ed8-b42c918d904f tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52c4d329-7684-3018-d3ea-082335b071a4, 'name': SearchDatastore_Task, 'duration_secs': 0.007305} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2647.134833] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-408db55b-4735-4b89-9ed8-b42c918d904f tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Reconfiguring VM instance instance-00000078 to detach disk 2000 {{(pid=62824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 2647.135110] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6f6db130-bea8-4d56-b3be-d3406cea62fb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2647.150181] env[62824]: DEBUG oslo_concurrency.lockutils [req-6a549b00-54bb-4586-89b3-8cb1f62e149b req-dc42ba9b-bf77-4541-9bee-c2b3377de910 service nova] Releasing lock "refresh_cache-ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2647.152851] env[62824]: DEBUG oslo_vmware.api [None req-408db55b-4735-4b89-9ed8-b42c918d904f tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2647.152851] env[62824]: value = "task-2147170" [ 2647.152851] env[62824]: _type = "Task" [ 2647.152851] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2647.160763] env[62824]: DEBUG oslo_vmware.api [None req-408db55b-4735-4b89-9ed8-b42c918d904f tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147170, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2647.278071] env[62824]: DEBUG oslo_vmware.api [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]522389af-ebaa-b688-9ac7-dfd7047a03a4, 'name': SearchDatastore_Task, 'duration_secs': 0.008032} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2647.278819] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3bf18677-a7db-4576-91e2-6538fcf42041 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2647.283732] env[62824]: DEBUG oslo_vmware.api [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Waiting for the task: (returnval){ [ 2647.283732] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5227a1c4-36ee-d887-3a3c-56bfdd632817" [ 2647.283732] env[62824]: _type = "Task" [ 2647.283732] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2647.291015] env[62824]: DEBUG oslo_vmware.api [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5227a1c4-36ee-d887-3a3c-56bfdd632817, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2647.662476] env[62824]: DEBUG oslo_vmware.api [None req-408db55b-4735-4b89-9ed8-b42c918d904f tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147170, 'name': ReconfigVM_Task, 'duration_secs': 0.22433} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2647.663557] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-408db55b-4735-4b89-9ed8-b42c918d904f tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Reconfigured VM instance instance-00000078 to detach disk 2000 {{(pid=62824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 2647.664390] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-183b8384-ce68-4019-ad67-7eec5ec11e55 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2647.688527] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-408db55b-4735-4b89-9ed8-b42c918d904f tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Reconfiguring VM instance instance-00000078 to attach disk [datastore2] eb0a5669-7b2a-4d2f-9596-6486e8b78289/eb0a5669-7b2a-4d2f-9596-6486e8b78289.vmdk or device None with type thin {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2647.689123] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5ba6c3d3-7b6d-486d-9726-30560211f2b6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2647.707043] env[62824]: DEBUG oslo_vmware.api [None req-408db55b-4735-4b89-9ed8-b42c918d904f tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2647.707043] env[62824]: value = "task-2147171" [ 2647.707043] env[62824]: _type = "Task" [ 2647.707043] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2647.714481] env[62824]: DEBUG oslo_vmware.api [None req-408db55b-4735-4b89-9ed8-b42c918d904f tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147171, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2647.793624] env[62824]: DEBUG oslo_vmware.api [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5227a1c4-36ee-d887-3a3c-56bfdd632817, 'name': SearchDatastore_Task, 'duration_secs': 0.00911} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2647.793868] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2647.794135] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea/ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2647.794375] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a6fcad2b-4501-4744-9e2f-107c1e9a8e7d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2647.800524] env[62824]: DEBUG oslo_vmware.api [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Waiting for the task: (returnval){ [ 2647.800524] env[62824]: value = "task-2147172" [ 2647.800524] env[62824]: _type = "Task" [ 2647.800524] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2647.808781] env[62824]: DEBUG oslo_vmware.api [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Task: {'id': task-2147172, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2648.217098] env[62824]: DEBUG oslo_vmware.api [None req-408db55b-4735-4b89-9ed8-b42c918d904f tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147171, 'name': ReconfigVM_Task, 'duration_secs': 0.292375} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2648.217568] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-408db55b-4735-4b89-9ed8-b42c918d904f tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Reconfigured VM instance instance-00000078 to attach disk [datastore2] eb0a5669-7b2a-4d2f-9596-6486e8b78289/eb0a5669-7b2a-4d2f-9596-6486e8b78289.vmdk or device None with type thin {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2648.218303] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64bc5a3c-0932-4f49-8931-8f09537e3d21 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2648.239288] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b07779ad-f04a-46e4-a92f-22effa8f4856 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2648.260122] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a718e3f-e72d-491e-b8ed-e380ec7bf1f5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2648.280503] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63667b74-a2ea-4031-928b-3bc499c7d3a1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2648.287146] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-408db55b-4735-4b89-9ed8-b42c918d904f tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2648.287367] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e694dfc3-7e20-428f-bf64-d0bd6837255f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2648.293289] env[62824]: DEBUG oslo_vmware.api [None req-408db55b-4735-4b89-9ed8-b42c918d904f tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2648.293289] env[62824]: value = "task-2147173" [ 2648.293289] env[62824]: _type = "Task" [ 2648.293289] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2648.300299] env[62824]: DEBUG oslo_vmware.api [None req-408db55b-4735-4b89-9ed8-b42c918d904f tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147173, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2648.308353] env[62824]: DEBUG oslo_vmware.api [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Task: {'id': task-2147172, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.442196} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2648.308584] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea/ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2648.308786] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] [instance: ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2648.308996] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0c5901d4-7a52-4a6b-89bb-0b594218dd76 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2648.314797] env[62824]: DEBUG oslo_vmware.api [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Waiting for the task: (returnval){ [ 2648.314797] env[62824]: value = "task-2147174" [ 2648.314797] env[62824]: _type = "Task" [ 2648.314797] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2648.321519] env[62824]: DEBUG oslo_vmware.api [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Task: {'id': task-2147174, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2648.802983] env[62824]: DEBUG oslo_vmware.api [None req-408db55b-4735-4b89-9ed8-b42c918d904f tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147173, 'name': PowerOnVM_Task, 'duration_secs': 0.390638} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2648.803296] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-408db55b-4735-4b89-9ed8-b42c918d904f tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2648.823443] env[62824]: DEBUG oslo_vmware.api [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Task: {'id': task-2147174, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063182} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2648.823696] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] [instance: ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2648.824460] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-955eabd1-c5af-4d81-b7c0-9d3cd26c9976 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2648.846474] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] [instance: ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea] Reconfiguring VM instance instance-0000007a to attach disk [datastore2] ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea/ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2648.846918] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0c3ca8ea-5d39-49cc-96d8-0b31e0f46772 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2648.867470] env[62824]: DEBUG oslo_vmware.api [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Waiting for the task: (returnval){ [ 2648.867470] env[62824]: value = "task-2147175" [ 2648.867470] env[62824]: _type = "Task" [ 2648.867470] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2648.874447] env[62824]: DEBUG oslo_vmware.api [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Task: {'id': task-2147175, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2649.377491] env[62824]: DEBUG oslo_vmware.api [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Task: {'id': task-2147175, 'name': ReconfigVM_Task, 'duration_secs': 0.307489} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2649.377869] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] [instance: ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea] Reconfigured VM instance instance-0000007a to attach disk [datastore2] ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea/ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2649.378594] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e316e6fa-6a96-4ed6-bee1-54f72ccfa267 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2649.385294] env[62824]: DEBUG oslo_vmware.api [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Waiting for the task: (returnval){ [ 2649.385294] env[62824]: value = "task-2147176" [ 2649.385294] env[62824]: _type = "Task" [ 2649.385294] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2649.392810] env[62824]: DEBUG oslo_vmware.api [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Task: {'id': task-2147176, 'name': Rename_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2649.851704] env[62824]: INFO nova.compute.manager [None req-408db55b-4735-4b89-9ed8-b42c918d904f tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Updating instance to original state: 'active' [ 2649.895320] env[62824]: DEBUG oslo_vmware.api [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Task: {'id': task-2147176, 'name': Rename_Task, 'duration_secs': 0.137459} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2649.895579] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] [instance: ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2649.895839] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6c5340fa-bcd3-47a6-9a01-71a5a82907a6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2649.901783] env[62824]: DEBUG oslo_vmware.api [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Waiting for the task: (returnval){ [ 2649.901783] env[62824]: value = "task-2147177" [ 2649.901783] env[62824]: _type = "Task" [ 2649.901783] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2649.908998] env[62824]: DEBUG oslo_vmware.api [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Task: {'id': task-2147177, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2650.412413] env[62824]: DEBUG oslo_vmware.api [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Task: {'id': task-2147177, 'name': PowerOnVM_Task, 'duration_secs': 0.492032} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2650.412803] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] [instance: ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2650.412803] env[62824]: INFO nova.compute.manager [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] [instance: ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea] Took 9.81 seconds to spawn the instance on the hypervisor. [ 2650.412982] env[62824]: DEBUG nova.compute.manager [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] [instance: ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2650.413739] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b56fec9-e744-4f4b-bdaf-e2b8de269644 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2650.929709] env[62824]: INFO nova.compute.manager [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] [instance: ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea] Took 14.52 seconds to build instance. [ 2651.432351] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d57a4a30-679e-4ca0-8475-2b0b002eb74c tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Lock "ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.031s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2651.751561] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f7727d22-7778-45a2-8962-ed25f053b7e7 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquiring lock "eb0a5669-7b2a-4d2f-9596-6486e8b78289" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2651.751828] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f7727d22-7778-45a2-8962-ed25f053b7e7 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lock "eb0a5669-7b2a-4d2f-9596-6486e8b78289" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2651.752084] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f7727d22-7778-45a2-8962-ed25f053b7e7 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquiring lock "eb0a5669-7b2a-4d2f-9596-6486e8b78289-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2651.752275] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f7727d22-7778-45a2-8962-ed25f053b7e7 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lock "eb0a5669-7b2a-4d2f-9596-6486e8b78289-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2651.752446] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f7727d22-7778-45a2-8962-ed25f053b7e7 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lock "eb0a5669-7b2a-4d2f-9596-6486e8b78289-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2651.754633] env[62824]: INFO nova.compute.manager [None req-f7727d22-7778-45a2-8962-ed25f053b7e7 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Terminating instance [ 2652.258851] env[62824]: DEBUG nova.compute.manager [None req-f7727d22-7778-45a2-8962-ed25f053b7e7 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2652.259146] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7727d22-7778-45a2-8962-ed25f053b7e7 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2652.259444] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a0d15baf-ea18-4017-b6a4-0b4a0749c36c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2652.268090] env[62824]: DEBUG oslo_vmware.api [None req-f7727d22-7778-45a2-8962-ed25f053b7e7 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2652.268090] env[62824]: value = "task-2147178" [ 2652.268090] env[62824]: _type = "Task" [ 2652.268090] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2652.277097] env[62824]: DEBUG oslo_vmware.api [None req-f7727d22-7778-45a2-8962-ed25f053b7e7 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147178, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2652.582165] env[62824]: DEBUG oslo_concurrency.lockutils [None req-07008d42-520c-4b2d-a672-6799211fc48e tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Acquiring lock "ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2652.582561] env[62824]: DEBUG oslo_concurrency.lockutils [None req-07008d42-520c-4b2d-a672-6799211fc48e tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Lock "ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2652.582690] env[62824]: DEBUG oslo_concurrency.lockutils [None req-07008d42-520c-4b2d-a672-6799211fc48e tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Acquiring lock "ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2652.582876] env[62824]: DEBUG oslo_concurrency.lockutils [None req-07008d42-520c-4b2d-a672-6799211fc48e tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Lock "ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2652.583065] env[62824]: DEBUG oslo_concurrency.lockutils [None req-07008d42-520c-4b2d-a672-6799211fc48e tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Lock "ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2652.585355] env[62824]: INFO nova.compute.manager [None req-07008d42-520c-4b2d-a672-6799211fc48e tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] [instance: ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea] Terminating instance [ 2652.777884] env[62824]: DEBUG oslo_vmware.api [None req-f7727d22-7778-45a2-8962-ed25f053b7e7 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147178, 'name': PowerOffVM_Task, 'duration_secs': 0.192475} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2652.778113] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7727d22-7778-45a2-8962-ed25f053b7e7 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2652.778313] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-f7727d22-7778-45a2-8962-ed25f053b7e7 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Volume detach. Driver type: vmdk {{(pid=62824) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 2652.778503] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-f7727d22-7778-45a2-8962-ed25f053b7e7 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-438794', 'volume_id': 'a0dcd1b7-ffe3-46f6-adc9-597e1c5487d6', 'name': 'volume-a0dcd1b7-ffe3-46f6-adc9-597e1c5487d6', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attaching', 'instance': 'eb0a5669-7b2a-4d2f-9596-6486e8b78289', 'attached_at': '2025-01-16T15:04:08.000000', 'detached_at': '', 'volume_id': 'a0dcd1b7-ffe3-46f6-adc9-597e1c5487d6', 'serial': 'a0dcd1b7-ffe3-46f6-adc9-597e1c5487d6'} {{(pid=62824) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 2652.779261] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa10f88b-7a4f-46f0-af70-b50349057baf {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2652.799861] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-908039ca-0b05-4033-ab70-2c3550900643 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2652.806217] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf389381-ae78-45fb-82f6-730293f617be {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2652.825705] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8af06a01-5298-424b-878b-c62dfc821cb1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2652.839926] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-f7727d22-7778-45a2-8962-ed25f053b7e7 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] The volume has not been displaced from its original location: [datastore1] volume-a0dcd1b7-ffe3-46f6-adc9-597e1c5487d6/volume-a0dcd1b7-ffe3-46f6-adc9-597e1c5487d6.vmdk. No consolidation needed. {{(pid=62824) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 2652.845181] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-f7727d22-7778-45a2-8962-ed25f053b7e7 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Reconfiguring VM instance instance-00000078 to detach disk 2001 {{(pid=62824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 2652.845479] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-acf82baa-0efe-4608-a19f-69f1566a4404 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2652.863847] env[62824]: DEBUG oslo_vmware.api [None req-f7727d22-7778-45a2-8962-ed25f053b7e7 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2652.863847] env[62824]: value = "task-2147179" [ 2652.863847] env[62824]: _type = "Task" [ 2652.863847] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2652.873804] env[62824]: DEBUG oslo_vmware.api [None req-f7727d22-7778-45a2-8962-ed25f053b7e7 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147179, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2653.088923] env[62824]: DEBUG nova.compute.manager [None req-07008d42-520c-4b2d-a672-6799211fc48e tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] [instance: ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2653.089206] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-07008d42-520c-4b2d-a672-6799211fc48e tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] [instance: ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2653.090119] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e6ca1c6-3ca4-476a-8787-23b36f62e0eb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2653.097886] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-07008d42-520c-4b2d-a672-6799211fc48e tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] [instance: ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2653.098168] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d1489ab8-a2f9-4669-b3d9-abbdc54fa162 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2653.104844] env[62824]: DEBUG oslo_vmware.api [None req-07008d42-520c-4b2d-a672-6799211fc48e tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Waiting for the task: (returnval){ [ 2653.104844] env[62824]: value = "task-2147180" [ 2653.104844] env[62824]: _type = "Task" [ 2653.104844] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2653.112672] env[62824]: DEBUG oslo_vmware.api [None req-07008d42-520c-4b2d-a672-6799211fc48e tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Task: {'id': task-2147180, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2653.373648] env[62824]: DEBUG oslo_vmware.api [None req-f7727d22-7778-45a2-8962-ed25f053b7e7 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147179, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2653.614724] env[62824]: DEBUG oslo_vmware.api [None req-07008d42-520c-4b2d-a672-6799211fc48e tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Task: {'id': task-2147180, 'name': PowerOffVM_Task, 'duration_secs': 0.31565} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2653.615051] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-07008d42-520c-4b2d-a672-6799211fc48e tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] [instance: ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2653.615177] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-07008d42-520c-4b2d-a672-6799211fc48e tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] [instance: ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2653.615419] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f3989c44-1581-449a-ac21-ab9ef0d2605f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2653.875313] env[62824]: DEBUG oslo_vmware.api [None req-f7727d22-7778-45a2-8962-ed25f053b7e7 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147179, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2654.234991] env[62824]: DEBUG oslo_concurrency.lockutils [None req-825167ae-501a-49a4-8587-d3696e19d907 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Acquiring lock "c41daca7-21fc-44fa-9957-d26767e06658" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2654.235258] env[62824]: DEBUG oslo_concurrency.lockutils [None req-825167ae-501a-49a4-8587-d3696e19d907 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Lock "c41daca7-21fc-44fa-9957-d26767e06658" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2654.375425] env[62824]: DEBUG oslo_vmware.api [None req-f7727d22-7778-45a2-8962-ed25f053b7e7 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147179, 'name': ReconfigVM_Task, 'duration_secs': 1.209212} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2654.375787] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-f7727d22-7778-45a2-8962-ed25f053b7e7 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Reconfigured VM instance instance-00000078 to detach disk 2001 {{(pid=62824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 2654.380250] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bd9daafe-21d1-4bf2-a271-7afeac44f102 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2654.395213] env[62824]: DEBUG oslo_vmware.api [None req-f7727d22-7778-45a2-8962-ed25f053b7e7 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2654.395213] env[62824]: value = "task-2147182" [ 2654.395213] env[62824]: _type = "Task" [ 2654.395213] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2654.402996] env[62824]: DEBUG oslo_vmware.api [None req-f7727d22-7778-45a2-8962-ed25f053b7e7 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147182, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2654.738614] env[62824]: DEBUG nova.compute.utils [None req-825167ae-501a-49a4-8587-d3696e19d907 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2654.904356] env[62824]: DEBUG oslo_vmware.api [None req-f7727d22-7778-45a2-8962-ed25f053b7e7 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147182, 'name': ReconfigVM_Task, 'duration_secs': 0.129079} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2654.904604] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-f7727d22-7778-45a2-8962-ed25f053b7e7 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-438794', 'volume_id': 'a0dcd1b7-ffe3-46f6-adc9-597e1c5487d6', 'name': 'volume-a0dcd1b7-ffe3-46f6-adc9-597e1c5487d6', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attaching', 'instance': 'eb0a5669-7b2a-4d2f-9596-6486e8b78289', 'attached_at': '2025-01-16T15:04:08.000000', 'detached_at': '', 'volume_id': 'a0dcd1b7-ffe3-46f6-adc9-597e1c5487d6', 'serial': 'a0dcd1b7-ffe3-46f6-adc9-597e1c5487d6'} {{(pid=62824) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 2654.904936] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-f7727d22-7778-45a2-8962-ed25f053b7e7 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2654.905689] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9b2602c-46be-4260-9866-9fcd7c0d3cf4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2654.911766] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-f7727d22-7778-45a2-8962-ed25f053b7e7 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2654.911973] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-65642d81-b3b2-4859-a8b5-85ceed79e8f3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2655.241818] env[62824]: DEBUG oslo_concurrency.lockutils [None req-825167ae-501a-49a4-8587-d3696e19d907 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Lock "c41daca7-21fc-44fa-9957-d26767e06658" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2656.379773] env[62824]: DEBUG oslo_concurrency.lockutils [None req-825167ae-501a-49a4-8587-d3696e19d907 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Acquiring lock "c41daca7-21fc-44fa-9957-d26767e06658" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2656.380154] env[62824]: DEBUG oslo_concurrency.lockutils [None req-825167ae-501a-49a4-8587-d3696e19d907 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Lock "c41daca7-21fc-44fa-9957-d26767e06658" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2656.380618] env[62824]: INFO nova.compute.manager [None req-825167ae-501a-49a4-8587-d3696e19d907 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c41daca7-21fc-44fa-9957-d26767e06658] Attaching volume ebf6dd97-00c0-41e6-88e2-256b4cb6b0d6 to /dev/sdb [ 2656.410249] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c400db1c-fca2-4cbb-8700-d13f7cde89cf {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2656.417166] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0e852fc-58aa-4f9d-b108-7509fdeab294 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2656.431876] env[62824]: DEBUG nova.virt.block_device [None req-825167ae-501a-49a4-8587-d3696e19d907 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c41daca7-21fc-44fa-9957-d26767e06658] Updating existing volume attachment record: 021c7e0c-1ffc-4310-8a13-d2171137f185 {{(pid=62824) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 2660.420052] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-f7727d22-7778-45a2-8962-ed25f053b7e7 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2660.420052] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-f7727d22-7778-45a2-8962-ed25f053b7e7 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Deleting contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2660.420052] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-f7727d22-7778-45a2-8962-ed25f053b7e7 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Deleting the datastore file [datastore2] eb0a5669-7b2a-4d2f-9596-6486e8b78289 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2660.420518] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1c6cb43e-1ac6-4e7a-aec9-9ec558835bc6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2660.422412] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-07008d42-520c-4b2d-a672-6799211fc48e tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] [instance: ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2660.422593] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-07008d42-520c-4b2d-a672-6799211fc48e tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] [instance: ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea] Deleting contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2660.422793] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-07008d42-520c-4b2d-a672-6799211fc48e tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Deleting the datastore file [datastore2] ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2660.423032] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d5e9e869-4c09-4922-8c9b-48158dfb42c1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2660.429418] env[62824]: DEBUG oslo_vmware.api [None req-f7727d22-7778-45a2-8962-ed25f053b7e7 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2660.429418] env[62824]: value = "task-2147186" [ 2660.429418] env[62824]: _type = "Task" [ 2660.429418] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2660.430536] env[62824]: DEBUG oslo_vmware.api [None req-07008d42-520c-4b2d-a672-6799211fc48e tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Waiting for the task: (returnval){ [ 2660.430536] env[62824]: value = "task-2147187" [ 2660.430536] env[62824]: _type = "Task" [ 2660.430536] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2660.442292] env[62824]: DEBUG oslo_vmware.api [None req-f7727d22-7778-45a2-8962-ed25f053b7e7 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147186, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2660.947773] env[62824]: DEBUG oslo_vmware.api [None req-f7727d22-7778-45a2-8962-ed25f053b7e7 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147186, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.140627} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2660.948027] env[62824]: DEBUG oslo_vmware.api [None req-07008d42-520c-4b2d-a672-6799211fc48e tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Task: {'id': task-2147187, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.139228} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2660.948255] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-f7727d22-7778-45a2-8962-ed25f053b7e7 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2660.948431] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-f7727d22-7778-45a2-8962-ed25f053b7e7 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Deleted contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2660.948606] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-f7727d22-7778-45a2-8962-ed25f053b7e7 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2660.948774] env[62824]: INFO nova.compute.manager [None req-f7727d22-7778-45a2-8962-ed25f053b7e7 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Took 8.69 seconds to destroy the instance on the hypervisor. [ 2660.949016] env[62824]: DEBUG oslo.service.loopingcall [None req-f7727d22-7778-45a2-8962-ed25f053b7e7 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2660.949207] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-07008d42-520c-4b2d-a672-6799211fc48e tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2660.949367] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-07008d42-520c-4b2d-a672-6799211fc48e tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] [instance: ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea] Deleted contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2660.949531] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-07008d42-520c-4b2d-a672-6799211fc48e tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] [instance: ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2660.949699] env[62824]: INFO nova.compute.manager [None req-07008d42-520c-4b2d-a672-6799211fc48e tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] [instance: ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea] Took 7.86 seconds to destroy the instance on the hypervisor. [ 2660.949906] env[62824]: DEBUG oslo.service.loopingcall [None req-07008d42-520c-4b2d-a672-6799211fc48e tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2660.950107] env[62824]: DEBUG nova.compute.manager [-] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2660.950203] env[62824]: DEBUG nova.network.neutron [-] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2660.951710] env[62824]: DEBUG nova.compute.manager [-] [instance: ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2660.951814] env[62824]: DEBUG nova.network.neutron [-] [instance: ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2660.973958] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-825167ae-501a-49a4-8587-d3696e19d907 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c41daca7-21fc-44fa-9957-d26767e06658] Volume attach. Driver type: vmdk {{(pid=62824) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 2660.974206] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-825167ae-501a-49a4-8587-d3696e19d907 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c41daca7-21fc-44fa-9957-d26767e06658] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-438799', 'volume_id': 'ebf6dd97-00c0-41e6-88e2-256b4cb6b0d6', 'name': 'volume-ebf6dd97-00c0-41e6-88e2-256b4cb6b0d6', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'c41daca7-21fc-44fa-9957-d26767e06658', 'attached_at': '', 'detached_at': '', 'volume_id': 'ebf6dd97-00c0-41e6-88e2-256b4cb6b0d6', 'serial': 'ebf6dd97-00c0-41e6-88e2-256b4cb6b0d6'} {{(pid=62824) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 2660.975064] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d094c1a-50ff-4ce0-a785-63166fe12106 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2660.990928] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-749dc540-abc8-40dc-b49e-e5c6a1dab036 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2661.015570] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-825167ae-501a-49a4-8587-d3696e19d907 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c41daca7-21fc-44fa-9957-d26767e06658] Reconfiguring VM instance instance-00000079 to attach disk [datastore1] volume-ebf6dd97-00c0-41e6-88e2-256b4cb6b0d6/volume-ebf6dd97-00c0-41e6-88e2-256b4cb6b0d6.vmdk or device None with type thin {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2661.015869] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a148962a-26ec-4843-b432-bc304e27a4ef {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2661.035289] env[62824]: DEBUG oslo_vmware.api [None req-825167ae-501a-49a4-8587-d3696e19d907 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Waiting for the task: (returnval){ [ 2661.035289] env[62824]: value = "task-2147188" [ 2661.035289] env[62824]: _type = "Task" [ 2661.035289] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2661.042311] env[62824]: DEBUG oslo_vmware.api [None req-825167ae-501a-49a4-8587-d3696e19d907 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147188, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2661.386590] env[62824]: DEBUG nova.compute.manager [req-7d66db5b-0b69-4089-b415-37702151d06c req-72560eb6-8b4c-4e20-9c7f-443ad69b8a1d service nova] [instance: ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea] Received event network-vif-deleted-8673757c-c371-4557-b6d6-cada5e787721 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2661.386779] env[62824]: INFO nova.compute.manager [req-7d66db5b-0b69-4089-b415-37702151d06c req-72560eb6-8b4c-4e20-9c7f-443ad69b8a1d service nova] [instance: ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea] Neutron deleted interface 8673757c-c371-4557-b6d6-cada5e787721; detaching it from the instance and deleting it from the info cache [ 2661.386945] env[62824]: DEBUG nova.network.neutron [req-7d66db5b-0b69-4089-b415-37702151d06c req-72560eb6-8b4c-4e20-9c7f-443ad69b8a1d service nova] [instance: ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2661.545139] env[62824]: DEBUG oslo_vmware.api [None req-825167ae-501a-49a4-8587-d3696e19d907 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147188, 'name': ReconfigVM_Task, 'duration_secs': 0.362721} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2661.545483] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-825167ae-501a-49a4-8587-d3696e19d907 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c41daca7-21fc-44fa-9957-d26767e06658] Reconfigured VM instance instance-00000079 to attach disk [datastore1] volume-ebf6dd97-00c0-41e6-88e2-256b4cb6b0d6/volume-ebf6dd97-00c0-41e6-88e2-256b4cb6b0d6.vmdk or device None with type thin {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2661.550084] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8cd8baf8-31c4-4d7f-abd2-fa3f26a55931 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2661.565725] env[62824]: DEBUG oslo_vmware.api [None req-825167ae-501a-49a4-8587-d3696e19d907 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Waiting for the task: (returnval){ [ 2661.565725] env[62824]: value = "task-2147189" [ 2661.565725] env[62824]: _type = "Task" [ 2661.565725] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2661.574837] env[62824]: DEBUG oslo_vmware.api [None req-825167ae-501a-49a4-8587-d3696e19d907 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147189, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2661.868568] env[62824]: DEBUG nova.network.neutron [-] [instance: ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2661.889980] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d292fe10-e7bf-4be1-8526-a7fd380cf540 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2661.899774] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cef0009-2a55-4f3c-a77f-e16a488bfff4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2661.924866] env[62824]: DEBUG nova.compute.manager [req-7d66db5b-0b69-4089-b415-37702151d06c req-72560eb6-8b4c-4e20-9c7f-443ad69b8a1d service nova] [instance: ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea] Detach interface failed, port_id=8673757c-c371-4557-b6d6-cada5e787721, reason: Instance ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea could not be found. {{(pid=62824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11482}} [ 2662.029580] env[62824]: DEBUG nova.network.neutron [-] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2662.077010] env[62824]: DEBUG oslo_vmware.api [None req-825167ae-501a-49a4-8587-d3696e19d907 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147189, 'name': ReconfigVM_Task, 'duration_secs': 0.133261} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2662.077352] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-825167ae-501a-49a4-8587-d3696e19d907 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c41daca7-21fc-44fa-9957-d26767e06658] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-438799', 'volume_id': 'ebf6dd97-00c0-41e6-88e2-256b4cb6b0d6', 'name': 'volume-ebf6dd97-00c0-41e6-88e2-256b4cb6b0d6', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'c41daca7-21fc-44fa-9957-d26767e06658', 'attached_at': '', 'detached_at': '', 'volume_id': 'ebf6dd97-00c0-41e6-88e2-256b4cb6b0d6', 'serial': 'ebf6dd97-00c0-41e6-88e2-256b4cb6b0d6'} {{(pid=62824) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 2662.371206] env[62824]: INFO nova.compute.manager [-] [instance: ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea] Took 1.42 seconds to deallocate network for instance. [ 2662.532340] env[62824]: INFO nova.compute.manager [-] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Took 1.58 seconds to deallocate network for instance. [ 2662.878642] env[62824]: DEBUG oslo_concurrency.lockutils [None req-07008d42-520c-4b2d-a672-6799211fc48e tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2662.878971] env[62824]: DEBUG oslo_concurrency.lockutils [None req-07008d42-520c-4b2d-a672-6799211fc48e tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2662.879170] env[62824]: DEBUG nova.objects.instance [None req-07008d42-520c-4b2d-a672-6799211fc48e tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Lazy-loading 'resources' on Instance uuid ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2663.073572] env[62824]: INFO nova.compute.manager [None req-f7727d22-7778-45a2-8962-ed25f053b7e7 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Took 0.54 seconds to detach 1 volumes for instance. [ 2663.109903] env[62824]: DEBUG nova.objects.instance [None req-825167ae-501a-49a4-8587-d3696e19d907 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Lazy-loading 'flavor' on Instance uuid c41daca7-21fc-44fa-9957-d26767e06658 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2663.415976] env[62824]: DEBUG nova.compute.manager [req-a7615300-3a44-446a-9301-17da0ad27cc7 req-59d313f9-5545-443a-927f-7fe84ff99e1b service nova] [instance: eb0a5669-7b2a-4d2f-9596-6486e8b78289] Received event network-vif-deleted-8be13d2b-3033-431e-92aa-3e05a180498f {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2663.434094] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce98eec0-0bfb-4730-a30d-7033b8de375f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2663.441448] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-132f7a83-b678-4e7a-8d0a-c27839f518ef {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2663.471435] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21b2ed7e-f61e-40c8-9fa4-d60c851703d0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2663.478630] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c305b49-ba13-4dcf-8a45-9ff9012db839 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2663.491452] env[62824]: DEBUG nova.compute.provider_tree [None req-07008d42-520c-4b2d-a672-6799211fc48e tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2663.579230] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f7727d22-7778-45a2-8962-ed25f053b7e7 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2663.615231] env[62824]: DEBUG oslo_concurrency.lockutils [None req-825167ae-501a-49a4-8587-d3696e19d907 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Lock "c41daca7-21fc-44fa-9957-d26767e06658" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.235s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2663.787690] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1e8c74cb-2c43-481a-8b24-7c39e0f87374 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Acquiring lock "c41daca7-21fc-44fa-9957-d26767e06658" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2663.787944] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1e8c74cb-2c43-481a-8b24-7c39e0f87374 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Lock "c41daca7-21fc-44fa-9957-d26767e06658" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2663.994957] env[62824]: DEBUG nova.scheduler.client.report [None req-07008d42-520c-4b2d-a672-6799211fc48e tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2664.291072] env[62824]: INFO nova.compute.manager [None req-1e8c74cb-2c43-481a-8b24-7c39e0f87374 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c41daca7-21fc-44fa-9957-d26767e06658] Detaching volume ebf6dd97-00c0-41e6-88e2-256b4cb6b0d6 [ 2664.326178] env[62824]: INFO nova.virt.block_device [None req-1e8c74cb-2c43-481a-8b24-7c39e0f87374 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c41daca7-21fc-44fa-9957-d26767e06658] Attempting to driver detach volume ebf6dd97-00c0-41e6-88e2-256b4cb6b0d6 from mountpoint /dev/sdb [ 2664.326433] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-1e8c74cb-2c43-481a-8b24-7c39e0f87374 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c41daca7-21fc-44fa-9957-d26767e06658] Volume detach. Driver type: vmdk {{(pid=62824) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 2664.326620] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-1e8c74cb-2c43-481a-8b24-7c39e0f87374 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c41daca7-21fc-44fa-9957-d26767e06658] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-438799', 'volume_id': 'ebf6dd97-00c0-41e6-88e2-256b4cb6b0d6', 'name': 'volume-ebf6dd97-00c0-41e6-88e2-256b4cb6b0d6', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'c41daca7-21fc-44fa-9957-d26767e06658', 'attached_at': '', 'detached_at': '', 'volume_id': 'ebf6dd97-00c0-41e6-88e2-256b4cb6b0d6', 'serial': 'ebf6dd97-00c0-41e6-88e2-256b4cb6b0d6'} {{(pid=62824) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 2664.327540] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d3d99ef-1b37-4129-9283-92eab2e04df8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2664.351368] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43c574de-1a0c-434a-8d6c-9e43ac8d053e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2664.358741] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1ab6fc1-eb39-4e66-9c19-cf3862cce447 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2664.378517] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a3d2811-8ec6-412e-9b77-7ba36d88b980 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2664.392875] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-1e8c74cb-2c43-481a-8b24-7c39e0f87374 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] The volume has not been displaced from its original location: [datastore1] volume-ebf6dd97-00c0-41e6-88e2-256b4cb6b0d6/volume-ebf6dd97-00c0-41e6-88e2-256b4cb6b0d6.vmdk. No consolidation needed. {{(pid=62824) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 2664.398199] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-1e8c74cb-2c43-481a-8b24-7c39e0f87374 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c41daca7-21fc-44fa-9957-d26767e06658] Reconfiguring VM instance instance-00000079 to detach disk 2001 {{(pid=62824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 2664.398468] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-33964d40-10d3-4209-ab69-a28ab493365a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2664.417553] env[62824]: DEBUG oslo_vmware.api [None req-1e8c74cb-2c43-481a-8b24-7c39e0f87374 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Waiting for the task: (returnval){ [ 2664.417553] env[62824]: value = "task-2147190" [ 2664.417553] env[62824]: _type = "Task" [ 2664.417553] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2664.425221] env[62824]: DEBUG oslo_vmware.api [None req-1e8c74cb-2c43-481a-8b24-7c39e0f87374 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147190, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2664.499758] env[62824]: DEBUG oslo_concurrency.lockutils [None req-07008d42-520c-4b2d-a672-6799211fc48e tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.621s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2664.502174] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f7727d22-7778-45a2-8962-ed25f053b7e7 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.923s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2664.502387] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f7727d22-7778-45a2-8962-ed25f053b7e7 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2664.521347] env[62824]: INFO nova.scheduler.client.report [None req-f7727d22-7778-45a2-8962-ed25f053b7e7 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Deleted allocations for instance eb0a5669-7b2a-4d2f-9596-6486e8b78289 [ 2664.523970] env[62824]: INFO nova.scheduler.client.report [None req-07008d42-520c-4b2d-a672-6799211fc48e tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Deleted allocations for instance ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea [ 2664.927182] env[62824]: DEBUG oslo_vmware.api [None req-1e8c74cb-2c43-481a-8b24-7c39e0f87374 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147190, 'name': ReconfigVM_Task, 'duration_secs': 0.202463} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2664.927465] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-1e8c74cb-2c43-481a-8b24-7c39e0f87374 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c41daca7-21fc-44fa-9957-d26767e06658] Reconfigured VM instance instance-00000079 to detach disk 2001 {{(pid=62824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 2664.932031] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6091d067-7c4b-476c-af44-13fba901e109 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2664.946229] env[62824]: DEBUG oslo_vmware.api [None req-1e8c74cb-2c43-481a-8b24-7c39e0f87374 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Waiting for the task: (returnval){ [ 2664.946229] env[62824]: value = "task-2147191" [ 2664.946229] env[62824]: _type = "Task" [ 2664.946229] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2664.953440] env[62824]: DEBUG oslo_vmware.api [None req-1e8c74cb-2c43-481a-8b24-7c39e0f87374 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147191, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2665.036125] env[62824]: DEBUG oslo_concurrency.lockutils [None req-f7727d22-7778-45a2-8962-ed25f053b7e7 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lock "eb0a5669-7b2a-4d2f-9596-6486e8b78289" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.284s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2665.037171] env[62824]: DEBUG oslo_concurrency.lockutils [None req-07008d42-520c-4b2d-a672-6799211fc48e tempest-ServerTagsTestJSON-1417683153 tempest-ServerTagsTestJSON-1417683153-project-member] Lock "ac63ebb4-0c01-44fd-ae2e-9fa9bf4ed3ea" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.455s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2665.456441] env[62824]: DEBUG oslo_vmware.api [None req-1e8c74cb-2c43-481a-8b24-7c39e0f87374 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147191, 'name': ReconfigVM_Task, 'duration_secs': 0.123241} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2665.456754] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-1e8c74cb-2c43-481a-8b24-7c39e0f87374 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c41daca7-21fc-44fa-9957-d26767e06658] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-438799', 'volume_id': 'ebf6dd97-00c0-41e6-88e2-256b4cb6b0d6', 'name': 'volume-ebf6dd97-00c0-41e6-88e2-256b4cb6b0d6', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'c41daca7-21fc-44fa-9957-d26767e06658', 'attached_at': '', 'detached_at': '', 'volume_id': 'ebf6dd97-00c0-41e6-88e2-256b4cb6b0d6', 'serial': 'ebf6dd97-00c0-41e6-88e2-256b4cb6b0d6'} {{(pid=62824) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 2666.003673] env[62824]: DEBUG nova.objects.instance [None req-1e8c74cb-2c43-481a-8b24-7c39e0f87374 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Lazy-loading 'flavor' on Instance uuid c41daca7-21fc-44fa-9957-d26767e06658 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2666.714573] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquiring lock "02c41e06-f6e9-4b3c-bb29-bee33ab5e13c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2666.714913] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lock "02c41e06-f6e9-4b3c-bb29-bee33ab5e13c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2667.016542] env[62824]: DEBUG oslo_concurrency.lockutils [None req-1e8c74cb-2c43-481a-8b24-7c39e0f87374 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Lock "c41daca7-21fc-44fa-9957-d26767e06658" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.228s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2667.217458] env[62824]: DEBUG nova.compute.manager [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2667.738041] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2667.738629] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2667.739955] env[62824]: INFO nova.compute.claims [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2668.054747] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b49905e6-2572-4683-9c68-e8ac53d547a0 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Acquiring lock "c41daca7-21fc-44fa-9957-d26767e06658" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2668.055378] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b49905e6-2572-4683-9c68-e8ac53d547a0 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Lock "c41daca7-21fc-44fa-9957-d26767e06658" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2668.055378] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b49905e6-2572-4683-9c68-e8ac53d547a0 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Acquiring lock "c41daca7-21fc-44fa-9957-d26767e06658-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2668.055378] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b49905e6-2572-4683-9c68-e8ac53d547a0 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Lock "c41daca7-21fc-44fa-9957-d26767e06658-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2668.055603] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b49905e6-2572-4683-9c68-e8ac53d547a0 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Lock "c41daca7-21fc-44fa-9957-d26767e06658-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2668.057818] env[62824]: INFO nova.compute.manager [None req-b49905e6-2572-4683-9c68-e8ac53d547a0 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c41daca7-21fc-44fa-9957-d26767e06658] Terminating instance [ 2668.561094] env[62824]: DEBUG nova.compute.manager [None req-b49905e6-2572-4683-9c68-e8ac53d547a0 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c41daca7-21fc-44fa-9957-d26767e06658] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2668.561414] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-b49905e6-2572-4683-9c68-e8ac53d547a0 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c41daca7-21fc-44fa-9957-d26767e06658] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2668.562346] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1912b60f-5bab-4e02-bec0-5d12f55d11b6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2668.570416] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-b49905e6-2572-4683-9c68-e8ac53d547a0 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c41daca7-21fc-44fa-9957-d26767e06658] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2668.570658] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1d873729-a5af-41af-9975-1ac1ab62875a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2668.577126] env[62824]: DEBUG oslo_vmware.api [None req-b49905e6-2572-4683-9c68-e8ac53d547a0 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Waiting for the task: (returnval){ [ 2668.577126] env[62824]: value = "task-2147192" [ 2668.577126] env[62824]: _type = "Task" [ 2668.577126] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2668.585059] env[62824]: DEBUG oslo_vmware.api [None req-b49905e6-2572-4683-9c68-e8ac53d547a0 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147192, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2668.791227] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a243352e-0023-4870-9ed4-7ac4971d3164 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2668.799248] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b4a45c9-ae3b-45f0-a0be-2034a961e58c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2668.830650] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f90b739-85ae-4dfd-a32b-84395613851b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2668.838347] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6b71d49-e5b4-4e26-9514-a6c0f75f672f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2668.851615] env[62824]: DEBUG nova.compute.provider_tree [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2669.087093] env[62824]: DEBUG oslo_vmware.api [None req-b49905e6-2572-4683-9c68-e8ac53d547a0 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147192, 'name': PowerOffVM_Task, 'duration_secs': 0.173464} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2669.087376] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-b49905e6-2572-4683-9c68-e8ac53d547a0 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c41daca7-21fc-44fa-9957-d26767e06658] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2669.087545] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-b49905e6-2572-4683-9c68-e8ac53d547a0 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c41daca7-21fc-44fa-9957-d26767e06658] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2669.087782] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e8b84fd4-81af-4098-800e-d79c5ce73fea {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2669.355035] env[62824]: DEBUG nova.scheduler.client.report [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2669.777764] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Acquiring lock "c68974df-ba6e-4f13-b682-737dfe78f265" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2669.777999] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Lock "c68974df-ba6e-4f13-b682-737dfe78f265" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2669.859478] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.121s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2669.859967] env[62824]: DEBUG nova.compute.manager [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2670.280646] env[62824]: DEBUG nova.compute.manager [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Starting instance... {{(pid=62824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 2670.365042] env[62824]: DEBUG nova.compute.utils [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2670.366266] env[62824]: DEBUG nova.compute.manager [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2670.366431] env[62824]: DEBUG nova.network.neutron [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2670.413303] env[62824]: DEBUG nova.policy [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e6ecc84f6587474f89419598f5c44dbb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0eb41836305847fca773afebdafa01db', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2670.672892] env[62824]: DEBUG nova.network.neutron [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Successfully created port: 2c611b6e-39e9-428c-a425-1eebe84169dd {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2670.803697] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2670.803952] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2670.805586] env[62824]: INFO nova.compute.claims [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2670.871187] env[62824]: DEBUG nova.compute.manager [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2671.861086] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89dd869f-51b1-4807-9c36-170ff069c39b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2671.868360] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9398cca8-8b4c-452e-9868-9ed1d68a84e1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2671.897193] env[62824]: DEBUG nova.compute.manager [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2671.899642] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38faef2c-38bc-493b-9eb2-0331cad032d7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2671.906307] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8368450-0d00-42f1-a079-ce1b5ac56ec8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2671.919110] env[62824]: DEBUG nova.compute.provider_tree [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2671.927459] env[62824]: DEBUG nova.virt.hardware [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2671.927674] env[62824]: DEBUG nova.virt.hardware [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2671.927827] env[62824]: DEBUG nova.virt.hardware [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2671.928015] env[62824]: DEBUG nova.virt.hardware [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2671.928167] env[62824]: DEBUG nova.virt.hardware [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2671.928314] env[62824]: DEBUG nova.virt.hardware [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2671.928521] env[62824]: DEBUG nova.virt.hardware [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2671.928677] env[62824]: DEBUG nova.virt.hardware [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2671.928845] env[62824]: DEBUG nova.virt.hardware [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2671.929010] env[62824]: DEBUG nova.virt.hardware [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2671.929191] env[62824]: DEBUG nova.virt.hardware [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2671.929902] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4cfce8e-0e1b-4120-8abe-0d3e64ff84f3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2671.936837] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48e98aac-98f5-4a84-881b-7f0915c657f3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2672.423873] env[62824]: DEBUG nova.scheduler.client.report [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2672.929497] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.125s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2672.930074] env[62824]: DEBUG nova.compute.manager [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Start building networks asynchronously for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 2673.435527] env[62824]: DEBUG nova.compute.utils [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Using /dev/sd instead of None {{(pid=62824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 2673.436970] env[62824]: DEBUG nova.compute.manager [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Allocating IP information in the background. {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 2673.437124] env[62824]: DEBUG nova.network.neutron [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] allocate_for_instance() {{(pid=62824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 2673.474987] env[62824]: DEBUG nova.policy [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7b85d4201e7946e6b5b40a6ab8a22d61', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3240be4f32254f54aa87cb045e3d8f31', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62824) authorize /opt/stack/nova/nova/policy.py:192}} [ 2673.729031] env[62824]: DEBUG nova.network.neutron [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Successfully created port: 18202605-3a10-4c27-989e-74430907f582 {{(pid=62824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 2673.939744] env[62824]: DEBUG nova.compute.manager [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Start building block device mappings for instance. {{(pid=62824) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 2674.951280] env[62824]: DEBUG nova.compute.manager [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Start spawning the instance on the hypervisor. {{(pid=62824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 2674.975542] env[62824]: DEBUG nova.virt.hardware [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-01-16T14:27:04Z,direct_url=,disk_format='vmdk',id=9e2a7d30-212d-4ab8-9606-c5c6d52629e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='22c5c40c2e004babaa15896c89307303',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-01-16T14:27:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2674.975798] env[62824]: DEBUG nova.virt.hardware [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2674.975960] env[62824]: DEBUG nova.virt.hardware [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2674.976184] env[62824]: DEBUG nova.virt.hardware [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2674.976335] env[62824]: DEBUG nova.virt.hardware [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2674.976481] env[62824]: DEBUG nova.virt.hardware [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2674.976682] env[62824]: DEBUG nova.virt.hardware [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2674.976839] env[62824]: DEBUG nova.virt.hardware [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2674.977011] env[62824]: DEBUG nova.virt.hardware [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2674.977189] env[62824]: DEBUG nova.virt.hardware [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2674.977359] env[62824]: DEBUG nova.virt.hardware [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2674.978243] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de41cded-4831-4f98-8f9b-1fdf6c603507 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2674.986062] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb692588-da42-4508-a664-b587d3b1e0fe {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2676.082841] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-b49905e6-2572-4683-9c68-e8ac53d547a0 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c41daca7-21fc-44fa-9957-d26767e06658] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2676.083131] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-b49905e6-2572-4683-9c68-e8ac53d547a0 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c41daca7-21fc-44fa-9957-d26767e06658] Deleting contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2676.083294] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-b49905e6-2572-4683-9c68-e8ac53d547a0 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Deleting the datastore file [datastore1] c41daca7-21fc-44fa-9957-d26767e06658 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2676.083582] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-94c7c817-7631-4faf-b8a1-5bb35def36e7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2676.090714] env[62824]: DEBUG oslo_vmware.api [None req-b49905e6-2572-4683-9c68-e8ac53d547a0 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Waiting for the task: (returnval){ [ 2676.090714] env[62824]: value = "task-2147194" [ 2676.090714] env[62824]: _type = "Task" [ 2676.090714] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2676.098393] env[62824]: DEBUG oslo_vmware.api [None req-b49905e6-2572-4683-9c68-e8ac53d547a0 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147194, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2676.600365] env[62824]: DEBUG oslo_vmware.api [None req-b49905e6-2572-4683-9c68-e8ac53d547a0 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Task: {'id': task-2147194, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.134083} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2676.600614] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-b49905e6-2572-4683-9c68-e8ac53d547a0 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2676.600801] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-b49905e6-2572-4683-9c68-e8ac53d547a0 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c41daca7-21fc-44fa-9957-d26767e06658] Deleted contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2676.600981] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-b49905e6-2572-4683-9c68-e8ac53d547a0 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c41daca7-21fc-44fa-9957-d26767e06658] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2676.601175] env[62824]: INFO nova.compute.manager [None req-b49905e6-2572-4683-9c68-e8ac53d547a0 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] [instance: c41daca7-21fc-44fa-9957-d26767e06658] Took 8.04 seconds to destroy the instance on the hypervisor. [ 2676.601434] env[62824]: DEBUG oslo.service.loopingcall [None req-b49905e6-2572-4683-9c68-e8ac53d547a0 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2676.601595] env[62824]: DEBUG nova.compute.manager [-] [instance: c41daca7-21fc-44fa-9957-d26767e06658] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2676.601692] env[62824]: DEBUG nova.network.neutron [-] [instance: c41daca7-21fc-44fa-9957-d26767e06658] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2677.194040] env[62824]: DEBUG nova.compute.manager [req-5f32226d-bc05-40a8-a3c8-9a2743cb677f req-dbaf4c37-8356-43e0-b36f-2d76f342f750 service nova] [instance: c41daca7-21fc-44fa-9957-d26767e06658] Received event network-vif-deleted-5dd607e5-efbc-43a8-ae77-a884eb5dbec4 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2677.194395] env[62824]: INFO nova.compute.manager [req-5f32226d-bc05-40a8-a3c8-9a2743cb677f req-dbaf4c37-8356-43e0-b36f-2d76f342f750 service nova] [instance: c41daca7-21fc-44fa-9957-d26767e06658] Neutron deleted interface 5dd607e5-efbc-43a8-ae77-a884eb5dbec4; detaching it from the instance and deleting it from the info cache [ 2677.194437] env[62824]: DEBUG nova.network.neutron [req-5f32226d-bc05-40a8-a3c8-9a2743cb677f req-dbaf4c37-8356-43e0-b36f-2d76f342f750 service nova] [instance: c41daca7-21fc-44fa-9957-d26767e06658] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2677.435184] env[62824]: DEBUG nova.network.neutron [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Successfully updated port: 2c611b6e-39e9-428c-a425-1eebe84169dd {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2677.459682] env[62824]: DEBUG nova.network.neutron [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Successfully updated port: 18202605-3a10-4c27-989e-74430907f582 {{(pid=62824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 2677.664672] env[62824]: DEBUG nova.network.neutron [-] [instance: c41daca7-21fc-44fa-9957-d26767e06658] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2677.698255] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e1f31480-f5f3-43d2-b113-158e5b26439a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2677.707754] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7e16cd6-bb94-4ba8-88ed-8a5be2bd5857 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2677.732045] env[62824]: DEBUG nova.compute.manager [req-5f32226d-bc05-40a8-a3c8-9a2743cb677f req-dbaf4c37-8356-43e0-b36f-2d76f342f750 service nova] [instance: c41daca7-21fc-44fa-9957-d26767e06658] Detach interface failed, port_id=5dd607e5-efbc-43a8-ae77-a884eb5dbec4, reason: Instance c41daca7-21fc-44fa-9957-d26767e06658 could not be found. {{(pid=62824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11482}} [ 2677.942485] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquiring lock "refresh_cache-02c41e06-f6e9-4b3c-bb29-bee33ab5e13c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2677.942634] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquired lock "refresh_cache-02c41e06-f6e9-4b3c-bb29-bee33ab5e13c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2677.942786] env[62824]: DEBUG nova.network.neutron [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2677.962403] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Acquiring lock "refresh_cache-c68974df-ba6e-4f13-b682-737dfe78f265" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2677.962557] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Acquired lock "refresh_cache-c68974df-ba6e-4f13-b682-737dfe78f265" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2677.962704] env[62824]: DEBUG nova.network.neutron [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2678.168101] env[62824]: INFO nova.compute.manager [-] [instance: c41daca7-21fc-44fa-9957-d26767e06658] Took 1.57 seconds to deallocate network for instance. [ 2678.483657] env[62824]: DEBUG nova.network.neutron [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2678.508389] env[62824]: DEBUG nova.network.neutron [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Instance cache missing network info. {{(pid=62824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 2678.674809] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b49905e6-2572-4683-9c68-e8ac53d547a0 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2678.675118] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b49905e6-2572-4683-9c68-e8ac53d547a0 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2678.675348] env[62824]: DEBUG nova.objects.instance [None req-b49905e6-2572-4683-9c68-e8ac53d547a0 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Lazy-loading 'resources' on Instance uuid c41daca7-21fc-44fa-9957-d26767e06658 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2678.701379] env[62824]: DEBUG nova.network.neutron [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Updating instance_info_cache with network_info: [{"id": "2c611b6e-39e9-428c-a425-1eebe84169dd", "address": "fa:16:3e:d2:d7:a6", "network": {"id": "5233a794-3514-4912-a857-7e76a66b1e8f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-910685188-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0eb41836305847fca773afebdafa01db", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd98a25d-a7a9-4fb5-8fef-e8df4dbbbf11", "external-id": "nsx-vlan-transportzone-707", "segmentation_id": 707, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2c611b6e-39", "ovs_interfaceid": "2c611b6e-39e9-428c-a425-1eebe84169dd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2678.704805] env[62824]: DEBUG nova.network.neutron [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Updating instance_info_cache with network_info: [{"id": "18202605-3a10-4c27-989e-74430907f582", "address": "fa:16:3e:29:4a:15", "network": {"id": "db233b0c-27f3-4a76-ae9c-2b5889bf3d3b", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1825671448-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3240be4f32254f54aa87cb045e3d8f31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7041d198-66a3-40de-bf7d-cfc036e6ed69", "external-id": "nsx-vlan-transportzone-278", "segmentation_id": 278, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap18202605-3a", "ovs_interfaceid": "18202605-3a10-4c27-989e-74430907f582", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2679.203373] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Releasing lock "refresh_cache-02c41e06-f6e9-4b3c-bb29-bee33ab5e13c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2679.203678] env[62824]: DEBUG nova.compute.manager [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Instance network_info: |[{"id": "2c611b6e-39e9-428c-a425-1eebe84169dd", "address": "fa:16:3e:d2:d7:a6", "network": {"id": "5233a794-3514-4912-a857-7e76a66b1e8f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-910685188-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0eb41836305847fca773afebdafa01db", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd98a25d-a7a9-4fb5-8fef-e8df4dbbbf11", "external-id": "nsx-vlan-transportzone-707", "segmentation_id": 707, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2c611b6e-39", "ovs_interfaceid": "2c611b6e-39e9-428c-a425-1eebe84169dd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2679.204097] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d2:d7:a6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dd98a25d-a7a9-4fb5-8fef-e8df4dbbbf11', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2c611b6e-39e9-428c-a425-1eebe84169dd', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2679.211707] env[62824]: DEBUG oslo.service.loopingcall [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2679.212080] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Releasing lock "refresh_cache-c68974df-ba6e-4f13-b682-737dfe78f265" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2679.212332] env[62824]: DEBUG nova.compute.manager [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Instance network_info: |[{"id": "18202605-3a10-4c27-989e-74430907f582", "address": "fa:16:3e:29:4a:15", "network": {"id": "db233b0c-27f3-4a76-ae9c-2b5889bf3d3b", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1825671448-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3240be4f32254f54aa87cb045e3d8f31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7041d198-66a3-40de-bf7d-cfc036e6ed69", "external-id": "nsx-vlan-transportzone-278", "segmentation_id": 278, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap18202605-3a", "ovs_interfaceid": "18202605-3a10-4c27-989e-74430907f582", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 2679.212549] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2679.212886] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:29:4a:15', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7041d198-66a3-40de-bf7d-cfc036e6ed69', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '18202605-3a10-4c27-989e-74430907f582', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2679.219950] env[62824]: DEBUG oslo.service.loopingcall [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2679.220299] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-01c199ca-af65-4d97-b091-d903bb5834dd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2679.238682] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2679.241039] env[62824]: DEBUG nova.compute.manager [req-9242ce03-127a-452c-b365-a6450b2362ff req-f20d8d91-fe2a-4598-8d8d-44c35e27f2a2 service nova] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Received event network-vif-plugged-2c611b6e-39e9-428c-a425-1eebe84169dd {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2679.241039] env[62824]: DEBUG oslo_concurrency.lockutils [req-9242ce03-127a-452c-b365-a6450b2362ff req-f20d8d91-fe2a-4598-8d8d-44c35e27f2a2 service nova] Acquiring lock "02c41e06-f6e9-4b3c-bb29-bee33ab5e13c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2679.241039] env[62824]: DEBUG oslo_concurrency.lockutils [req-9242ce03-127a-452c-b365-a6450b2362ff req-f20d8d91-fe2a-4598-8d8d-44c35e27f2a2 service nova] Lock "02c41e06-f6e9-4b3c-bb29-bee33ab5e13c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2679.241039] env[62824]: DEBUG oslo_concurrency.lockutils [req-9242ce03-127a-452c-b365-a6450b2362ff req-f20d8d91-fe2a-4598-8d8d-44c35e27f2a2 service nova] Lock "02c41e06-f6e9-4b3c-bb29-bee33ab5e13c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2679.241039] env[62824]: DEBUG nova.compute.manager [req-9242ce03-127a-452c-b365-a6450b2362ff req-f20d8d91-fe2a-4598-8d8d-44c35e27f2a2 service nova] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] No waiting events found dispatching network-vif-plugged-2c611b6e-39e9-428c-a425-1eebe84169dd {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2679.241322] env[62824]: WARNING nova.compute.manager [req-9242ce03-127a-452c-b365-a6450b2362ff req-f20d8d91-fe2a-4598-8d8d-44c35e27f2a2 service nova] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Received unexpected event network-vif-plugged-2c611b6e-39e9-428c-a425-1eebe84169dd for instance with vm_state building and task_state spawning. [ 2679.241322] env[62824]: DEBUG nova.compute.manager [req-9242ce03-127a-452c-b365-a6450b2362ff req-f20d8d91-fe2a-4598-8d8d-44c35e27f2a2 service nova] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Received event network-vif-plugged-18202605-3a10-4c27-989e-74430907f582 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2679.241441] env[62824]: DEBUG oslo_concurrency.lockutils [req-9242ce03-127a-452c-b365-a6450b2362ff req-f20d8d91-fe2a-4598-8d8d-44c35e27f2a2 service nova] Acquiring lock "c68974df-ba6e-4f13-b682-737dfe78f265-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2679.241616] env[62824]: DEBUG oslo_concurrency.lockutils [req-9242ce03-127a-452c-b365-a6450b2362ff req-f20d8d91-fe2a-4598-8d8d-44c35e27f2a2 service nova] Lock "c68974df-ba6e-4f13-b682-737dfe78f265-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2679.241769] env[62824]: DEBUG oslo_concurrency.lockutils [req-9242ce03-127a-452c-b365-a6450b2362ff req-f20d8d91-fe2a-4598-8d8d-44c35e27f2a2 service nova] Lock "c68974df-ba6e-4f13-b682-737dfe78f265-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2679.241922] env[62824]: DEBUG nova.compute.manager [req-9242ce03-127a-452c-b365-a6450b2362ff req-f20d8d91-fe2a-4598-8d8d-44c35e27f2a2 service nova] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] No waiting events found dispatching network-vif-plugged-18202605-3a10-4c27-989e-74430907f582 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2679.242096] env[62824]: WARNING nova.compute.manager [req-9242ce03-127a-452c-b365-a6450b2362ff req-f20d8d91-fe2a-4598-8d8d-44c35e27f2a2 service nova] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Received unexpected event network-vif-plugged-18202605-3a10-4c27-989e-74430907f582 for instance with vm_state building and task_state spawning. [ 2679.242253] env[62824]: DEBUG nova.compute.manager [req-9242ce03-127a-452c-b365-a6450b2362ff req-f20d8d91-fe2a-4598-8d8d-44c35e27f2a2 service nova] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Received event network-changed-2c611b6e-39e9-428c-a425-1eebe84169dd {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2679.242400] env[62824]: DEBUG nova.compute.manager [req-9242ce03-127a-452c-b365-a6450b2362ff req-f20d8d91-fe2a-4598-8d8d-44c35e27f2a2 service nova] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Refreshing instance network info cache due to event network-changed-2c611b6e-39e9-428c-a425-1eebe84169dd. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2679.242574] env[62824]: DEBUG oslo_concurrency.lockutils [req-9242ce03-127a-452c-b365-a6450b2362ff req-f20d8d91-fe2a-4598-8d8d-44c35e27f2a2 service nova] Acquiring lock "refresh_cache-02c41e06-f6e9-4b3c-bb29-bee33ab5e13c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2679.242705] env[62824]: DEBUG oslo_concurrency.lockutils [req-9242ce03-127a-452c-b365-a6450b2362ff req-f20d8d91-fe2a-4598-8d8d-44c35e27f2a2 service nova] Acquired lock "refresh_cache-02c41e06-f6e9-4b3c-bb29-bee33ab5e13c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2679.242854] env[62824]: DEBUG nova.network.neutron [req-9242ce03-127a-452c-b365-a6450b2362ff req-f20d8d91-fe2a-4598-8d8d-44c35e27f2a2 service nova] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Refreshing network info cache for port 2c611b6e-39e9-428c-a425-1eebe84169dd {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2679.244046] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-413bf097-c0e2-4343-b30d-14b1906dbc44 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2679.267509] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2679.267509] env[62824]: value = "task-2147195" [ 2679.267509] env[62824]: _type = "Task" [ 2679.267509] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2679.272208] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2679.272208] env[62824]: value = "task-2147196" [ 2679.272208] env[62824]: _type = "Task" [ 2679.272208] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2679.280763] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager.update_available_resource {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2679.281910] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2147195, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2679.287405] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2147196, 'name': CreateVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2679.292668] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb410ac9-08a7-4d2f-9ec7-7afc4d0a3fed {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2679.301992] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7375fca-2a82-4c0a-8976-b23f738a300b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2679.336571] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9748cf3f-5334-4f5f-90cf-3db6c54dbdb4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2679.347504] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-264ed1a8-7f1c-4b7e-ab09-e529ef108b45 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2679.361811] env[62824]: DEBUG nova.compute.provider_tree [None req-b49905e6-2572-4683-9c68-e8ac53d547a0 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2679.473235] env[62824]: DEBUG nova.network.neutron [req-9242ce03-127a-452c-b365-a6450b2362ff req-f20d8d91-fe2a-4598-8d8d-44c35e27f2a2 service nova] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Updated VIF entry in instance network info cache for port 2c611b6e-39e9-428c-a425-1eebe84169dd. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2679.473586] env[62824]: DEBUG nova.network.neutron [req-9242ce03-127a-452c-b365-a6450b2362ff req-f20d8d91-fe2a-4598-8d8d-44c35e27f2a2 service nova] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Updating instance_info_cache with network_info: [{"id": "2c611b6e-39e9-428c-a425-1eebe84169dd", "address": "fa:16:3e:d2:d7:a6", "network": {"id": "5233a794-3514-4912-a857-7e76a66b1e8f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-910685188-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0eb41836305847fca773afebdafa01db", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd98a25d-a7a9-4fb5-8fef-e8df4dbbbf11", "external-id": "nsx-vlan-transportzone-707", "segmentation_id": 707, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2c611b6e-39", "ovs_interfaceid": "2c611b6e-39e9-428c-a425-1eebe84169dd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2679.780297] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2147195, 'name': CreateVM_Task, 'duration_secs': 0.439417} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2679.780779] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2679.781507] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2679.781705] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2679.782030] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2679.785129] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3208f693-a421-4b9a-a0af-48193c842436 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2679.787166] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2679.787695] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2147196, 'name': CreateVM_Task, 'duration_secs': 0.431699} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2679.787845] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2679.788689] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2679.790641] env[62824]: DEBUG oslo_vmware.api [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2679.790641] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]529f4f49-7e35-b9c8-881d-8553b81a2142" [ 2679.790641] env[62824]: _type = "Task" [ 2679.790641] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2679.797779] env[62824]: DEBUG oslo_vmware.api [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]529f4f49-7e35-b9c8-881d-8553b81a2142, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2679.866843] env[62824]: DEBUG nova.scheduler.client.report [None req-b49905e6-2572-4683-9c68-e8ac53d547a0 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2679.976125] env[62824]: DEBUG oslo_concurrency.lockutils [req-9242ce03-127a-452c-b365-a6450b2362ff req-f20d8d91-fe2a-4598-8d8d-44c35e27f2a2 service nova] Releasing lock "refresh_cache-02c41e06-f6e9-4b3c-bb29-bee33ab5e13c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2679.976411] env[62824]: DEBUG nova.compute.manager [req-9242ce03-127a-452c-b365-a6450b2362ff req-f20d8d91-fe2a-4598-8d8d-44c35e27f2a2 service nova] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Received event network-changed-18202605-3a10-4c27-989e-74430907f582 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2679.976576] env[62824]: DEBUG nova.compute.manager [req-9242ce03-127a-452c-b365-a6450b2362ff req-f20d8d91-fe2a-4598-8d8d-44c35e27f2a2 service nova] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Refreshing instance network info cache due to event network-changed-18202605-3a10-4c27-989e-74430907f582. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2679.976781] env[62824]: DEBUG oslo_concurrency.lockutils [req-9242ce03-127a-452c-b365-a6450b2362ff req-f20d8d91-fe2a-4598-8d8d-44c35e27f2a2 service nova] Acquiring lock "refresh_cache-c68974df-ba6e-4f13-b682-737dfe78f265" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2679.976921] env[62824]: DEBUG oslo_concurrency.lockutils [req-9242ce03-127a-452c-b365-a6450b2362ff req-f20d8d91-fe2a-4598-8d8d-44c35e27f2a2 service nova] Acquired lock "refresh_cache-c68974df-ba6e-4f13-b682-737dfe78f265" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2679.977101] env[62824]: DEBUG nova.network.neutron [req-9242ce03-127a-452c-b365-a6450b2362ff req-f20d8d91-fe2a-4598-8d8d-44c35e27f2a2 service nova] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Refreshing network info cache for port 18202605-3a10-4c27-989e-74430907f582 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2680.300622] env[62824]: DEBUG oslo_vmware.api [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]529f4f49-7e35-b9c8-881d-8553b81a2142, 'name': SearchDatastore_Task, 'duration_secs': 0.009195} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2680.300872] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2680.301140] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2680.301376] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2680.301517] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2680.301692] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2680.301965] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2680.302287] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2680.302508] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0369dbba-e727-498c-af95-a215cec80b1c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2680.304385] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2a18075d-df40-4beb-a23e-e30f18157a4d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2680.308610] env[62824]: DEBUG oslo_vmware.api [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Waiting for the task: (returnval){ [ 2680.308610] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]529df4b4-e61d-af56-cec2-914c504eabe6" [ 2680.308610] env[62824]: _type = "Task" [ 2680.308610] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2680.312521] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2680.312702] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2680.313456] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-33ecffe1-e194-4eff-9c08-4973e4bbaf51 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2680.318141] env[62824]: DEBUG oslo_vmware.api [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]529df4b4-e61d-af56-cec2-914c504eabe6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2680.320770] env[62824]: DEBUG oslo_vmware.api [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2680.320770] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]525f7968-2f5e-72af-7fc9-089f51a34264" [ 2680.320770] env[62824]: _type = "Task" [ 2680.320770] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2680.327714] env[62824]: DEBUG oslo_vmware.api [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]525f7968-2f5e-72af-7fc9-089f51a34264, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2680.371647] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b49905e6-2572-4683-9c68-e8ac53d547a0 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.696s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2680.373878] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.587s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2680.374077] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2680.374238] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62824) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 2680.375088] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1727f48e-c5c6-4e8f-8222-975ac7f48aeb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2680.382841] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bee8b5c-9868-4683-bbbe-7b6098d9f71b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2680.388544] env[62824]: INFO nova.scheduler.client.report [None req-b49905e6-2572-4683-9c68-e8ac53d547a0 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Deleted allocations for instance c41daca7-21fc-44fa-9957-d26767e06658 [ 2680.399900] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c6f6ddf-edb6-4c55-9650-2d1a63d860cb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2680.406833] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-189045d5-e1ac-42e5-8fb9-7df433be4186 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2680.437736] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181017MB free_disk=175GB free_vcpus=48 pci_devices=None {{(pid=62824) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 2680.437890] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2680.438094] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2680.666692] env[62824]: DEBUG nova.network.neutron [req-9242ce03-127a-452c-b365-a6450b2362ff req-f20d8d91-fe2a-4598-8d8d-44c35e27f2a2 service nova] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Updated VIF entry in instance network info cache for port 18202605-3a10-4c27-989e-74430907f582. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2680.667050] env[62824]: DEBUG nova.network.neutron [req-9242ce03-127a-452c-b365-a6450b2362ff req-f20d8d91-fe2a-4598-8d8d-44c35e27f2a2 service nova] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Updating instance_info_cache with network_info: [{"id": "18202605-3a10-4c27-989e-74430907f582", "address": "fa:16:3e:29:4a:15", "network": {"id": "db233b0c-27f3-4a76-ae9c-2b5889bf3d3b", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1825671448-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3240be4f32254f54aa87cb045e3d8f31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7041d198-66a3-40de-bf7d-cfc036e6ed69", "external-id": "nsx-vlan-transportzone-278", "segmentation_id": 278, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap18202605-3a", "ovs_interfaceid": "18202605-3a10-4c27-989e-74430907f582", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2680.819079] env[62824]: DEBUG oslo_vmware.api [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]529df4b4-e61d-af56-cec2-914c504eabe6, 'name': SearchDatastore_Task, 'duration_secs': 0.008594} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2680.819452] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2680.819542] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Processing image 9e2a7d30-212d-4ab8-9606-c5c6d52629e8 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2680.819778] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2680.829012] env[62824]: DEBUG oslo_vmware.api [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]525f7968-2f5e-72af-7fc9-089f51a34264, 'name': SearchDatastore_Task, 'duration_secs': 0.007776} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2680.829759] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-af1924fb-4a63-40d8-9fcb-2879e12a62ca {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2680.834435] env[62824]: DEBUG oslo_vmware.api [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2680.834435] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]525370a1-0cc5-3ea5-015c-b12befbb79af" [ 2680.834435] env[62824]: _type = "Task" [ 2680.834435] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2680.841600] env[62824]: DEBUG oslo_vmware.api [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]525370a1-0cc5-3ea5-015c-b12befbb79af, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2680.905019] env[62824]: DEBUG oslo_concurrency.lockutils [None req-b49905e6-2572-4683-9c68-e8ac53d547a0 tempest-AttachVolumeNegativeTest-1225950244 tempest-AttachVolumeNegativeTest-1225950244-project-member] Lock "c41daca7-21fc-44fa-9957-d26767e06658" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.850s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2681.169658] env[62824]: DEBUG oslo_concurrency.lockutils [req-9242ce03-127a-452c-b365-a6450b2362ff req-f20d8d91-fe2a-4598-8d8d-44c35e27f2a2 service nova] Releasing lock "refresh_cache-c68974df-ba6e-4f13-b682-737dfe78f265" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2681.344313] env[62824]: DEBUG oslo_vmware.api [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]525370a1-0cc5-3ea5-015c-b12befbb79af, 'name': SearchDatastore_Task, 'duration_secs': 0.008525} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2681.344589] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2681.344846] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c/02c41e06-f6e9-4b3c-bb29-bee33ab5e13c.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2681.345156] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2681.345345] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2681.345560] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a19f2fe7-1ed3-4285-80bb-2e73de3ce078 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2681.347435] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2bf54329-19f3-44eb-b9b1-c67c5c0394a7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2681.353882] env[62824]: DEBUG oslo_vmware.api [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2681.353882] env[62824]: value = "task-2147197" [ 2681.353882] env[62824]: _type = "Task" [ 2681.353882] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2681.357511] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2681.357685] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2681.358600] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3d9c9730-1099-4042-b51a-a601f8f40770 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2681.363375] env[62824]: DEBUG oslo_vmware.api [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147197, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2681.365834] env[62824]: DEBUG oslo_vmware.api [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Waiting for the task: (returnval){ [ 2681.365834] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52081b73-b9b4-2849-7cfe-98908fa12601" [ 2681.365834] env[62824]: _type = "Task" [ 2681.365834] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2681.372631] env[62824]: DEBUG oslo_vmware.api [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52081b73-b9b4-2849-7cfe-98908fa12601, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2681.461933] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2681.462146] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance c68974df-ba6e-4f13-b682-737dfe78f265 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2681.462336] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Total usable vcpus: 48, total allocated vcpus: 2 {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 2681.462481] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=896MB phys_disk=200GB used_disk=2GB total_vcpus=48 used_vcpus=2 pci_stats=[] {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 2681.499891] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a014ba21-d48c-4cab-990e-bc90aa04a02b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2681.507898] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d97ebef4-13e7-4836-af3c-e5f85200f29c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2681.538959] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf690169-5c61-48d2-8d96-f8c052d4c599 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2681.546854] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2755c3a-c609-4bf2-b6ee-86167c91777c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2681.560542] env[62824]: DEBUG nova.compute.provider_tree [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2681.864376] env[62824]: DEBUG oslo_vmware.api [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147197, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.459774} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2681.864769] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c/02c41e06-f6e9-4b3c-bb29-bee33ab5e13c.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2681.864872] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2681.865146] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e968fbd8-4440-44f9-a363-e160029e74b8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2681.874579] env[62824]: DEBUG oslo_vmware.api [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52081b73-b9b4-2849-7cfe-98908fa12601, 'name': SearchDatastore_Task, 'duration_secs': 0.008639} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2681.876158] env[62824]: DEBUG oslo_vmware.api [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2681.876158] env[62824]: value = "task-2147199" [ 2681.876158] env[62824]: _type = "Task" [ 2681.876158] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2681.876345] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-628554c1-2dc6-4282-94d2-20f4e1305cc4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2681.883647] env[62824]: DEBUG oslo_vmware.api [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Waiting for the task: (returnval){ [ 2681.883647] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52c7fe2f-4756-de00-9ffb-f8f0e0a169cc" [ 2681.883647] env[62824]: _type = "Task" [ 2681.883647] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2681.886325] env[62824]: DEBUG oslo_vmware.api [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147199, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2681.895142] env[62824]: DEBUG oslo_vmware.api [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52c7fe2f-4756-de00-9ffb-f8f0e0a169cc, 'name': SearchDatastore_Task, 'duration_secs': 0.008877} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2681.895368] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2681.895627] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] c68974df-ba6e-4f13-b682-737dfe78f265/c68974df-ba6e-4f13-b682-737dfe78f265.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2681.895836] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-eca4cc05-50e8-4757-9112-b0b273dead1e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2681.901475] env[62824]: DEBUG oslo_vmware.api [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Waiting for the task: (returnval){ [ 2681.901475] env[62824]: value = "task-2147200" [ 2681.901475] env[62824]: _type = "Task" [ 2681.901475] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2681.908568] env[62824]: DEBUG oslo_vmware.api [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147200, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2682.064144] env[62824]: DEBUG nova.scheduler.client.report [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2682.388073] env[62824]: DEBUG oslo_vmware.api [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147199, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066217} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2682.388407] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2682.389219] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4594c70c-f342-4367-aff7-684e4a229693 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2682.411393] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Reconfiguring VM instance instance-0000007b to attach disk [datastore2] 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c/02c41e06-f6e9-4b3c-bb29-bee33ab5e13c.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2682.414377] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d7c11427-9d5d-41c6-93c7-c2a1dc3d23fc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2682.433730] env[62824]: DEBUG oslo_vmware.api [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147200, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.400315} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2682.434847] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9e2a7d30-212d-4ab8-9606-c5c6d52629e8/9e2a7d30-212d-4ab8-9606-c5c6d52629e8.vmdk to [datastore2] c68974df-ba6e-4f13-b682-737dfe78f265/c68974df-ba6e-4f13-b682-737dfe78f265.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2682.435138] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Extending root virtual disk to 1048576 {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 2682.435462] env[62824]: DEBUG oslo_vmware.api [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2682.435462] env[62824]: value = "task-2147201" [ 2682.435462] env[62824]: _type = "Task" [ 2682.435462] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2682.435700] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-83f91f07-cc17-411a-9c79-896036401264 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2682.444530] env[62824]: DEBUG oslo_vmware.api [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147201, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2682.445680] env[62824]: DEBUG oslo_vmware.api [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Waiting for the task: (returnval){ [ 2682.445680] env[62824]: value = "task-2147202" [ 2682.445680] env[62824]: _type = "Task" [ 2682.445680] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2682.453288] env[62824]: DEBUG oslo_vmware.api [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147202, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2682.569349] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62824) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 2682.569583] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.131s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2682.946767] env[62824]: DEBUG oslo_vmware.api [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147201, 'name': ReconfigVM_Task, 'duration_secs': 0.269068} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2682.949510] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Reconfigured VM instance instance-0000007b to attach disk [datastore2] 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c/02c41e06-f6e9-4b3c-bb29-bee33ab5e13c.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2682.950133] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b47e3698-7dba-43e4-b200-7e8dc9a5d06e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2682.956538] env[62824]: DEBUG oslo_vmware.api [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147202, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075082} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2682.957617] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Extended root virtual disk {{(pid=62824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 2682.957953] env[62824]: DEBUG oslo_vmware.api [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2682.957953] env[62824]: value = "task-2147203" [ 2682.957953] env[62824]: _type = "Task" [ 2682.957953] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2682.958638] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82d50be7-4d24-4e86-9a5c-87683d744d25 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2682.987894] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Reconfiguring VM instance instance-0000007c to attach disk [datastore2] c68974df-ba6e-4f13-b682-737dfe78f265/c68974df-ba6e-4f13-b682-737dfe78f265.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2682.988182] env[62824]: DEBUG oslo_vmware.api [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147203, 'name': Rename_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2682.988432] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d720b142-d713-44eb-a846-29441c6821ba {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2683.008485] env[62824]: DEBUG oslo_vmware.api [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Waiting for the task: (returnval){ [ 2683.008485] env[62824]: value = "task-2147204" [ 2683.008485] env[62824]: _type = "Task" [ 2683.008485] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2683.019594] env[62824]: DEBUG oslo_vmware.api [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147204, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2683.057252] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2683.057478] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2683.470216] env[62824]: DEBUG oslo_vmware.api [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147203, 'name': Rename_Task, 'duration_secs': 0.165523} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2683.470497] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2683.470747] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1569b961-d3e4-45e5-97ab-919659626647 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2683.477231] env[62824]: DEBUG oslo_vmware.api [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2683.477231] env[62824]: value = "task-2147205" [ 2683.477231] env[62824]: _type = "Task" [ 2683.477231] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2683.484675] env[62824]: DEBUG oslo_vmware.api [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147205, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2683.520576] env[62824]: DEBUG oslo_vmware.api [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147204, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2683.773524] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2683.773752] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Starting heal instance info cache {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10445}} [ 2683.773894] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Rebuilding the list of instances to heal {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10449}} [ 2683.987082] env[62824]: DEBUG oslo_vmware.api [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147205, 'name': PowerOnVM_Task, 'duration_secs': 0.47099} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2683.987393] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2683.987540] env[62824]: INFO nova.compute.manager [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Took 12.09 seconds to spawn the instance on the hypervisor. [ 2683.987741] env[62824]: DEBUG nova.compute.manager [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2683.988520] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29290b15-0884-4f9b-9873-f6592e87f2f2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2684.016932] env[62824]: DEBUG oslo_vmware.api [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147204, 'name': ReconfigVM_Task, 'duration_secs': 0.51593} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2684.017226] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Reconfigured VM instance instance-0000007c to attach disk [datastore2] c68974df-ba6e-4f13-b682-737dfe78f265/c68974df-ba6e-4f13-b682-737dfe78f265.vmdk or device None with type sparse {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2684.017809] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1c6cf142-ec11-41fe-aa38-1e4b4a546960 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2684.024137] env[62824]: DEBUG oslo_vmware.api [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Waiting for the task: (returnval){ [ 2684.024137] env[62824]: value = "task-2147206" [ 2684.024137] env[62824]: _type = "Task" [ 2684.024137] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2684.031781] env[62824]: DEBUG oslo_vmware.api [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147206, 'name': Rename_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2684.277551] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Skipping network cache update for instance because it is Building. {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10458}} [ 2684.277733] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Skipping network cache update for instance because it is Building. {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10458}} [ 2684.277787] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Didn't find any instances for network info cache update. {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10531}} [ 2684.506078] env[62824]: INFO nova.compute.manager [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Took 16.78 seconds to build instance. [ 2684.533975] env[62824]: DEBUG oslo_vmware.api [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147206, 'name': Rename_Task, 'duration_secs': 0.157265} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2684.534278] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2684.534525] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-18580454-2d27-46f7-8fdd-fd2d0bfcc566 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2684.541891] env[62824]: DEBUG oslo_vmware.api [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Waiting for the task: (returnval){ [ 2684.541891] env[62824]: value = "task-2147208" [ 2684.541891] env[62824]: _type = "Task" [ 2684.541891] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2684.550400] env[62824]: DEBUG oslo_vmware.api [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147208, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2684.772663] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2685.008246] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d3b86c32-9974-437c-ae8e-0d67271842dd tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lock "02c41e06-f6e9-4b3c-bb29-bee33ab5e13c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.293s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2685.052069] env[62824]: DEBUG oslo_vmware.api [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147208, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2685.353921] env[62824]: DEBUG nova.compute.manager [req-f9e26c0b-47ae-4ec2-9398-8c74f0546f20 req-dec2ea2f-e368-476d-b267-2fa12be839ea service nova] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Received event network-changed-2c611b6e-39e9-428c-a425-1eebe84169dd {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2685.353921] env[62824]: DEBUG nova.compute.manager [req-f9e26c0b-47ae-4ec2-9398-8c74f0546f20 req-dec2ea2f-e368-476d-b267-2fa12be839ea service nova] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Refreshing instance network info cache due to event network-changed-2c611b6e-39e9-428c-a425-1eebe84169dd. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2685.354045] env[62824]: DEBUG oslo_concurrency.lockutils [req-f9e26c0b-47ae-4ec2-9398-8c74f0546f20 req-dec2ea2f-e368-476d-b267-2fa12be839ea service nova] Acquiring lock "refresh_cache-02c41e06-f6e9-4b3c-bb29-bee33ab5e13c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2685.354089] env[62824]: DEBUG oslo_concurrency.lockutils [req-f9e26c0b-47ae-4ec2-9398-8c74f0546f20 req-dec2ea2f-e368-476d-b267-2fa12be839ea service nova] Acquired lock "refresh_cache-02c41e06-f6e9-4b3c-bb29-bee33ab5e13c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2685.354248] env[62824]: DEBUG nova.network.neutron [req-f9e26c0b-47ae-4ec2-9398-8c74f0546f20 req-dec2ea2f-e368-476d-b267-2fa12be839ea service nova] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Refreshing network info cache for port 2c611b6e-39e9-428c-a425-1eebe84169dd {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2685.552502] env[62824]: DEBUG oslo_vmware.api [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147208, 'name': PowerOnVM_Task, 'duration_secs': 0.528097} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2685.552838] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2685.553126] env[62824]: INFO nova.compute.manager [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Took 10.60 seconds to spawn the instance on the hypervisor. [ 2685.553357] env[62824]: DEBUG nova.compute.manager [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2685.554183] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-229feb50-e97e-466b-8db0-3e1005add524 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2685.772765] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2686.074472] env[62824]: DEBUG nova.network.neutron [req-f9e26c0b-47ae-4ec2-9398-8c74f0546f20 req-dec2ea2f-e368-476d-b267-2fa12be839ea service nova] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Updated VIF entry in instance network info cache for port 2c611b6e-39e9-428c-a425-1eebe84169dd. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2686.074801] env[62824]: DEBUG nova.network.neutron [req-f9e26c0b-47ae-4ec2-9398-8c74f0546f20 req-dec2ea2f-e368-476d-b267-2fa12be839ea service nova] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Updating instance_info_cache with network_info: [{"id": "2c611b6e-39e9-428c-a425-1eebe84169dd", "address": "fa:16:3e:d2:d7:a6", "network": {"id": "5233a794-3514-4912-a857-7e76a66b1e8f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-910685188-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.129", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0eb41836305847fca773afebdafa01db", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd98a25d-a7a9-4fb5-8fef-e8df4dbbbf11", "external-id": "nsx-vlan-transportzone-707", "segmentation_id": 707, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2c611b6e-39", "ovs_interfaceid": "2c611b6e-39e9-428c-a425-1eebe84169dd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2686.075999] env[62824]: INFO nova.compute.manager [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Took 15.29 seconds to build instance. [ 2686.578369] env[62824]: DEBUG oslo_concurrency.lockutils [req-f9e26c0b-47ae-4ec2-9398-8c74f0546f20 req-dec2ea2f-e368-476d-b267-2fa12be839ea service nova] Releasing lock "refresh_cache-02c41e06-f6e9-4b3c-bb29-bee33ab5e13c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2686.579104] env[62824]: DEBUG oslo_concurrency.lockutils [None req-d1cab692-7e67-4b5a-95a5-06ce9ad9a2ad tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Lock "c68974df-ba6e-4f13-b682-737dfe78f265" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.801s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2686.773571] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2687.384265] env[62824]: DEBUG nova.compute.manager [req-54efb003-3e8c-4387-8fe6-1b59ab291808 req-a6da2524-f4c6-4c9d-a29b-0eb3289b1349 service nova] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Received event network-changed-18202605-3a10-4c27-989e-74430907f582 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2687.384265] env[62824]: DEBUG nova.compute.manager [req-54efb003-3e8c-4387-8fe6-1b59ab291808 req-a6da2524-f4c6-4c9d-a29b-0eb3289b1349 service nova] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Refreshing instance network info cache due to event network-changed-18202605-3a10-4c27-989e-74430907f582. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2687.384510] env[62824]: DEBUG oslo_concurrency.lockutils [req-54efb003-3e8c-4387-8fe6-1b59ab291808 req-a6da2524-f4c6-4c9d-a29b-0eb3289b1349 service nova] Acquiring lock "refresh_cache-c68974df-ba6e-4f13-b682-737dfe78f265" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2687.384510] env[62824]: DEBUG oslo_concurrency.lockutils [req-54efb003-3e8c-4387-8fe6-1b59ab291808 req-a6da2524-f4c6-4c9d-a29b-0eb3289b1349 service nova] Acquired lock "refresh_cache-c68974df-ba6e-4f13-b682-737dfe78f265" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2687.384636] env[62824]: DEBUG nova.network.neutron [req-54efb003-3e8c-4387-8fe6-1b59ab291808 req-a6da2524-f4c6-4c9d-a29b-0eb3289b1349 service nova] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Refreshing network info cache for port 18202605-3a10-4c27-989e-74430907f582 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2688.092277] env[62824]: DEBUG nova.network.neutron [req-54efb003-3e8c-4387-8fe6-1b59ab291808 req-a6da2524-f4c6-4c9d-a29b-0eb3289b1349 service nova] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Updated VIF entry in instance network info cache for port 18202605-3a10-4c27-989e-74430907f582. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2688.092632] env[62824]: DEBUG nova.network.neutron [req-54efb003-3e8c-4387-8fe6-1b59ab291808 req-a6da2524-f4c6-4c9d-a29b-0eb3289b1349 service nova] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Updating instance_info_cache with network_info: [{"id": "18202605-3a10-4c27-989e-74430907f582", "address": "fa:16:3e:29:4a:15", "network": {"id": "db233b0c-27f3-4a76-ae9c-2b5889bf3d3b", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1825671448-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.192", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3240be4f32254f54aa87cb045e3d8f31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7041d198-66a3-40de-bf7d-cfc036e6ed69", "external-id": "nsx-vlan-transportzone-278", "segmentation_id": 278, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap18202605-3a", "ovs_interfaceid": "18202605-3a10-4c27-989e-74430907f582", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2688.595175] env[62824]: DEBUG oslo_concurrency.lockutils [req-54efb003-3e8c-4387-8fe6-1b59ab291808 req-a6da2524-f4c6-4c9d-a29b-0eb3289b1349 service nova] Releasing lock "refresh_cache-c68974df-ba6e-4f13-b682-737dfe78f265" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2688.772810] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2690.772985] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2690.773406] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62824) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11064}} [ 2698.768653] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2723.149538] env[62824]: DEBUG oslo_concurrency.lockutils [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquiring lock "02c41e06-f6e9-4b3c-bb29-bee33ab5e13c" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2723.149979] env[62824]: DEBUG oslo_concurrency.lockutils [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lock "02c41e06-f6e9-4b3c-bb29-bee33ab5e13c" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2723.150077] env[62824]: INFO nova.compute.manager [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Shelving [ 2724.160752] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2724.161195] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c913814e-290d-4806-951d-78f3228605fe {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2724.168249] env[62824]: DEBUG oslo_vmware.api [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2724.168249] env[62824]: value = "task-2147210" [ 2724.168249] env[62824]: _type = "Task" [ 2724.168249] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2724.176665] env[62824]: DEBUG oslo_vmware.api [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147210, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2724.475693] env[62824]: DEBUG oslo_concurrency.lockutils [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Acquiring lock "c68974df-ba6e-4f13-b682-737dfe78f265" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2724.475961] env[62824]: DEBUG oslo_concurrency.lockutils [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Lock "c68974df-ba6e-4f13-b682-737dfe78f265" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2724.476170] env[62824]: INFO nova.compute.manager [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Shelving [ 2724.677733] env[62824]: DEBUG oslo_vmware.api [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147210, 'name': PowerOffVM_Task, 'duration_secs': 0.209104} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2724.678027] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2724.678789] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca211da1-0c31-4c34-a3b9-6288370f9d36 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2724.697506] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82cd9514-d135-47c3-a3ab-ab05ac0b1431 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2725.206854] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Creating Snapshot of the VM instance {{(pid=62824) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 2725.207182] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-e5c9cb7a-9725-409f-94bf-30a435755ade {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2725.214845] env[62824]: DEBUG oslo_vmware.api [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2725.214845] env[62824]: value = "task-2147211" [ 2725.214845] env[62824]: _type = "Task" [ 2725.214845] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2725.222611] env[62824]: DEBUG oslo_vmware.api [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147211, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2725.485223] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2725.485513] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7d1c1332-06d8-4c52-9cbf-35577cbc0d00 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2725.492699] env[62824]: DEBUG oslo_vmware.api [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Waiting for the task: (returnval){ [ 2725.492699] env[62824]: value = "task-2147212" [ 2725.492699] env[62824]: _type = "Task" [ 2725.492699] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2725.500547] env[62824]: DEBUG oslo_vmware.api [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147212, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2725.724177] env[62824]: DEBUG oslo_vmware.api [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147211, 'name': CreateSnapshot_Task, 'duration_secs': 0.40867} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2725.724438] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Created Snapshot of the VM instance {{(pid=62824) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 2725.725159] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b63be6c9-2c41-4ca3-9b0e-613f98c2351c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2726.002994] env[62824]: DEBUG oslo_vmware.api [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147212, 'name': PowerOffVM_Task, 'duration_secs': 0.203783} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2726.003335] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2726.004115] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0feb3642-c63c-4c90-a91e-cb481c84231e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2726.021857] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20b2efe4-5451-4808-a089-b20bf9af670f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2726.242265] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Creating linked-clone VM from snapshot {{(pid=62824) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 2726.242582] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-5b465d5f-619a-44d5-a421-f1107972a61d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2726.251394] env[62824]: DEBUG oslo_vmware.api [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2726.251394] env[62824]: value = "task-2147213" [ 2726.251394] env[62824]: _type = "Task" [ 2726.251394] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2726.259288] env[62824]: DEBUG oslo_vmware.api [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147213, 'name': CloneVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2726.532775] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Creating Snapshot of the VM instance {{(pid=62824) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 2726.533097] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-ff77ac58-f6a9-4c2a-a32f-30375f6fb128 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2726.541935] env[62824]: DEBUG oslo_vmware.api [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Waiting for the task: (returnval){ [ 2726.541935] env[62824]: value = "task-2147214" [ 2726.541935] env[62824]: _type = "Task" [ 2726.541935] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2726.553816] env[62824]: DEBUG oslo_vmware.api [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147214, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2726.761204] env[62824]: DEBUG oslo_vmware.api [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147213, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2727.052154] env[62824]: DEBUG oslo_vmware.api [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147214, 'name': CreateSnapshot_Task, 'duration_secs': 0.42312} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2727.052419] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Created Snapshot of the VM instance {{(pid=62824) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 2727.053141] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-daea07c3-0f95-4b5e-a5f5-4a409006734e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2727.262223] env[62824]: DEBUG oslo_vmware.api [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147213, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2727.570275] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Creating linked-clone VM from snapshot {{(pid=62824) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 2727.570594] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-84e0b04d-572d-44d5-9c06-e2e6e6f6ec53 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2727.579574] env[62824]: DEBUG oslo_vmware.api [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Waiting for the task: (returnval){ [ 2727.579574] env[62824]: value = "task-2147215" [ 2727.579574] env[62824]: _type = "Task" [ 2727.579574] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2727.587196] env[62824]: DEBUG oslo_vmware.api [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147215, 'name': CloneVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2727.763196] env[62824]: DEBUG oslo_vmware.api [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147213, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2728.090541] env[62824]: DEBUG oslo_vmware.api [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147215, 'name': CloneVM_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2728.263934] env[62824]: DEBUG oslo_vmware.api [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147213, 'name': CloneVM_Task} progress is 95%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2728.590058] env[62824]: DEBUG oslo_vmware.api [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147215, 'name': CloneVM_Task} progress is 95%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2728.764629] env[62824]: DEBUG oslo_vmware.api [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147213, 'name': CloneVM_Task, 'duration_secs': 2.473679} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2728.764875] env[62824]: INFO nova.virt.vmwareapi.vmops [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Created linked-clone VM from snapshot [ 2728.765622] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0169a5a4-3bd0-4c71-be03-2622b85ad7a8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2728.772705] env[62824]: DEBUG nova.virt.vmwareapi.images [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Uploading image 585e45af-6ab3-438f-b022-209c5a367d1a {{(pid=62824) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 2728.825784] env[62824]: DEBUG oslo_vmware.rw_handles [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 2728.825784] env[62824]: value = "vm-438803" [ 2728.825784] env[62824]: _type = "VirtualMachine" [ 2728.825784] env[62824]: }. {{(pid=62824) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 2728.826082] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-4252e2e1-2faa-4551-8292-19c3b1599e58 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2728.834529] env[62824]: DEBUG oslo_vmware.rw_handles [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lease: (returnval){ [ 2728.834529] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52df05c8-2a4e-6cb6-e360-869a00b0cb31" [ 2728.834529] env[62824]: _type = "HttpNfcLease" [ 2728.834529] env[62824]: } obtained for exporting VM: (result){ [ 2728.834529] env[62824]: value = "vm-438803" [ 2728.834529] env[62824]: _type = "VirtualMachine" [ 2728.834529] env[62824]: }. {{(pid=62824) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 2728.834900] env[62824]: DEBUG oslo_vmware.api [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the lease: (returnval){ [ 2728.834900] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52df05c8-2a4e-6cb6-e360-869a00b0cb31" [ 2728.834900] env[62824]: _type = "HttpNfcLease" [ 2728.834900] env[62824]: } to be ready. {{(pid=62824) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 2728.840921] env[62824]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 2728.840921] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52df05c8-2a4e-6cb6-e360-869a00b0cb31" [ 2728.840921] env[62824]: _type = "HttpNfcLease" [ 2728.840921] env[62824]: } is initializing. {{(pid=62824) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 2729.091133] env[62824]: DEBUG oslo_vmware.api [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147215, 'name': CloneVM_Task, 'duration_secs': 1.227495} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2729.091371] env[62824]: INFO nova.virt.vmwareapi.vmops [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Created linked-clone VM from snapshot [ 2729.092113] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b03de052-04d4-49ab-9944-94149d817156 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2729.099104] env[62824]: DEBUG nova.virt.vmwareapi.images [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Uploading image 1c41327c-708c-44f8-b176-faab82421719 {{(pid=62824) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 2729.118341] env[62824]: DEBUG oslo_vmware.rw_handles [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 2729.118341] env[62824]: value = "vm-438805" [ 2729.118341] env[62824]: _type = "VirtualMachine" [ 2729.118341] env[62824]: }. {{(pid=62824) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 2729.118567] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-2a3d524b-3f5c-438f-9afe-999c3cf35834 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2729.125261] env[62824]: DEBUG oslo_vmware.rw_handles [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Lease: (returnval){ [ 2729.125261] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52492595-10e9-df97-6a47-6369fc41248c" [ 2729.125261] env[62824]: _type = "HttpNfcLease" [ 2729.125261] env[62824]: } obtained for exporting VM: (result){ [ 2729.125261] env[62824]: value = "vm-438805" [ 2729.125261] env[62824]: _type = "VirtualMachine" [ 2729.125261] env[62824]: }. {{(pid=62824) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 2729.125542] env[62824]: DEBUG oslo_vmware.api [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Waiting for the lease: (returnval){ [ 2729.125542] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52492595-10e9-df97-6a47-6369fc41248c" [ 2729.125542] env[62824]: _type = "HttpNfcLease" [ 2729.125542] env[62824]: } to be ready. {{(pid=62824) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 2729.131485] env[62824]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 2729.131485] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52492595-10e9-df97-6a47-6369fc41248c" [ 2729.131485] env[62824]: _type = "HttpNfcLease" [ 2729.131485] env[62824]: } is initializing. {{(pid=62824) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 2729.342815] env[62824]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 2729.342815] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52df05c8-2a4e-6cb6-e360-869a00b0cb31" [ 2729.342815] env[62824]: _type = "HttpNfcLease" [ 2729.342815] env[62824]: } is ready. {{(pid=62824) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 2729.343288] env[62824]: DEBUG oslo_vmware.rw_handles [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 2729.343288] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52df05c8-2a4e-6cb6-e360-869a00b0cb31" [ 2729.343288] env[62824]: _type = "HttpNfcLease" [ 2729.343288] env[62824]: }. {{(pid=62824) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 2729.343783] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9979732-f342-4a56-849a-6f3791668c84 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2729.350844] env[62824]: DEBUG oslo_vmware.rw_handles [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5229982f-58f6-46fd-35a3-a44e57b326e9/disk-0.vmdk from lease info. {{(pid=62824) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 2729.351027] env[62824]: DEBUG oslo_vmware.rw_handles [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5229982f-58f6-46fd-35a3-a44e57b326e9/disk-0.vmdk for reading. {{(pid=62824) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 2729.438758] env[62824]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-9e0accd0-3cba-44ba-98a6-a196b479c973 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2729.634780] env[62824]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 2729.634780] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52492595-10e9-df97-6a47-6369fc41248c" [ 2729.634780] env[62824]: _type = "HttpNfcLease" [ 2729.634780] env[62824]: } is ready. {{(pid=62824) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 2729.635150] env[62824]: DEBUG oslo_vmware.rw_handles [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 2729.635150] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52492595-10e9-df97-6a47-6369fc41248c" [ 2729.635150] env[62824]: _type = "HttpNfcLease" [ 2729.635150] env[62824]: }. {{(pid=62824) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 2729.635923] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73c7d01d-8e54-4e2b-b2b7-e3fb949b8efb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2729.645649] env[62824]: DEBUG oslo_vmware.rw_handles [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/521c2d76-bf6a-c826-2600-44762bba72d5/disk-0.vmdk from lease info. {{(pid=62824) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 2729.645919] env[62824]: DEBUG oslo_vmware.rw_handles [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/521c2d76-bf6a-c826-2600-44762bba72d5/disk-0.vmdk for reading. {{(pid=62824) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 2729.740058] env[62824]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-0bbee905-b506-4251-99a7-4d12fda95350 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2736.466976] env[62824]: DEBUG oslo_vmware.rw_handles [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5229982f-58f6-46fd-35a3-a44e57b326e9/disk-0.vmdk. {{(pid=62824) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 2736.468070] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8011436e-9683-4cb7-adf0-c01905ad264e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2736.475252] env[62824]: DEBUG oslo_vmware.rw_handles [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5229982f-58f6-46fd-35a3-a44e57b326e9/disk-0.vmdk is in state: ready. {{(pid=62824) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 2736.475475] env[62824]: ERROR oslo_vmware.rw_handles [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5229982f-58f6-46fd-35a3-a44e57b326e9/disk-0.vmdk due to incomplete transfer. [ 2736.475785] env[62824]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-09a4bf32-4f03-4417-b16d-feb968983496 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2736.483288] env[62824]: DEBUG oslo_vmware.rw_handles [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5229982f-58f6-46fd-35a3-a44e57b326e9/disk-0.vmdk. {{(pid=62824) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 2736.483526] env[62824]: DEBUG nova.virt.vmwareapi.images [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Uploaded image 585e45af-6ab3-438f-b022-209c5a367d1a to the Glance image server {{(pid=62824) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 2736.486166] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Destroying the VM {{(pid=62824) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 2736.486450] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-1199ebb5-44a9-44f8-82c0-84459469139f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2736.492165] env[62824]: DEBUG oslo_vmware.api [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2736.492165] env[62824]: value = "task-2147218" [ 2736.492165] env[62824]: _type = "Task" [ 2736.492165] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2736.500380] env[62824]: DEBUG oslo_vmware.api [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147218, 'name': Destroy_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2736.941258] env[62824]: DEBUG oslo_vmware.rw_handles [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/521c2d76-bf6a-c826-2600-44762bba72d5/disk-0.vmdk. {{(pid=62824) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 2736.942168] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87385f10-c789-4b50-aa81-7ad0b715a2ef {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2736.948385] env[62824]: DEBUG oslo_vmware.rw_handles [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/521c2d76-bf6a-c826-2600-44762bba72d5/disk-0.vmdk is in state: ready. {{(pid=62824) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 2736.948543] env[62824]: ERROR oslo_vmware.rw_handles [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/521c2d76-bf6a-c826-2600-44762bba72d5/disk-0.vmdk due to incomplete transfer. [ 2736.948741] env[62824]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-f277e39d-02e6-465d-a94d-9763c89572b5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2736.956192] env[62824]: DEBUG oslo_vmware.rw_handles [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/521c2d76-bf6a-c826-2600-44762bba72d5/disk-0.vmdk. {{(pid=62824) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 2736.956379] env[62824]: DEBUG nova.virt.vmwareapi.images [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Uploaded image 1c41327c-708c-44f8-b176-faab82421719 to the Glance image server {{(pid=62824) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 2736.957993] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Destroying the VM {{(pid=62824) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 2736.958224] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-64d683d9-5475-4cf7-b8d7-e75cb27567c3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2736.964401] env[62824]: DEBUG oslo_vmware.api [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Waiting for the task: (returnval){ [ 2736.964401] env[62824]: value = "task-2147219" [ 2736.964401] env[62824]: _type = "Task" [ 2736.964401] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2736.971879] env[62824]: DEBUG oslo_vmware.api [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147219, 'name': Destroy_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2737.000962] env[62824]: DEBUG oslo_vmware.api [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147218, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2737.473824] env[62824]: DEBUG oslo_vmware.api [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147219, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2737.501598] env[62824]: DEBUG oslo_vmware.api [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147218, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2737.974257] env[62824]: DEBUG oslo_vmware.api [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147219, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2738.002097] env[62824]: DEBUG oslo_vmware.api [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147218, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2738.474925] env[62824]: DEBUG oslo_vmware.api [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147219, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2738.502315] env[62824]: DEBUG oslo_vmware.api [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147218, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2738.772526] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager.update_available_resource {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2738.975763] env[62824]: DEBUG oslo_vmware.api [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147219, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2739.004035] env[62824]: DEBUG oslo_vmware.api [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147218, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2739.276464] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2739.276744] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2739.276830] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2739.276972] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62824) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 2739.277863] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18aa9bb5-12fa-46b3-9ec0-dad4c4991d73 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2739.286146] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51e810b0-3931-430d-bcb1-9f26c4471d35 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2739.301160] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d99279b-b287-484b-8358-3514918ebe65 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2739.307252] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17eed5a4-8f86-44dc-82f8-0810cbd89281 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2739.337955] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180838MB free_disk=175GB free_vcpus=48 pci_devices=None {{(pid=62824) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 2739.338135] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2739.338314] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2739.476696] env[62824]: DEBUG oslo_vmware.api [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147219, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2739.504451] env[62824]: DEBUG oslo_vmware.api [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147218, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2739.977193] env[62824]: DEBUG oslo_vmware.api [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147219, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2740.005253] env[62824]: DEBUG oslo_vmware.api [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147218, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2740.363491] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2740.363648] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance c68974df-ba6e-4f13-b682-737dfe78f265 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2740.363832] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Total usable vcpus: 48, total allocated vcpus: 2 {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 2740.363964] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=896MB phys_disk=200GB used_disk=2GB total_vcpus=48 used_vcpus=2 pci_stats=[] {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 2740.379733] env[62824]: DEBUG nova.scheduler.client.report [None req-41517485-76de-4e8f-a181-a012faec848d None None] Refreshing inventories for resource provider bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 2740.391628] env[62824]: DEBUG nova.scheduler.client.report [None req-41517485-76de-4e8f-a181-a012faec848d None None] Updating ProviderTree inventory for provider bf95157b-a274-42de-9ccf-9851128a44a1 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 2740.391809] env[62824]: DEBUG nova.compute.provider_tree [None req-41517485-76de-4e8f-a181-a012faec848d None None] Updating inventory in ProviderTree for provider bf95157b-a274-42de-9ccf-9851128a44a1 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 2740.402102] env[62824]: DEBUG nova.scheduler.client.report [None req-41517485-76de-4e8f-a181-a012faec848d None None] Refreshing aggregate associations for resource provider bf95157b-a274-42de-9ccf-9851128a44a1, aggregates: None {{(pid=62824) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 2740.419871] env[62824]: DEBUG nova.scheduler.client.report [None req-41517485-76de-4e8f-a181-a012faec848d None None] Refreshing trait associations for resource provider bf95157b-a274-42de-9ccf-9851128a44a1, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_ISO {{(pid=62824) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 2740.453797] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f1b023d-3de5-4214-8369-a510ebf1d248 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2740.461322] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47d45334-0774-4ae5-a90a-2e55710c0590 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2740.493896] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-262b2f31-c44e-4006-8030-e37d36829cd8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2740.504670] env[62824]: DEBUG oslo_vmware.api [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147219, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2740.506155] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ee0f641-e9ad-4749-98dd-ad8aff94ab22 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2740.512365] env[62824]: DEBUG oslo_vmware.api [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147218, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2740.521858] env[62824]: DEBUG nova.compute.provider_tree [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2741.000753] env[62824]: DEBUG oslo_vmware.api [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147219, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2741.008788] env[62824]: DEBUG oslo_vmware.api [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147218, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2741.024737] env[62824]: DEBUG nova.scheduler.client.report [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2741.500457] env[62824]: DEBUG oslo_vmware.api [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147219, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2741.507836] env[62824]: DEBUG oslo_vmware.api [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147218, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2741.529721] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62824) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 2741.529909] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.192s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2742.001810] env[62824]: DEBUG oslo_vmware.api [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147219, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2742.009564] env[62824]: DEBUG oslo_vmware.api [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147218, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2742.502464] env[62824]: DEBUG oslo_vmware.api [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147219, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2742.509798] env[62824]: DEBUG oslo_vmware.api [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147218, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2742.525375] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2743.004179] env[62824]: DEBUG oslo_vmware.api [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147219, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2743.011404] env[62824]: DEBUG oslo_vmware.api [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147218, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2743.504794] env[62824]: DEBUG oslo_vmware.api [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147219, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2743.512077] env[62824]: DEBUG oslo_vmware.api [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147218, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2744.005115] env[62824]: DEBUG oslo_vmware.api [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147219, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2744.012379] env[62824]: DEBUG oslo_vmware.api [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147218, 'name': Destroy_Task} progress is 33%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2744.506351] env[62824]: DEBUG oslo_vmware.api [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147219, 'name': Destroy_Task, 'duration_secs': 7.466544} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2744.509369] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Destroyed the VM [ 2744.509621] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Deleting Snapshot of the VM instance {{(pid=62824) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 2744.509872] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-405d0c4b-30f2-425f-a6d9-27fa2d4472b8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2744.516284] env[62824]: DEBUG oslo_vmware.api [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147218, 'name': Destroy_Task, 'duration_secs': 7.93332} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2744.517315] env[62824]: INFO nova.virt.vmwareapi.vm_util [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Destroyed the VM [ 2744.517538] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Deleting Snapshot of the VM instance {{(pid=62824) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 2744.517816] env[62824]: DEBUG oslo_vmware.api [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Waiting for the task: (returnval){ [ 2744.517816] env[62824]: value = "task-2147220" [ 2744.517816] env[62824]: _type = "Task" [ 2744.517816] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2744.517990] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-ed2375c3-de67-4cb4-b53d-5e983c81498b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2744.527113] env[62824]: DEBUG oslo_vmware.api [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147220, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2744.528197] env[62824]: DEBUG oslo_vmware.api [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2744.528197] env[62824]: value = "task-2147221" [ 2744.528197] env[62824]: _type = "Task" [ 2744.528197] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2744.535009] env[62824]: DEBUG oslo_vmware.api [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147221, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2744.772998] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2744.773218] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Starting heal instance info cache {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10445}} [ 2744.773283] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Rebuilding the list of instances to heal {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10449}} [ 2745.028851] env[62824]: DEBUG oslo_vmware.api [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147220, 'name': RemoveSnapshot_Task, 'duration_secs': 0.315125} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2745.031793] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Deleted Snapshot of the VM instance {{(pid=62824) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 2745.032099] env[62824]: DEBUG nova.compute.manager [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2745.032813] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-019006a8-6896-492b-9a73-d58d9915143e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2745.043358] env[62824]: DEBUG oslo_vmware.api [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147221, 'name': RemoveSnapshot_Task, 'duration_secs': 0.311309} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2745.043802] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Deleted Snapshot of the VM instance {{(pid=62824) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 2745.044071] env[62824]: DEBUG nova.compute.manager [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2745.044775] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a453db4-d951-4ddd-8fb4-15b855536ced {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2745.277215] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "refresh_cache-02c41e06-f6e9-4b3c-bb29-bee33ab5e13c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2745.277517] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquired lock "refresh_cache-02c41e06-f6e9-4b3c-bb29-bee33ab5e13c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2745.277517] env[62824]: DEBUG nova.network.neutron [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Forcefully refreshing network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 2745.277660] env[62824]: DEBUG nova.objects.instance [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lazy-loading 'info_cache' on Instance uuid 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2745.547821] env[62824]: INFO nova.compute.manager [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Shelve offloading [ 2745.554191] env[62824]: INFO nova.compute.manager [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Shelve offloading [ 2746.051493] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2746.051634] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e58f767c-d348-448b-8aed-1f2a18f25ed7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2746.057203] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2746.057434] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-225642ce-d6a5-4fc3-9afe-de032b4cd6b4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2746.060087] env[62824]: DEBUG oslo_vmware.api [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Waiting for the task: (returnval){ [ 2746.060087] env[62824]: value = "task-2147222" [ 2746.060087] env[62824]: _type = "Task" [ 2746.060087] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2746.068400] env[62824]: DEBUG oslo_vmware.api [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147222, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2746.069530] env[62824]: DEBUG oslo_vmware.api [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2746.069530] env[62824]: value = "task-2147223" [ 2746.069530] env[62824]: _type = "Task" [ 2746.069530] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2746.076297] env[62824]: DEBUG oslo_vmware.api [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147223, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2746.573996] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] VM already powered off {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 2746.574335] env[62824]: DEBUG nova.compute.manager [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2746.574887] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11a7a92a-b8fe-4bbc-bbac-23ac9902dc11 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2746.583299] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] VM already powered off {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 2746.583487] env[62824]: DEBUG nova.compute.manager [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2746.583792] env[62824]: DEBUG oslo_concurrency.lockutils [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Acquiring lock "refresh_cache-c68974df-ba6e-4f13-b682-737dfe78f265" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2746.583949] env[62824]: DEBUG oslo_concurrency.lockutils [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Acquired lock "refresh_cache-c68974df-ba6e-4f13-b682-737dfe78f265" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2746.584126] env[62824]: DEBUG nova.network.neutron [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2746.585691] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ad86752-6759-491e-9030-e2cf5a8c36ea {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2746.591626] env[62824]: DEBUG oslo_concurrency.lockutils [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquiring lock "refresh_cache-02c41e06-f6e9-4b3c-bb29-bee33ab5e13c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2747.018099] env[62824]: DEBUG nova.network.neutron [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Updating instance_info_cache with network_info: [{"id": "2c611b6e-39e9-428c-a425-1eebe84169dd", "address": "fa:16:3e:d2:d7:a6", "network": {"id": "5233a794-3514-4912-a857-7e76a66b1e8f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-910685188-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.129", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0eb41836305847fca773afebdafa01db", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd98a25d-a7a9-4fb5-8fef-e8df4dbbbf11", "external-id": "nsx-vlan-transportzone-707", "segmentation_id": 707, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2c611b6e-39", "ovs_interfaceid": "2c611b6e-39e9-428c-a425-1eebe84169dd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2747.291512] env[62824]: DEBUG nova.network.neutron [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Updating instance_info_cache with network_info: [{"id": "18202605-3a10-4c27-989e-74430907f582", "address": "fa:16:3e:29:4a:15", "network": {"id": "db233b0c-27f3-4a76-ae9c-2b5889bf3d3b", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1825671448-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.192", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3240be4f32254f54aa87cb045e3d8f31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7041d198-66a3-40de-bf7d-cfc036e6ed69", "external-id": "nsx-vlan-transportzone-278", "segmentation_id": 278, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap18202605-3a", "ovs_interfaceid": "18202605-3a10-4c27-989e-74430907f582", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2747.521142] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Releasing lock "refresh_cache-02c41e06-f6e9-4b3c-bb29-bee33ab5e13c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2747.521371] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Updated the network info_cache for instance {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10516}} [ 2747.521632] env[62824]: DEBUG oslo_concurrency.lockutils [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquired lock "refresh_cache-02c41e06-f6e9-4b3c-bb29-bee33ab5e13c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2747.521803] env[62824]: DEBUG nova.network.neutron [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2747.522990] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2747.523194] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2747.523349] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2747.523498] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2747.793852] env[62824]: DEBUG oslo_concurrency.lockutils [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Releasing lock "refresh_cache-c68974df-ba6e-4f13-b682-737dfe78f265" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2748.062698] env[62824]: DEBUG nova.compute.manager [req-a3387ac2-378e-4e22-ab15-7ffd580fbf8d req-01b5c8ae-1233-4b5b-a12b-cb2e13b8be72 service nova] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Received event network-vif-unplugged-18202605-3a10-4c27-989e-74430907f582 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2748.062931] env[62824]: DEBUG oslo_concurrency.lockutils [req-a3387ac2-378e-4e22-ab15-7ffd580fbf8d req-01b5c8ae-1233-4b5b-a12b-cb2e13b8be72 service nova] Acquiring lock "c68974df-ba6e-4f13-b682-737dfe78f265-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2748.063164] env[62824]: DEBUG oslo_concurrency.lockutils [req-a3387ac2-378e-4e22-ab15-7ffd580fbf8d req-01b5c8ae-1233-4b5b-a12b-cb2e13b8be72 service nova] Lock "c68974df-ba6e-4f13-b682-737dfe78f265-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2748.063334] env[62824]: DEBUG oslo_concurrency.lockutils [req-a3387ac2-378e-4e22-ab15-7ffd580fbf8d req-01b5c8ae-1233-4b5b-a12b-cb2e13b8be72 service nova] Lock "c68974df-ba6e-4f13-b682-737dfe78f265-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2748.063494] env[62824]: DEBUG nova.compute.manager [req-a3387ac2-378e-4e22-ab15-7ffd580fbf8d req-01b5c8ae-1233-4b5b-a12b-cb2e13b8be72 service nova] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] No waiting events found dispatching network-vif-unplugged-18202605-3a10-4c27-989e-74430907f582 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2748.063694] env[62824]: WARNING nova.compute.manager [req-a3387ac2-378e-4e22-ab15-7ffd580fbf8d req-01b5c8ae-1233-4b5b-a12b-cb2e13b8be72 service nova] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Received unexpected event network-vif-unplugged-18202605-3a10-4c27-989e-74430907f582 for instance with vm_state shelved and task_state shelving_offloading. [ 2748.209558] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2748.210488] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e704c039-449f-4aff-b8ac-d6cf99f52eb0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2748.217773] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2748.218016] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c864b3b9-5a36-41ac-af48-ee74d2c1ed1a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2748.310731] env[62824]: DEBUG nova.network.neutron [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Updating instance_info_cache with network_info: [{"id": "2c611b6e-39e9-428c-a425-1eebe84169dd", "address": "fa:16:3e:d2:d7:a6", "network": {"id": "5233a794-3514-4912-a857-7e76a66b1e8f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-910685188-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.129", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0eb41836305847fca773afebdafa01db", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd98a25d-a7a9-4fb5-8fef-e8df4dbbbf11", "external-id": "nsx-vlan-transportzone-707", "segmentation_id": 707, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2c611b6e-39", "ovs_interfaceid": "2c611b6e-39e9-428c-a425-1eebe84169dd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2748.514386] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2748.514633] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Deleting contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2748.514771] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Deleting the datastore file [datastore2] c68974df-ba6e-4f13-b682-737dfe78f265 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2748.515060] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e5ecd396-07a5-4c3a-82f0-9d7d43f00db1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2748.522075] env[62824]: DEBUG oslo_vmware.api [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Waiting for the task: (returnval){ [ 2748.522075] env[62824]: value = "task-2147225" [ 2748.522075] env[62824]: _type = "Task" [ 2748.522075] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2748.529678] env[62824]: DEBUG oslo_vmware.api [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147225, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2748.773405] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2748.813411] env[62824]: DEBUG oslo_concurrency.lockutils [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Releasing lock "refresh_cache-02c41e06-f6e9-4b3c-bb29-bee33ab5e13c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2749.033930] env[62824]: DEBUG oslo_vmware.api [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147225, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.131459} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2749.034579] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2749.034778] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Deleted contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2749.034962] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2749.059806] env[62824]: INFO nova.scheduler.client.report [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Deleted allocations for instance c68974df-ba6e-4f13-b682-737dfe78f265 [ 2749.117084] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2749.117961] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-399fe7c4-8720-4ed3-b08a-2bd861f41eee {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2749.125870] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2749.126113] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a4ec9bb6-9e52-4e84-a8c9-3cd030f09d92 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2749.292048] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2749.292357] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Deleting contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2749.292614] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Deleting the datastore file [datastore2] 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2749.292961] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0dfb0cee-1f7e-4f97-91a8-7a6ddbb7f781 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2749.300367] env[62824]: DEBUG oslo_vmware.api [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2749.300367] env[62824]: value = "task-2147227" [ 2749.300367] env[62824]: _type = "Task" [ 2749.300367] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2749.308963] env[62824]: DEBUG oslo_vmware.api [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147227, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2749.564141] env[62824]: DEBUG oslo_concurrency.lockutils [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2749.564404] env[62824]: DEBUG oslo_concurrency.lockutils [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2749.564624] env[62824]: DEBUG nova.objects.instance [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Lazy-loading 'resources' on Instance uuid c68974df-ba6e-4f13-b682-737dfe78f265 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2749.810374] env[62824]: DEBUG oslo_vmware.api [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147227, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.126168} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2749.810629] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2749.810825] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Deleted contents of the VM from datastore datastore2 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2749.811010] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2749.826693] env[62824]: INFO nova.scheduler.client.report [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Deleted allocations for instance 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c [ 2750.067491] env[62824]: DEBUG nova.objects.instance [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Lazy-loading 'numa_topology' on Instance uuid c68974df-ba6e-4f13-b682-737dfe78f265 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2750.091731] env[62824]: DEBUG nova.compute.manager [req-c68ec581-3189-48c3-9b64-109a4abbf829 req-10a43993-d0fd-4845-a6e7-4f7075a6a8f8 service nova] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Received event network-changed-18202605-3a10-4c27-989e-74430907f582 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2750.091942] env[62824]: DEBUG nova.compute.manager [req-c68ec581-3189-48c3-9b64-109a4abbf829 req-10a43993-d0fd-4845-a6e7-4f7075a6a8f8 service nova] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Refreshing instance network info cache due to event network-changed-18202605-3a10-4c27-989e-74430907f582. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2750.092189] env[62824]: DEBUG oslo_concurrency.lockutils [req-c68ec581-3189-48c3-9b64-109a4abbf829 req-10a43993-d0fd-4845-a6e7-4f7075a6a8f8 service nova] Acquiring lock "refresh_cache-c68974df-ba6e-4f13-b682-737dfe78f265" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2750.092334] env[62824]: DEBUG oslo_concurrency.lockutils [req-c68ec581-3189-48c3-9b64-109a4abbf829 req-10a43993-d0fd-4845-a6e7-4f7075a6a8f8 service nova] Acquired lock "refresh_cache-c68974df-ba6e-4f13-b682-737dfe78f265" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2750.092499] env[62824]: DEBUG nova.network.neutron [req-c68ec581-3189-48c3-9b64-109a4abbf829 req-10a43993-d0fd-4845-a6e7-4f7075a6a8f8 service nova] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Refreshing network info cache for port 18202605-3a10-4c27-989e-74430907f582 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2750.331098] env[62824]: DEBUG oslo_concurrency.lockutils [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2750.569974] env[62824]: DEBUG nova.objects.base [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=62824) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 2750.584743] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44d06950-abbe-4293-8058-be8d49c5514c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2750.592332] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b49d79ba-0395-4760-994c-ab7884de90cb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2750.628022] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54cb7886-17e6-4769-9ca0-f3834381345d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2750.635770] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c89c267c-a40c-4b61-bfe5-0c560b1a0a4b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2750.649339] env[62824]: DEBUG nova.compute.provider_tree [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2750.836060] env[62824]: DEBUG nova.network.neutron [req-c68ec581-3189-48c3-9b64-109a4abbf829 req-10a43993-d0fd-4845-a6e7-4f7075a6a8f8 service nova] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Updated VIF entry in instance network info cache for port 18202605-3a10-4c27-989e-74430907f582. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2750.836494] env[62824]: DEBUG nova.network.neutron [req-c68ec581-3189-48c3-9b64-109a4abbf829 req-10a43993-d0fd-4845-a6e7-4f7075a6a8f8 service nova] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Updating instance_info_cache with network_info: [{"id": "18202605-3a10-4c27-989e-74430907f582", "address": "fa:16:3e:29:4a:15", "network": {"id": "db233b0c-27f3-4a76-ae9c-2b5889bf3d3b", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-1825671448-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.192", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3240be4f32254f54aa87cb045e3d8f31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap18202605-3a", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2751.152983] env[62824]: DEBUG nova.scheduler.client.report [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2751.339271] env[62824]: DEBUG oslo_concurrency.lockutils [req-c68ec581-3189-48c3-9b64-109a4abbf829 req-10a43993-d0fd-4845-a6e7-4f7075a6a8f8 service nova] Releasing lock "refresh_cache-c68974df-ba6e-4f13-b682-737dfe78f265" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2751.339532] env[62824]: DEBUG nova.compute.manager [req-c68ec581-3189-48c3-9b64-109a4abbf829 req-10a43993-d0fd-4845-a6e7-4f7075a6a8f8 service nova] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Received event network-vif-unplugged-2c611b6e-39e9-428c-a425-1eebe84169dd {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2751.339726] env[62824]: DEBUG oslo_concurrency.lockutils [req-c68ec581-3189-48c3-9b64-109a4abbf829 req-10a43993-d0fd-4845-a6e7-4f7075a6a8f8 service nova] Acquiring lock "02c41e06-f6e9-4b3c-bb29-bee33ab5e13c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2751.339943] env[62824]: DEBUG oslo_concurrency.lockutils [req-c68ec581-3189-48c3-9b64-109a4abbf829 req-10a43993-d0fd-4845-a6e7-4f7075a6a8f8 service nova] Lock "02c41e06-f6e9-4b3c-bb29-bee33ab5e13c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2751.340124] env[62824]: DEBUG oslo_concurrency.lockutils [req-c68ec581-3189-48c3-9b64-109a4abbf829 req-10a43993-d0fd-4845-a6e7-4f7075a6a8f8 service nova] Lock "02c41e06-f6e9-4b3c-bb29-bee33ab5e13c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2751.340303] env[62824]: DEBUG nova.compute.manager [req-c68ec581-3189-48c3-9b64-109a4abbf829 req-10a43993-d0fd-4845-a6e7-4f7075a6a8f8 service nova] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] No waiting events found dispatching network-vif-unplugged-2c611b6e-39e9-428c-a425-1eebe84169dd {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2751.340470] env[62824]: WARNING nova.compute.manager [req-c68ec581-3189-48c3-9b64-109a4abbf829 req-10a43993-d0fd-4845-a6e7-4f7075a6a8f8 service nova] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Received unexpected event network-vif-unplugged-2c611b6e-39e9-428c-a425-1eebe84169dd for instance with vm_state shelved_offloaded and task_state None. [ 2751.340633] env[62824]: DEBUG nova.compute.manager [req-c68ec581-3189-48c3-9b64-109a4abbf829 req-10a43993-d0fd-4845-a6e7-4f7075a6a8f8 service nova] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Received event network-changed-2c611b6e-39e9-428c-a425-1eebe84169dd {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2751.340793] env[62824]: DEBUG nova.compute.manager [req-c68ec581-3189-48c3-9b64-109a4abbf829 req-10a43993-d0fd-4845-a6e7-4f7075a6a8f8 service nova] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Refreshing instance network info cache due to event network-changed-2c611b6e-39e9-428c-a425-1eebe84169dd. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2751.340980] env[62824]: DEBUG oslo_concurrency.lockutils [req-c68ec581-3189-48c3-9b64-109a4abbf829 req-10a43993-d0fd-4845-a6e7-4f7075a6a8f8 service nova] Acquiring lock "refresh_cache-02c41e06-f6e9-4b3c-bb29-bee33ab5e13c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2751.341133] env[62824]: DEBUG oslo_concurrency.lockutils [req-c68ec581-3189-48c3-9b64-109a4abbf829 req-10a43993-d0fd-4845-a6e7-4f7075a6a8f8 service nova] Acquired lock "refresh_cache-02c41e06-f6e9-4b3c-bb29-bee33ab5e13c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2751.341290] env[62824]: DEBUG nova.network.neutron [req-c68ec581-3189-48c3-9b64-109a4abbf829 req-10a43993-d0fd-4845-a6e7-4f7075a6a8f8 service nova] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Refreshing network info cache for port 2c611b6e-39e9-428c-a425-1eebe84169dd {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2751.657872] env[62824]: DEBUG oslo_concurrency.lockutils [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.093s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2751.660201] env[62824]: DEBUG oslo_concurrency.lockutils [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.329s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2751.660428] env[62824]: DEBUG nova.objects.instance [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lazy-loading 'resources' on Instance uuid 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2751.772677] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2751.772838] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62824) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11064}} [ 2751.880652] env[62824]: DEBUG oslo_concurrency.lockutils [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquiring lock "02c41e06-f6e9-4b3c-bb29-bee33ab5e13c" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2751.929972] env[62824]: DEBUG oslo_concurrency.lockutils [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Acquiring lock "c68974df-ba6e-4f13-b682-737dfe78f265" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2752.164094] env[62824]: DEBUG nova.objects.instance [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lazy-loading 'numa_topology' on Instance uuid 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2752.166927] env[62824]: DEBUG oslo_concurrency.lockutils [None req-618f1ee6-c70a-4b7f-add5-1d02fcf3d8d5 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Lock "c68974df-ba6e-4f13-b682-737dfe78f265" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 27.691s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2752.171021] env[62824]: DEBUG oslo_concurrency.lockutils [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Lock "c68974df-ba6e-4f13-b682-737dfe78f265" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 0.238s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2752.171021] env[62824]: INFO nova.compute.manager [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Unshelving [ 2752.236927] env[62824]: DEBUG nova.network.neutron [req-c68ec581-3189-48c3-9b64-109a4abbf829 req-10a43993-d0fd-4845-a6e7-4f7075a6a8f8 service nova] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Updated VIF entry in instance network info cache for port 2c611b6e-39e9-428c-a425-1eebe84169dd. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2752.237303] env[62824]: DEBUG nova.network.neutron [req-c68ec581-3189-48c3-9b64-109a4abbf829 req-10a43993-d0fd-4845-a6e7-4f7075a6a8f8 service nova] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Updating instance_info_cache with network_info: [{"id": "2c611b6e-39e9-428c-a425-1eebe84169dd", "address": "fa:16:3e:d2:d7:a6", "network": {"id": "5233a794-3514-4912-a857-7e76a66b1e8f", "bridge": null, "label": "tempest-ServerActionsTestOtherB-910685188-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.129", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0eb41836305847fca773afebdafa01db", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap2c611b6e-39", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2752.667865] env[62824]: DEBUG nova.objects.base [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Object Instance<02c41e06-f6e9-4b3c-bb29-bee33ab5e13c> lazy-loaded attributes: resources,numa_topology {{(pid=62824) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 2752.731336] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98f3b3b1-f569-42a6-85a6-9ab78b49df3d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2752.739281] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1e3558d-8924-4b66-bb8a-790f32403758 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2752.742366] env[62824]: DEBUG oslo_concurrency.lockutils [req-c68ec581-3189-48c3-9b64-109a4abbf829 req-10a43993-d0fd-4845-a6e7-4f7075a6a8f8 service nova] Releasing lock "refresh_cache-02c41e06-f6e9-4b3c-bb29-bee33ab5e13c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2752.769017] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6981e648-2e17-4629-8930-077e0ed1acc0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2752.776203] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4836b783-aaea-4a3c-9662-e00682bb7b4b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2752.789890] env[62824]: DEBUG nova.compute.provider_tree [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2753.190122] env[62824]: DEBUG oslo_concurrency.lockutils [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2753.293430] env[62824]: DEBUG nova.scheduler.client.report [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2753.798621] env[62824]: DEBUG oslo_concurrency.lockutils [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.138s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2753.801354] env[62824]: DEBUG oslo_concurrency.lockutils [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.611s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2753.801569] env[62824]: DEBUG nova.objects.instance [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Lazy-loading 'pci_requests' on Instance uuid c68974df-ba6e-4f13-b682-737dfe78f265 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2754.307822] env[62824]: DEBUG nova.objects.instance [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Lazy-loading 'numa_topology' on Instance uuid c68974df-ba6e-4f13-b682-737dfe78f265 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2754.309363] env[62824]: DEBUG oslo_concurrency.lockutils [None req-13954d1f-f00c-4a21-a953-5976d56c1f12 tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lock "02c41e06-f6e9-4b3c-bb29-bee33ab5e13c" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 31.160s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2754.310119] env[62824]: DEBUG oslo_concurrency.lockutils [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lock "02c41e06-f6e9-4b3c-bb29-bee33ab5e13c" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 2.430s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2754.310301] env[62824]: INFO nova.compute.manager [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Unshelving [ 2754.810593] env[62824]: INFO nova.compute.claims [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2755.331350] env[62824]: DEBUG oslo_concurrency.lockutils [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2755.857030] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a6c70b0-86e6-4016-be2d-613c6551823e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2755.864781] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5025a98-6cab-40da-8f55-de3b50b57efb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2755.893830] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b66421f-03ef-4a6c-b290-83197e28921f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2755.901025] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6aba52e4-449f-4ff6-94a4-d2422a915e91 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2755.912970] env[62824]: DEBUG nova.compute.provider_tree [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2756.415991] env[62824]: DEBUG nova.scheduler.client.report [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2756.921658] env[62824]: DEBUG oslo_concurrency.lockutils [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.120s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2756.923898] env[62824]: DEBUG oslo_concurrency.lockutils [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.593s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2756.924147] env[62824]: DEBUG nova.objects.instance [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lazy-loading 'pci_requests' on Instance uuid 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2756.954143] env[62824]: INFO nova.network.neutron [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Updating port 18202605-3a10-4c27-989e-74430907f582 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 2757.428761] env[62824]: DEBUG nova.objects.instance [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lazy-loading 'numa_topology' on Instance uuid 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2757.933493] env[62824]: INFO nova.compute.claims [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 2758.986655] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbe1b444-a9bf-4897-b45a-2c2b5a5f4c9d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2758.994432] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ece5b793-4b96-4040-8e31-27049247d300 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2759.023319] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2762644-1ed0-457e-8801-b5f6c67b8d9f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2759.029929] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61207dcc-03a2-40a9-8510-d696140648eb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2759.042596] env[62824]: DEBUG nova.compute.provider_tree [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2759.546549] env[62824]: DEBUG nova.scheduler.client.report [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2760.051601] env[62824]: DEBUG oslo_concurrency.lockutils [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.128s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2760.081762] env[62824]: INFO nova.network.neutron [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Updating port 2c611b6e-39e9-428c-a425-1eebe84169dd with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 2761.452745] env[62824]: DEBUG nova.compute.manager [req-817c1f5c-6b79-46ce-a96e-14477146e824 req-f466a83c-2084-4dee-ac97-b8bebb2a8480 service nova] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Received event network-vif-plugged-2c611b6e-39e9-428c-a425-1eebe84169dd {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2761.453118] env[62824]: DEBUG oslo_concurrency.lockutils [req-817c1f5c-6b79-46ce-a96e-14477146e824 req-f466a83c-2084-4dee-ac97-b8bebb2a8480 service nova] Acquiring lock "02c41e06-f6e9-4b3c-bb29-bee33ab5e13c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2761.453246] env[62824]: DEBUG oslo_concurrency.lockutils [req-817c1f5c-6b79-46ce-a96e-14477146e824 req-f466a83c-2084-4dee-ac97-b8bebb2a8480 service nova] Lock "02c41e06-f6e9-4b3c-bb29-bee33ab5e13c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2761.453385] env[62824]: DEBUG oslo_concurrency.lockutils [req-817c1f5c-6b79-46ce-a96e-14477146e824 req-f466a83c-2084-4dee-ac97-b8bebb2a8480 service nova] Lock "02c41e06-f6e9-4b3c-bb29-bee33ab5e13c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2761.453552] env[62824]: DEBUG nova.compute.manager [req-817c1f5c-6b79-46ce-a96e-14477146e824 req-f466a83c-2084-4dee-ac97-b8bebb2a8480 service nova] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] No waiting events found dispatching network-vif-plugged-2c611b6e-39e9-428c-a425-1eebe84169dd {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2761.453712] env[62824]: WARNING nova.compute.manager [req-817c1f5c-6b79-46ce-a96e-14477146e824 req-f466a83c-2084-4dee-ac97-b8bebb2a8480 service nova] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Received unexpected event network-vif-plugged-2c611b6e-39e9-428c-a425-1eebe84169dd for instance with vm_state shelved_offloaded and task_state spawning. [ 2761.540200] env[62824]: DEBUG oslo_concurrency.lockutils [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquiring lock "refresh_cache-02c41e06-f6e9-4b3c-bb29-bee33ab5e13c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2761.540200] env[62824]: DEBUG oslo_concurrency.lockutils [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquired lock "refresh_cache-02c41e06-f6e9-4b3c-bb29-bee33ab5e13c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2761.540200] env[62824]: DEBUG nova.network.neutron [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2761.921564] env[62824]: DEBUG oslo_concurrency.lockutils [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Acquiring lock "refresh_cache-c68974df-ba6e-4f13-b682-737dfe78f265" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2761.921751] env[62824]: DEBUG oslo_concurrency.lockutils [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Acquired lock "refresh_cache-c68974df-ba6e-4f13-b682-737dfe78f265" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2761.921932] env[62824]: DEBUG nova.network.neutron [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Building network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 2762.241412] env[62824]: DEBUG nova.network.neutron [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Updating instance_info_cache with network_info: [{"id": "2c611b6e-39e9-428c-a425-1eebe84169dd", "address": "fa:16:3e:d2:d7:a6", "network": {"id": "5233a794-3514-4912-a857-7e76a66b1e8f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-910685188-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.129", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0eb41836305847fca773afebdafa01db", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd98a25d-a7a9-4fb5-8fef-e8df4dbbbf11", "external-id": "nsx-vlan-transportzone-707", "segmentation_id": 707, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2c611b6e-39", "ovs_interfaceid": "2c611b6e-39e9-428c-a425-1eebe84169dd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2762.621697] env[62824]: DEBUG nova.network.neutron [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Updating instance_info_cache with network_info: [{"id": "18202605-3a10-4c27-989e-74430907f582", "address": "fa:16:3e:29:4a:15", "network": {"id": "db233b0c-27f3-4a76-ae9c-2b5889bf3d3b", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1825671448-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.192", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3240be4f32254f54aa87cb045e3d8f31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7041d198-66a3-40de-bf7d-cfc036e6ed69", "external-id": "nsx-vlan-transportzone-278", "segmentation_id": 278, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap18202605-3a", "ovs_interfaceid": "18202605-3a10-4c27-989e-74430907f582", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2762.744388] env[62824]: DEBUG oslo_concurrency.lockutils [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Releasing lock "refresh_cache-02c41e06-f6e9-4b3c-bb29-bee33ab5e13c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2762.771475] env[62824]: DEBUG nova.virt.hardware [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='f2cfa8ceaeddbcc86dc52d1cc5062f9f',container_format='bare',created_at=2025-01-16T15:05:26Z,direct_url=,disk_format='vmdk',id=585e45af-6ab3-438f-b022-209c5a367d1a,min_disk=1,min_ram=0,name='tempest-ServerActionsTestOtherB-server-1066615763-shelved',owner='0eb41836305847fca773afebdafa01db',properties=ImageMetaProps,protected=,size=31668736,status='active',tags=,updated_at=2025-01-16T15:05:40Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2762.771720] env[62824]: DEBUG nova.virt.hardware [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2762.771875] env[62824]: DEBUG nova.virt.hardware [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2762.772072] env[62824]: DEBUG nova.virt.hardware [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2762.772222] env[62824]: DEBUG nova.virt.hardware [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2762.772368] env[62824]: DEBUG nova.virt.hardware [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2762.772572] env[62824]: DEBUG nova.virt.hardware [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2762.772731] env[62824]: DEBUG nova.virt.hardware [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2762.772897] env[62824]: DEBUG nova.virt.hardware [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2762.773074] env[62824]: DEBUG nova.virt.hardware [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2762.773252] env[62824]: DEBUG nova.virt.hardware [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2762.774126] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bea34699-36d1-4934-b222-1fcd247cc463 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2762.781946] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7616b75-e2c9-47c1-9041-e47aafe60722 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2762.794996] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d2:d7:a6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dd98a25d-a7a9-4fb5-8fef-e8df4dbbbf11', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2c611b6e-39e9-428c-a425-1eebe84169dd', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2762.802348] env[62824]: DEBUG oslo.service.loopingcall [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2762.802570] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2762.802765] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c4749330-26db-4040-a76d-b3d78f827e87 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2762.823272] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2762.823272] env[62824]: value = "task-2147228" [ 2762.823272] env[62824]: _type = "Task" [ 2762.823272] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2762.830467] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2147228, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2763.124986] env[62824]: DEBUG oslo_concurrency.lockutils [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Releasing lock "refresh_cache-c68974df-ba6e-4f13-b682-737dfe78f265" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2763.147472] env[62824]: DEBUG nova.virt.hardware [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-01-16T14:27:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='b5361c16f66c293f08e0709357c9761f',container_format='bare',created_at=2025-01-16T15:05:28Z,direct_url=,disk_format='vmdk',id=1c41327c-708c-44f8-b176-faab82421719,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-723048545-shelved',owner='3240be4f32254f54aa87cb045e3d8f31',properties=ImageMetaProps,protected=,size=31669248,status='active',tags=,updated_at=2025-01-16T15:05:41Z,virtual_size=,visibility=), allow threads: False {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 2763.147710] env[62824]: DEBUG nova.virt.hardware [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Flavor limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 2763.147894] env[62824]: DEBUG nova.virt.hardware [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Image limits 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 2763.148115] env[62824]: DEBUG nova.virt.hardware [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Flavor pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 2763.148269] env[62824]: DEBUG nova.virt.hardware [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Image pref 0:0:0 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 2763.148417] env[62824]: DEBUG nova.virt.hardware [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 2763.148624] env[62824]: DEBUG nova.virt.hardware [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 2763.148781] env[62824]: DEBUG nova.virt.hardware [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 2763.148948] env[62824]: DEBUG nova.virt.hardware [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Got 1 possible topologies {{(pid=62824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 2763.149134] env[62824]: DEBUG nova.virt.hardware [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 2763.149306] env[62824]: DEBUG nova.virt.hardware [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 2763.150490] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2550ee32-fbca-45f1-ac1e-ab5fc65ba822 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2763.158304] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c15ed09d-761f-408d-86a3-7a4bd218e26e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2763.171150] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:29:4a:15', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7041d198-66a3-40de-bf7d-cfc036e6ed69', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '18202605-3a10-4c27-989e-74430907f582', 'vif_model': 'vmxnet3'}] {{(pid=62824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 2763.178350] env[62824]: DEBUG oslo.service.loopingcall [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2763.178555] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Creating VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 2763.178743] env[62824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ce6a080b-a1e2-4b4c-a9e8-8535dad2a49d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2763.197032] env[62824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 2763.197032] env[62824]: value = "task-2147229" [ 2763.197032] env[62824]: _type = "Task" [ 2763.197032] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2763.205324] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2147229, 'name': CreateVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2763.332582] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2147228, 'name': CreateVM_Task, 'duration_secs': 0.31505} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2763.332772] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2763.333382] env[62824]: DEBUG oslo_concurrency.lockutils [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/585e45af-6ab3-438f-b022-209c5a367d1a" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2763.333550] env[62824]: DEBUG oslo_concurrency.lockutils [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquired lock "[datastore1] devstack-image-cache_base/585e45af-6ab3-438f-b022-209c5a367d1a" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2763.333960] env[62824]: DEBUG oslo_concurrency.lockutils [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/585e45af-6ab3-438f-b022-209c5a367d1a" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2763.334220] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bee35f73-1051-43b2-ad7e-85a29fab1a0d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2763.338613] env[62824]: DEBUG oslo_vmware.api [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2763.338613] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52f7bdb0-42cb-9a00-aed4-3fbfdeb1b5a9" [ 2763.338613] env[62824]: _type = "Task" [ 2763.338613] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2763.345897] env[62824]: DEBUG oslo_vmware.api [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52f7bdb0-42cb-9a00-aed4-3fbfdeb1b5a9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2763.483960] env[62824]: DEBUG nova.compute.manager [req-fb7ad16c-3cd1-422b-8661-2482e201977c req-523fa192-c644-4f3a-b3f7-0f9cf6f0c600 service nova] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Received event network-changed-2c611b6e-39e9-428c-a425-1eebe84169dd {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2763.484174] env[62824]: DEBUG nova.compute.manager [req-fb7ad16c-3cd1-422b-8661-2482e201977c req-523fa192-c644-4f3a-b3f7-0f9cf6f0c600 service nova] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Refreshing instance network info cache due to event network-changed-2c611b6e-39e9-428c-a425-1eebe84169dd. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2763.484386] env[62824]: DEBUG oslo_concurrency.lockutils [req-fb7ad16c-3cd1-422b-8661-2482e201977c req-523fa192-c644-4f3a-b3f7-0f9cf6f0c600 service nova] Acquiring lock "refresh_cache-02c41e06-f6e9-4b3c-bb29-bee33ab5e13c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2763.484529] env[62824]: DEBUG oslo_concurrency.lockutils [req-fb7ad16c-3cd1-422b-8661-2482e201977c req-523fa192-c644-4f3a-b3f7-0f9cf6f0c600 service nova] Acquired lock "refresh_cache-02c41e06-f6e9-4b3c-bb29-bee33ab5e13c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2763.484687] env[62824]: DEBUG nova.network.neutron [req-fb7ad16c-3cd1-422b-8661-2482e201977c req-523fa192-c644-4f3a-b3f7-0f9cf6f0c600 service nova] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Refreshing network info cache for port 2c611b6e-39e9-428c-a425-1eebe84169dd {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2763.707398] env[62824]: DEBUG oslo_vmware.api [-] Task: {'id': task-2147229, 'name': CreateVM_Task, 'duration_secs': 0.427526} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2763.707744] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Created VM on the ESX host {{(pid=62824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 2763.708337] env[62824]: DEBUG oslo_concurrency.lockutils [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1c41327c-708c-44f8-b176-faab82421719" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2763.708545] env[62824]: DEBUG oslo_concurrency.lockutils [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1c41327c-708c-44f8-b176-faab82421719" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2763.708939] env[62824]: DEBUG oslo_concurrency.lockutils [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1c41327c-708c-44f8-b176-faab82421719" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 2763.709223] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1e969fc7-933d-4af2-8104-e8fef5a72d0b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2763.713710] env[62824]: DEBUG oslo_vmware.api [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Waiting for the task: (returnval){ [ 2763.713710] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]524805e2-7ccc-7a9f-ac1e-60d156672944" [ 2763.713710] env[62824]: _type = "Task" [ 2763.713710] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2763.721615] env[62824]: DEBUG oslo_vmware.api [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]524805e2-7ccc-7a9f-ac1e-60d156672944, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2763.849745] env[62824]: DEBUG oslo_concurrency.lockutils [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Releasing lock "[datastore1] devstack-image-cache_base/585e45af-6ab3-438f-b022-209c5a367d1a" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2763.849964] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Processing image 585e45af-6ab3-438f-b022-209c5a367d1a {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2763.850240] env[62824]: DEBUG oslo_concurrency.lockutils [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/585e45af-6ab3-438f-b022-209c5a367d1a/585e45af-6ab3-438f-b022-209c5a367d1a.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2763.850418] env[62824]: DEBUG oslo_concurrency.lockutils [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquired lock "[datastore1] devstack-image-cache_base/585e45af-6ab3-438f-b022-209c5a367d1a/585e45af-6ab3-438f-b022-209c5a367d1a.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2763.850636] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2763.850892] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f194923e-ce1e-40ea-8ced-1c30c2362449 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2763.868397] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2763.868586] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2763.869329] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f779f972-6ccc-457e-b010-d279f5cc97cb {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2763.874682] env[62824]: DEBUG oslo_vmware.api [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2763.874682] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52886113-5dc3-d13b-4527-db709cde808a" [ 2763.874682] env[62824]: _type = "Task" [ 2763.874682] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2763.887126] env[62824]: DEBUG oslo_vmware.api [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]52886113-5dc3-d13b-4527-db709cde808a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2764.184668] env[62824]: DEBUG nova.network.neutron [req-fb7ad16c-3cd1-422b-8661-2482e201977c req-523fa192-c644-4f3a-b3f7-0f9cf6f0c600 service nova] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Updated VIF entry in instance network info cache for port 2c611b6e-39e9-428c-a425-1eebe84169dd. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2764.185038] env[62824]: DEBUG nova.network.neutron [req-fb7ad16c-3cd1-422b-8661-2482e201977c req-523fa192-c644-4f3a-b3f7-0f9cf6f0c600 service nova] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Updating instance_info_cache with network_info: [{"id": "2c611b6e-39e9-428c-a425-1eebe84169dd", "address": "fa:16:3e:d2:d7:a6", "network": {"id": "5233a794-3514-4912-a857-7e76a66b1e8f", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-910685188-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.129", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0eb41836305847fca773afebdafa01db", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd98a25d-a7a9-4fb5-8fef-e8df4dbbbf11", "external-id": "nsx-vlan-transportzone-707", "segmentation_id": 707, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2c611b6e-39", "ovs_interfaceid": "2c611b6e-39e9-428c-a425-1eebe84169dd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2764.223233] env[62824]: DEBUG oslo_concurrency.lockutils [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1c41327c-708c-44f8-b176-faab82421719" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2764.223466] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Processing image 1c41327c-708c-44f8-b176-faab82421719 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 2764.223692] env[62824]: DEBUG oslo_concurrency.lockutils [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1c41327c-708c-44f8-b176-faab82421719/1c41327c-708c-44f8-b176-faab82421719.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2764.223838] env[62824]: DEBUG oslo_concurrency.lockutils [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1c41327c-708c-44f8-b176-faab82421719/1c41327c-708c-44f8-b176-faab82421719.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2764.224028] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2764.224265] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-123e346b-d97a-4920-bf65-5436af195e9b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2764.239798] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2764.239964] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 2764.240629] env[62824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a92df5b7-ddf6-4e11-a3bb-524b21ce1379 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2764.245130] env[62824]: DEBUG oslo_vmware.api [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Waiting for the task: (returnval){ [ 2764.245130] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]5275ddb6-51e2-3925-231e-0cbf7ec090e4" [ 2764.245130] env[62824]: _type = "Task" [ 2764.245130] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2764.252116] env[62824]: DEBUG oslo_vmware.api [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': session[52762108-595b-6c80-c46c-9e1189c5cddf]5275ddb6-51e2-3925-231e-0cbf7ec090e4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2764.384059] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Preparing fetch location {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 2764.384268] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Fetch image to [datastore1] OSTACK_IMG_8d47de28-38d6-4c56-a822-f75358a06088/OSTACK_IMG_8d47de28-38d6-4c56-a822-f75358a06088.vmdk {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 2764.384456] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Downloading stream optimized image 585e45af-6ab3-438f-b022-209c5a367d1a to [datastore1] OSTACK_IMG_8d47de28-38d6-4c56-a822-f75358a06088/OSTACK_IMG_8d47de28-38d6-4c56-a822-f75358a06088.vmdk on the data store datastore1 as vApp {{(pid=62824) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 2764.384627] env[62824]: DEBUG nova.virt.vmwareapi.images [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Downloading image file data 585e45af-6ab3-438f-b022-209c5a367d1a to the ESX as VM named 'OSTACK_IMG_8d47de28-38d6-4c56-a822-f75358a06088' {{(pid=62824) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 2764.447611] env[62824]: DEBUG oslo_vmware.rw_handles [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 2764.447611] env[62824]: value = "resgroup-9" [ 2764.447611] env[62824]: _type = "ResourcePool" [ 2764.447611] env[62824]: }. {{(pid=62824) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 2764.447886] env[62824]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-d3924e07-e9e0-4ebe-922e-8bbd62684444 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2764.470707] env[62824]: DEBUG oslo_vmware.rw_handles [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lease: (returnval){ [ 2764.470707] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52f461b6-80ac-c8be-a3c3-7c520f481891" [ 2764.470707] env[62824]: _type = "HttpNfcLease" [ 2764.470707] env[62824]: } obtained for vApp import into resource pool (val){ [ 2764.470707] env[62824]: value = "resgroup-9" [ 2764.470707] env[62824]: _type = "ResourcePool" [ 2764.470707] env[62824]: }. {{(pid=62824) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 2764.470931] env[62824]: DEBUG oslo_vmware.api [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the lease: (returnval){ [ 2764.470931] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52f461b6-80ac-c8be-a3c3-7c520f481891" [ 2764.470931] env[62824]: _type = "HttpNfcLease" [ 2764.470931] env[62824]: } to be ready. {{(pid=62824) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 2764.476693] env[62824]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 2764.476693] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52f461b6-80ac-c8be-a3c3-7c520f481891" [ 2764.476693] env[62824]: _type = "HttpNfcLease" [ 2764.476693] env[62824]: } is initializing. {{(pid=62824) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 2764.688646] env[62824]: DEBUG oslo_concurrency.lockutils [req-fb7ad16c-3cd1-422b-8661-2482e201977c req-523fa192-c644-4f3a-b3f7-0f9cf6f0c600 service nova] Releasing lock "refresh_cache-02c41e06-f6e9-4b3c-bb29-bee33ab5e13c" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2764.688809] env[62824]: DEBUG nova.compute.manager [req-fb7ad16c-3cd1-422b-8661-2482e201977c req-523fa192-c644-4f3a-b3f7-0f9cf6f0c600 service nova] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Received event network-vif-plugged-18202605-3a10-4c27-989e-74430907f582 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2764.688946] env[62824]: DEBUG oslo_concurrency.lockutils [req-fb7ad16c-3cd1-422b-8661-2482e201977c req-523fa192-c644-4f3a-b3f7-0f9cf6f0c600 service nova] Acquiring lock "c68974df-ba6e-4f13-b682-737dfe78f265-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2764.689205] env[62824]: DEBUG oslo_concurrency.lockutils [req-fb7ad16c-3cd1-422b-8661-2482e201977c req-523fa192-c644-4f3a-b3f7-0f9cf6f0c600 service nova] Lock "c68974df-ba6e-4f13-b682-737dfe78f265-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2764.689405] env[62824]: DEBUG oslo_concurrency.lockutils [req-fb7ad16c-3cd1-422b-8661-2482e201977c req-523fa192-c644-4f3a-b3f7-0f9cf6f0c600 service nova] Lock "c68974df-ba6e-4f13-b682-737dfe78f265-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2764.689614] env[62824]: DEBUG nova.compute.manager [req-fb7ad16c-3cd1-422b-8661-2482e201977c req-523fa192-c644-4f3a-b3f7-0f9cf6f0c600 service nova] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] No waiting events found dispatching network-vif-plugged-18202605-3a10-4c27-989e-74430907f582 {{(pid=62824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 2764.689813] env[62824]: WARNING nova.compute.manager [req-fb7ad16c-3cd1-422b-8661-2482e201977c req-523fa192-c644-4f3a-b3f7-0f9cf6f0c600 service nova] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Received unexpected event network-vif-plugged-18202605-3a10-4c27-989e-74430907f582 for instance with vm_state shelved_offloaded and task_state spawning. [ 2764.689997] env[62824]: DEBUG nova.compute.manager [req-fb7ad16c-3cd1-422b-8661-2482e201977c req-523fa192-c644-4f3a-b3f7-0f9cf6f0c600 service nova] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Received event network-changed-18202605-3a10-4c27-989e-74430907f582 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2764.690209] env[62824]: DEBUG nova.compute.manager [req-fb7ad16c-3cd1-422b-8661-2482e201977c req-523fa192-c644-4f3a-b3f7-0f9cf6f0c600 service nova] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Refreshing instance network info cache due to event network-changed-18202605-3a10-4c27-989e-74430907f582. {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11653}} [ 2764.690420] env[62824]: DEBUG oslo_concurrency.lockutils [req-fb7ad16c-3cd1-422b-8661-2482e201977c req-523fa192-c644-4f3a-b3f7-0f9cf6f0c600 service nova] Acquiring lock "refresh_cache-c68974df-ba6e-4f13-b682-737dfe78f265" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2764.690597] env[62824]: DEBUG oslo_concurrency.lockutils [req-fb7ad16c-3cd1-422b-8661-2482e201977c req-523fa192-c644-4f3a-b3f7-0f9cf6f0c600 service nova] Acquired lock "refresh_cache-c68974df-ba6e-4f13-b682-737dfe78f265" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2764.690776] env[62824]: DEBUG nova.network.neutron [req-fb7ad16c-3cd1-422b-8661-2482e201977c req-523fa192-c644-4f3a-b3f7-0f9cf6f0c600 service nova] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Refreshing network info cache for port 18202605-3a10-4c27-989e-74430907f582 {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 2764.755525] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Preparing fetch location {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 2764.755861] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Fetch image to [datastore1] OSTACK_IMG_172ab770-ee4b-490b-a451-6a5553efe024/OSTACK_IMG_172ab770-ee4b-490b-a451-6a5553efe024.vmdk {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 2764.756059] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Downloading stream optimized image 1c41327c-708c-44f8-b176-faab82421719 to [datastore1] OSTACK_IMG_172ab770-ee4b-490b-a451-6a5553efe024/OSTACK_IMG_172ab770-ee4b-490b-a451-6a5553efe024.vmdk on the data store datastore1 as vApp {{(pid=62824) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 2764.756181] env[62824]: DEBUG nova.virt.vmwareapi.images [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Downloading image file data 1c41327c-708c-44f8-b176-faab82421719 to the ESX as VM named 'OSTACK_IMG_172ab770-ee4b-490b-a451-6a5553efe024' {{(pid=62824) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 2764.824078] env[62824]: DEBUG oslo_vmware.rw_handles [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 2764.824078] env[62824]: value = "resgroup-9" [ 2764.824078] env[62824]: _type = "ResourcePool" [ 2764.824078] env[62824]: }. {{(pid=62824) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 2764.824403] env[62824]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-e4ca9d7b-5702-47db-b7df-194abbce418b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2764.844584] env[62824]: DEBUG oslo_vmware.rw_handles [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Lease: (returnval){ [ 2764.844584] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52208d99-21a6-1255-1e32-2fb75b16dfd9" [ 2764.844584] env[62824]: _type = "HttpNfcLease" [ 2764.844584] env[62824]: } obtained for vApp import into resource pool (val){ [ 2764.844584] env[62824]: value = "resgroup-9" [ 2764.844584] env[62824]: _type = "ResourcePool" [ 2764.844584] env[62824]: }. {{(pid=62824) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 2764.844821] env[62824]: DEBUG oslo_vmware.api [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Waiting for the lease: (returnval){ [ 2764.844821] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52208d99-21a6-1255-1e32-2fb75b16dfd9" [ 2764.844821] env[62824]: _type = "HttpNfcLease" [ 2764.844821] env[62824]: } to be ready. {{(pid=62824) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 2764.850814] env[62824]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 2764.850814] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52208d99-21a6-1255-1e32-2fb75b16dfd9" [ 2764.850814] env[62824]: _type = "HttpNfcLease" [ 2764.850814] env[62824]: } is initializing. {{(pid=62824) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 2764.979621] env[62824]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 2764.979621] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52f461b6-80ac-c8be-a3c3-7c520f481891" [ 2764.979621] env[62824]: _type = "HttpNfcLease" [ 2764.979621] env[62824]: } is ready. {{(pid=62824) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 2764.979944] env[62824]: DEBUG oslo_vmware.rw_handles [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 2764.979944] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52f461b6-80ac-c8be-a3c3-7c520f481891" [ 2764.979944] env[62824]: _type = "HttpNfcLease" [ 2764.979944] env[62824]: }. {{(pid=62824) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 2764.980735] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21c19ecc-c01a-435f-83c7-776536d5cc8b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2764.988667] env[62824]: DEBUG oslo_vmware.rw_handles [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5236d7aa-9be0-cc79-1178-d249baa4039e/disk-0.vmdk from lease info. {{(pid=62824) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 2764.988896] env[62824]: DEBUG oslo_vmware.rw_handles [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Creating HTTP connection to write to file with size = 31668736 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5236d7aa-9be0-cc79-1178-d249baa4039e/disk-0.vmdk. {{(pid=62824) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 2765.054322] env[62824]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-57209a04-0b08-463f-b024-dfbf6669164b {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2765.352439] env[62824]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 2765.352439] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52208d99-21a6-1255-1e32-2fb75b16dfd9" [ 2765.352439] env[62824]: _type = "HttpNfcLease" [ 2765.352439] env[62824]: } is ready. {{(pid=62824) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 2765.352744] env[62824]: DEBUG oslo_vmware.rw_handles [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 2765.352744] env[62824]: value = "session[52762108-595b-6c80-c46c-9e1189c5cddf]52208d99-21a6-1255-1e32-2fb75b16dfd9" [ 2765.352744] env[62824]: _type = "HttpNfcLease" [ 2765.352744] env[62824]: }. {{(pid=62824) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 2765.353447] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-817cf826-4cf8-4778-815b-5ed1f8d8fb7d {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2765.360308] env[62824]: DEBUG oslo_vmware.rw_handles [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52de5d5e-1d0d-e6fb-7535-0cd69c01cb04/disk-0.vmdk from lease info. {{(pid=62824) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 2765.360477] env[62824]: DEBUG oslo_vmware.rw_handles [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Creating HTTP connection to write to file with size = 31669248 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52de5d5e-1d0d-e6fb-7535-0cd69c01cb04/disk-0.vmdk. {{(pid=62824) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 2765.426946] env[62824]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-02f094a2-8856-4529-9691-99aad20a993e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2765.458989] env[62824]: DEBUG nova.network.neutron [req-fb7ad16c-3cd1-422b-8661-2482e201977c req-523fa192-c644-4f3a-b3f7-0f9cf6f0c600 service nova] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Updated VIF entry in instance network info cache for port 18202605-3a10-4c27-989e-74430907f582. {{(pid=62824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 2765.459401] env[62824]: DEBUG nova.network.neutron [req-fb7ad16c-3cd1-422b-8661-2482e201977c req-523fa192-c644-4f3a-b3f7-0f9cf6f0c600 service nova] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Updating instance_info_cache with network_info: [{"id": "18202605-3a10-4c27-989e-74430907f582", "address": "fa:16:3e:29:4a:15", "network": {"id": "db233b0c-27f3-4a76-ae9c-2b5889bf3d3b", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1825671448-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.192", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3240be4f32254f54aa87cb045e3d8f31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7041d198-66a3-40de-bf7d-cfc036e6ed69", "external-id": "nsx-vlan-transportzone-278", "segmentation_id": 278, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap18202605-3a", "ovs_interfaceid": "18202605-3a10-4c27-989e-74430907f582", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2765.964368] env[62824]: DEBUG oslo_concurrency.lockutils [req-fb7ad16c-3cd1-422b-8661-2482e201977c req-523fa192-c644-4f3a-b3f7-0f9cf6f0c600 service nova] Releasing lock "refresh_cache-c68974df-ba6e-4f13-b682-737dfe78f265" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2766.340930] env[62824]: DEBUG oslo_vmware.rw_handles [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Completed reading data from the image iterator. {{(pid=62824) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 2766.341308] env[62824]: DEBUG oslo_vmware.rw_handles [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5236d7aa-9be0-cc79-1178-d249baa4039e/disk-0.vmdk. {{(pid=62824) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 2766.342431] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e538de5a-981c-457d-8278-8d70373ac0c2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2766.353497] env[62824]: DEBUG oslo_vmware.rw_handles [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5236d7aa-9be0-cc79-1178-d249baa4039e/disk-0.vmdk is in state: ready. {{(pid=62824) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 2766.353687] env[62824]: DEBUG oslo_vmware.rw_handles [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5236d7aa-9be0-cc79-1178-d249baa4039e/disk-0.vmdk. {{(pid=62824) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 2766.353970] env[62824]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-99faae79-f399-48f3-84e2-dc4b3ce15c81 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2766.515606] env[62824]: DEBUG oslo_vmware.rw_handles [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Completed reading data from the image iterator. {{(pid=62824) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 2766.515852] env[62824]: DEBUG oslo_vmware.rw_handles [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52de5d5e-1d0d-e6fb-7535-0cd69c01cb04/disk-0.vmdk. {{(pid=62824) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 2766.516790] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65e9783c-7cf4-42b0-be97-128a5b861fca {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2766.523498] env[62824]: DEBUG oslo_vmware.rw_handles [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52de5d5e-1d0d-e6fb-7535-0cd69c01cb04/disk-0.vmdk is in state: ready. {{(pid=62824) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 2766.523662] env[62824]: DEBUG oslo_vmware.rw_handles [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52de5d5e-1d0d-e6fb-7535-0cd69c01cb04/disk-0.vmdk. {{(pid=62824) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 2766.523924] env[62824]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-25dee493-8fb3-4247-95de-131034b51905 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2766.566719] env[62824]: DEBUG oslo_vmware.rw_handles [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5236d7aa-9be0-cc79-1178-d249baa4039e/disk-0.vmdk. {{(pid=62824) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 2766.566978] env[62824]: INFO nova.virt.vmwareapi.images [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Downloaded image file data 585e45af-6ab3-438f-b022-209c5a367d1a [ 2766.567877] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d277cb5-952c-4c9a-9348-8a76784f09e9 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2766.584369] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-19c31214-bbb5-42fb-a440-3c3f84455525 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2766.614021] env[62824]: INFO nova.virt.vmwareapi.images [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] The imported VM was unregistered [ 2766.616594] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Caching image {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 2766.616827] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Creating directory with path [datastore1] devstack-image-cache_base/585e45af-6ab3-438f-b022-209c5a367d1a {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2766.617148] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-13cec2b7-dd93-463f-b931-97edfbeaae94 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2766.626950] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Created directory with path [datastore1] devstack-image-cache_base/585e45af-6ab3-438f-b022-209c5a367d1a {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2766.627118] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_8d47de28-38d6-4c56-a822-f75358a06088/OSTACK_IMG_8d47de28-38d6-4c56-a822-f75358a06088.vmdk to [datastore1] devstack-image-cache_base/585e45af-6ab3-438f-b022-209c5a367d1a/585e45af-6ab3-438f-b022-209c5a367d1a.vmdk. {{(pid=62824) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 2766.627355] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-8fe5d9a9-00af-4e71-8937-0f6e5e92b61a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2766.633884] env[62824]: DEBUG oslo_vmware.api [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2766.633884] env[62824]: value = "task-2147233" [ 2766.633884] env[62824]: _type = "Task" [ 2766.633884] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2766.641116] env[62824]: DEBUG oslo_vmware.api [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147233, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2766.723478] env[62824]: DEBUG oslo_vmware.rw_handles [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52de5d5e-1d0d-e6fb-7535-0cd69c01cb04/disk-0.vmdk. {{(pid=62824) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 2766.723746] env[62824]: INFO nova.virt.vmwareapi.images [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Downloaded image file data 1c41327c-708c-44f8-b176-faab82421719 [ 2766.724742] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acea89a7-4b7f-430a-9f69-9515255e88a8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2766.742865] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ad7f220a-6eeb-45e6-b938-9f53101e3134 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2766.767193] env[62824]: INFO nova.virt.vmwareapi.images [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] The imported VM was unregistered [ 2766.769076] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Caching image {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 2766.769318] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Creating directory with path [datastore1] devstack-image-cache_base/1c41327c-708c-44f8-b176-faab82421719 {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 2766.769610] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8f7ff405-7335-4524-9e7e-92f5c4e5bc96 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2766.779245] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Created directory with path [datastore1] devstack-image-cache_base/1c41327c-708c-44f8-b176-faab82421719 {{(pid=62824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 2766.779439] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_172ab770-ee4b-490b-a451-6a5553efe024/OSTACK_IMG_172ab770-ee4b-490b-a451-6a5553efe024.vmdk to [datastore1] devstack-image-cache_base/1c41327c-708c-44f8-b176-faab82421719/1c41327c-708c-44f8-b176-faab82421719.vmdk. {{(pid=62824) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 2766.779680] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-86a2c164-5a4a-4ee0-9a3d-0f3389cda8ca {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2766.786284] env[62824]: DEBUG oslo_vmware.api [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Waiting for the task: (returnval){ [ 2766.786284] env[62824]: value = "task-2147235" [ 2766.786284] env[62824]: _type = "Task" [ 2766.786284] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2766.794089] env[62824]: DEBUG oslo_vmware.api [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147235, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2767.144803] env[62824]: DEBUG oslo_vmware.api [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147233, 'name': MoveVirtualDisk_Task} progress is 24%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2767.297171] env[62824]: DEBUG oslo_vmware.api [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147235, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2767.645574] env[62824]: DEBUG oslo_vmware.api [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147233, 'name': MoveVirtualDisk_Task} progress is 46%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2767.798491] env[62824]: DEBUG oslo_vmware.api [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147235, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2768.146014] env[62824]: DEBUG oslo_vmware.api [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147233, 'name': MoveVirtualDisk_Task} progress is 71%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2768.299879] env[62824]: DEBUG oslo_vmware.api [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147235, 'name': MoveVirtualDisk_Task} progress is 9%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2768.647186] env[62824]: DEBUG oslo_vmware.api [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147233, 'name': MoveVirtualDisk_Task} progress is 91%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2768.799257] env[62824]: DEBUG oslo_vmware.api [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147235, 'name': MoveVirtualDisk_Task} progress is 12%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2769.147991] env[62824]: DEBUG oslo_vmware.api [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147233, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2769.300102] env[62824]: DEBUG oslo_vmware.api [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147235, 'name': MoveVirtualDisk_Task} progress is 32%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2769.648786] env[62824]: DEBUG oslo_vmware.api [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147233, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.912119} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2769.649133] env[62824]: INFO nova.virt.vmwareapi.ds_util [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_8d47de28-38d6-4c56-a822-f75358a06088/OSTACK_IMG_8d47de28-38d6-4c56-a822-f75358a06088.vmdk to [datastore1] devstack-image-cache_base/585e45af-6ab3-438f-b022-209c5a367d1a/585e45af-6ab3-438f-b022-209c5a367d1a.vmdk. [ 2769.649352] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Cleaning up location [datastore1] OSTACK_IMG_8d47de28-38d6-4c56-a822-f75358a06088 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 2769.649519] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_8d47de28-38d6-4c56-a822-f75358a06088 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2769.649788] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-705ad63a-30ee-4315-aa00-993750b526f5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2769.657786] env[62824]: DEBUG oslo_vmware.api [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2769.657786] env[62824]: value = "task-2147236" [ 2769.657786] env[62824]: _type = "Task" [ 2769.657786] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2769.666195] env[62824]: DEBUG oslo_vmware.api [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147236, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2769.802189] env[62824]: DEBUG oslo_vmware.api [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147235, 'name': MoveVirtualDisk_Task} progress is 54%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2770.169197] env[62824]: DEBUG oslo_vmware.api [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147236, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.329439} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2770.169653] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2770.169789] env[62824]: DEBUG oslo_concurrency.lockutils [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Releasing lock "[datastore1] devstack-image-cache_base/585e45af-6ab3-438f-b022-209c5a367d1a/585e45af-6ab3-438f-b022-209c5a367d1a.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2770.170055] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/585e45af-6ab3-438f-b022-209c5a367d1a/585e45af-6ab3-438f-b022-209c5a367d1a.vmdk to [datastore1] 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c/02c41e06-f6e9-4b3c-bb29-bee33ab5e13c.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2770.170427] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7ed49fee-8179-4c78-9143-077523ac48d3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2770.179838] env[62824]: DEBUG oslo_vmware.api [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2770.179838] env[62824]: value = "task-2147237" [ 2770.179838] env[62824]: _type = "Task" [ 2770.179838] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2770.188455] env[62824]: DEBUG oslo_vmware.api [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147237, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2770.301464] env[62824]: DEBUG oslo_vmware.api [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147235, 'name': MoveVirtualDisk_Task} progress is 77%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2770.690572] env[62824]: DEBUG oslo_vmware.api [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147237, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2770.804074] env[62824]: DEBUG oslo_vmware.api [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147235, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2771.191449] env[62824]: DEBUG oslo_vmware.api [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147237, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2771.303236] env[62824]: DEBUG oslo_vmware.api [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147235, 'name': MoveVirtualDisk_Task, 'duration_secs': 4.027562} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2771.303523] env[62824]: INFO nova.virt.vmwareapi.ds_util [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_172ab770-ee4b-490b-a451-6a5553efe024/OSTACK_IMG_172ab770-ee4b-490b-a451-6a5553efe024.vmdk to [datastore1] devstack-image-cache_base/1c41327c-708c-44f8-b176-faab82421719/1c41327c-708c-44f8-b176-faab82421719.vmdk. [ 2771.303712] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Cleaning up location [datastore1] OSTACK_IMG_172ab770-ee4b-490b-a451-6a5553efe024 {{(pid=62824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 2771.303876] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_172ab770-ee4b-490b-a451-6a5553efe024 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2771.304207] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-46618859-2099-404c-b9b6-e73ffad5ba1a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2771.311777] env[62824]: DEBUG oslo_vmware.api [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Waiting for the task: (returnval){ [ 2771.311777] env[62824]: value = "task-2147238" [ 2771.311777] env[62824]: _type = "Task" [ 2771.311777] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2771.320670] env[62824]: DEBUG oslo_vmware.api [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147238, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2771.692643] env[62824]: DEBUG oslo_vmware.api [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147237, 'name': CopyVirtualDisk_Task} progress is 46%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2771.822709] env[62824]: DEBUG oslo_vmware.api [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147238, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.287952} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2771.823079] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2771.823264] env[62824]: DEBUG oslo_concurrency.lockutils [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1c41327c-708c-44f8-b176-faab82421719/1c41327c-708c-44f8-b176-faab82421719.vmdk" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2771.823593] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1c41327c-708c-44f8-b176-faab82421719/1c41327c-708c-44f8-b176-faab82421719.vmdk to [datastore1] c68974df-ba6e-4f13-b682-737dfe78f265/c68974df-ba6e-4f13-b682-737dfe78f265.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 2771.823973] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6af2ae00-3aaf-424a-8fb5-f81a501b8a46 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2771.831396] env[62824]: DEBUG oslo_vmware.api [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Waiting for the task: (returnval){ [ 2771.831396] env[62824]: value = "task-2147239" [ 2771.831396] env[62824]: _type = "Task" [ 2771.831396] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2771.840204] env[62824]: DEBUG oslo_vmware.api [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147239, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2772.192222] env[62824]: DEBUG oslo_vmware.api [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147237, 'name': CopyVirtualDisk_Task} progress is 69%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2772.342701] env[62824]: DEBUG oslo_vmware.api [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147239, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2772.692588] env[62824]: DEBUG oslo_vmware.api [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147237, 'name': CopyVirtualDisk_Task} progress is 91%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2772.842509] env[62824]: DEBUG oslo_vmware.api [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147239, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2773.194201] env[62824]: DEBUG oslo_vmware.api [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147237, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.68579} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2773.194600] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/585e45af-6ab3-438f-b022-209c5a367d1a/585e45af-6ab3-438f-b022-209c5a367d1a.vmdk to [datastore1] 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c/02c41e06-f6e9-4b3c-bb29-bee33ab5e13c.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2773.195400] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e004e788-1c19-4ae0-b68d-9798072edbb4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2773.219469] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Reconfiguring VM instance instance-0000007b to attach disk [datastore1] 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c/02c41e06-f6e9-4b3c-bb29-bee33ab5e13c.vmdk or device None with type streamOptimized {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2773.219817] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-16b53dba-8027-45a2-b0bd-eda653019427 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2773.241515] env[62824]: DEBUG oslo_vmware.api [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2773.241515] env[62824]: value = "task-2147240" [ 2773.241515] env[62824]: _type = "Task" [ 2773.241515] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2773.250702] env[62824]: DEBUG oslo_vmware.api [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147240, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2773.344076] env[62824]: DEBUG oslo_vmware.api [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147239, 'name': CopyVirtualDisk_Task} progress is 26%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2773.752951] env[62824]: DEBUG oslo_vmware.api [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147240, 'name': ReconfigVM_Task, 'duration_secs': 0.292494} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2773.753292] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Reconfigured VM instance instance-0000007b to attach disk [datastore1] 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c/02c41e06-f6e9-4b3c-bb29-bee33ab5e13c.vmdk or device None with type streamOptimized {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2773.754197] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-993449ae-a999-4c6a-a1a9-88cbef239ea4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2773.761755] env[62824]: DEBUG oslo_vmware.api [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2773.761755] env[62824]: value = "task-2147241" [ 2773.761755] env[62824]: _type = "Task" [ 2773.761755] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2773.770383] env[62824]: DEBUG oslo_vmware.api [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147241, 'name': Rename_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2773.844257] env[62824]: DEBUG oslo_vmware.api [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147239, 'name': CopyVirtualDisk_Task} progress is 49%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2774.275045] env[62824]: DEBUG oslo_vmware.api [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147241, 'name': Rename_Task, 'duration_secs': 0.161484} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2774.275639] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2774.275916] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ecc7e34b-ea1e-438a-88ae-ba68c56cd572 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2774.284197] env[62824]: DEBUG oslo_vmware.api [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2774.284197] env[62824]: value = "task-2147242" [ 2774.284197] env[62824]: _type = "Task" [ 2774.284197] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2774.298584] env[62824]: DEBUG oslo_vmware.api [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147242, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2774.344043] env[62824]: DEBUG oslo_vmware.api [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147239, 'name': CopyVirtualDisk_Task} progress is 71%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2774.795939] env[62824]: DEBUG oslo_vmware.api [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147242, 'name': PowerOnVM_Task, 'duration_secs': 0.490423} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2774.796278] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2774.845368] env[62824]: DEBUG oslo_vmware.api [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147239, 'name': CopyVirtualDisk_Task} progress is 94%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2774.898283] env[62824]: DEBUG nova.compute.manager [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2774.899323] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c4be2fb-c9ef-40df-a785-4af8f6d3beea {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2775.346994] env[62824]: DEBUG oslo_vmware.api [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147239, 'name': CopyVirtualDisk_Task, 'duration_secs': 3.124022} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2775.348519] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1c41327c-708c-44f8-b176-faab82421719/1c41327c-708c-44f8-b176-faab82421719.vmdk to [datastore1] c68974df-ba6e-4f13-b682-737dfe78f265/c68974df-ba6e-4f13-b682-737dfe78f265.vmdk {{(pid=62824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 2775.352053] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ae15f9b-9c95-4ff1-add0-51e09e9bda6e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2775.375286] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Reconfiguring VM instance instance-0000007c to attach disk [datastore1] c68974df-ba6e-4f13-b682-737dfe78f265/c68974df-ba6e-4f13-b682-737dfe78f265.vmdk or device None with type streamOptimized {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 2775.376440] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-18340c15-7560-4cb6-8dcf-0d0ad2b93202 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2775.396510] env[62824]: DEBUG oslo_vmware.api [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Waiting for the task: (returnval){ [ 2775.396510] env[62824]: value = "task-2147243" [ 2775.396510] env[62824]: _type = "Task" [ 2775.396510] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2775.403885] env[62824]: DEBUG oslo_vmware.api [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147243, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2775.418918] env[62824]: DEBUG oslo_concurrency.lockutils [None req-61a44dff-4d5f-428f-ac61-0145e2318afa tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lock "02c41e06-f6e9-4b3c-bb29-bee33ab5e13c" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 21.107s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2775.852260] env[62824]: DEBUG oslo_concurrency.lockutils [None req-65e37330-681b-4bbe-a545-944f970dea6d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquiring lock "02c41e06-f6e9-4b3c-bb29-bee33ab5e13c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2775.852504] env[62824]: DEBUG oslo_concurrency.lockutils [None req-65e37330-681b-4bbe-a545-944f970dea6d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lock "02c41e06-f6e9-4b3c-bb29-bee33ab5e13c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2775.852714] env[62824]: DEBUG oslo_concurrency.lockutils [None req-65e37330-681b-4bbe-a545-944f970dea6d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquiring lock "02c41e06-f6e9-4b3c-bb29-bee33ab5e13c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2775.852905] env[62824]: DEBUG oslo_concurrency.lockutils [None req-65e37330-681b-4bbe-a545-944f970dea6d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lock "02c41e06-f6e9-4b3c-bb29-bee33ab5e13c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2775.853096] env[62824]: DEBUG oslo_concurrency.lockutils [None req-65e37330-681b-4bbe-a545-944f970dea6d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lock "02c41e06-f6e9-4b3c-bb29-bee33ab5e13c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2775.855229] env[62824]: INFO nova.compute.manager [None req-65e37330-681b-4bbe-a545-944f970dea6d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Terminating instance [ 2775.905239] env[62824]: DEBUG oslo_vmware.api [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147243, 'name': ReconfigVM_Task, 'duration_secs': 0.27072} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2775.905487] env[62824]: DEBUG nova.virt.vmwareapi.volumeops [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Reconfigured VM instance instance-0000007c to attach disk [datastore1] c68974df-ba6e-4f13-b682-737dfe78f265/c68974df-ba6e-4f13-b682-737dfe78f265.vmdk or device None with type streamOptimized {{(pid=62824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 2775.906098] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cde1c739-701d-486a-a1b2-bd0a876d7536 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2775.911964] env[62824]: DEBUG oslo_vmware.api [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Waiting for the task: (returnval){ [ 2775.911964] env[62824]: value = "task-2147244" [ 2775.911964] env[62824]: _type = "Task" [ 2775.911964] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2775.920379] env[62824]: DEBUG oslo_vmware.api [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147244, 'name': Rename_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2776.359475] env[62824]: DEBUG nova.compute.manager [None req-65e37330-681b-4bbe-a545-944f970dea6d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2776.359781] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-65e37330-681b-4bbe-a545-944f970dea6d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2776.360659] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c13edac-afa4-4dfb-9e63-f300461e13d6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2776.368343] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-65e37330-681b-4bbe-a545-944f970dea6d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2776.368571] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fa2f3808-c783-40c5-9b27-7778ed8824ef {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2776.375046] env[62824]: DEBUG oslo_vmware.api [None req-65e37330-681b-4bbe-a545-944f970dea6d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2776.375046] env[62824]: value = "task-2147245" [ 2776.375046] env[62824]: _type = "Task" [ 2776.375046] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2776.382589] env[62824]: DEBUG oslo_vmware.api [None req-65e37330-681b-4bbe-a545-944f970dea6d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147245, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2776.421403] env[62824]: DEBUG oslo_vmware.api [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147244, 'name': Rename_Task} progress is 14%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2776.885388] env[62824]: DEBUG oslo_vmware.api [None req-65e37330-681b-4bbe-a545-944f970dea6d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147245, 'name': PowerOffVM_Task, 'duration_secs': 0.186951} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2776.885647] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-65e37330-681b-4bbe-a545-944f970dea6d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2776.885812] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-65e37330-681b-4bbe-a545-944f970dea6d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2776.886093] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b6ed0910-06c5-42f7-a41a-e0a8c2470bf7 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2776.921463] env[62824]: DEBUG oslo_vmware.api [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147244, 'name': Rename_Task, 'duration_secs': 0.82116} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2776.921710] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Powering on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 2776.921933] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dae921b4-4389-44be-b0f2-72ebe5e1c2d0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2776.928158] env[62824]: DEBUG oslo_vmware.api [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Waiting for the task: (returnval){ [ 2776.928158] env[62824]: value = "task-2147247" [ 2776.928158] env[62824]: _type = "Task" [ 2776.928158] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2776.935387] env[62824]: DEBUG oslo_vmware.api [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147247, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2777.437342] env[62824]: DEBUG oslo_vmware.api [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147247, 'name': PowerOnVM_Task, 'duration_secs': 0.415502} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2777.437614] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Powered on the VM {{(pid=62824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 2777.521401] env[62824]: DEBUG nova.compute.manager [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Checking state {{(pid=62824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 2777.522314] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ea04f16-f589-4d72-8668-c063e231b6c1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2778.041315] env[62824]: DEBUG oslo_concurrency.lockutils [None req-43b7aa08-6965-4cd9-88cb-eb5c839e5d29 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Lock "c68974df-ba6e-4f13-b682-737dfe78f265" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 25.873s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2778.335244] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-65e37330-681b-4bbe-a545-944f970dea6d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2778.335451] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-65e37330-681b-4bbe-a545-944f970dea6d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Deleting contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2778.335670] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-65e37330-681b-4bbe-a545-944f970dea6d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Deleting the datastore file [datastore1] 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2778.335940] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6969e8aa-b5e7-43c1-84cf-5aafe7902cf3 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2778.343286] env[62824]: DEBUG oslo_vmware.api [None req-65e37330-681b-4bbe-a545-944f970dea6d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for the task: (returnval){ [ 2778.343286] env[62824]: value = "task-2147248" [ 2778.343286] env[62824]: _type = "Task" [ 2778.343286] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2778.351168] env[62824]: DEBUG oslo_vmware.api [None req-65e37330-681b-4bbe-a545-944f970dea6d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147248, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2778.856026] env[62824]: DEBUG oslo_vmware.api [None req-65e37330-681b-4bbe-a545-944f970dea6d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Task: {'id': task-2147248, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.130363} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2778.856378] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-65e37330-681b-4bbe-a545-944f970dea6d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2778.856594] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-65e37330-681b-4bbe-a545-944f970dea6d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Deleted contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2778.856839] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-65e37330-681b-4bbe-a545-944f970dea6d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2778.857114] env[62824]: INFO nova.compute.manager [None req-65e37330-681b-4bbe-a545-944f970dea6d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Took 2.50 seconds to destroy the instance on the hypervisor. [ 2778.857473] env[62824]: DEBUG oslo.service.loopingcall [None req-65e37330-681b-4bbe-a545-944f970dea6d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2778.857726] env[62824]: DEBUG nova.compute.manager [-] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2778.857865] env[62824]: DEBUG nova.network.neutron [-] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2779.297270] env[62824]: DEBUG nova.compute.manager [req-24d830a7-b95c-4250-a3cf-fa8d070c1460 req-5b02b531-080c-4444-98f3-052e2a2fc128 service nova] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Received event network-vif-deleted-2c611b6e-39e9-428c-a425-1eebe84169dd {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2779.297473] env[62824]: INFO nova.compute.manager [req-24d830a7-b95c-4250-a3cf-fa8d070c1460 req-5b02b531-080c-4444-98f3-052e2a2fc128 service nova] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Neutron deleted interface 2c611b6e-39e9-428c-a425-1eebe84169dd; detaching it from the instance and deleting it from the info cache [ 2779.297706] env[62824]: DEBUG nova.network.neutron [req-24d830a7-b95c-4250-a3cf-fa8d070c1460 req-5b02b531-080c-4444-98f3-052e2a2fc128 service nova] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2779.770780] env[62824]: DEBUG nova.network.neutron [-] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2779.800886] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-feaf509c-b4c3-4081-8db6-3657a5d1ace5 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2779.811131] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94a76024-bc5f-4220-8a8f-2d712068399f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2779.836619] env[62824]: DEBUG nova.compute.manager [req-24d830a7-b95c-4250-a3cf-fa8d070c1460 req-5b02b531-080c-4444-98f3-052e2a2fc128 service nova] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Detach interface failed, port_id=2c611b6e-39e9-428c-a425-1eebe84169dd, reason: Instance 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c could not be found. {{(pid=62824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11482}} [ 2780.273536] env[62824]: INFO nova.compute.manager [-] [instance: 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c] Took 1.42 seconds to deallocate network for instance. [ 2780.780842] env[62824]: DEBUG oslo_concurrency.lockutils [None req-65e37330-681b-4bbe-a545-944f970dea6d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2780.781113] env[62824]: DEBUG oslo_concurrency.lockutils [None req-65e37330-681b-4bbe-a545-944f970dea6d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2780.781361] env[62824]: DEBUG nova.objects.instance [None req-65e37330-681b-4bbe-a545-944f970dea6d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lazy-loading 'resources' on Instance uuid 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2781.324360] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6af869dd-2d67-4aba-9167-2e549bb84cd2 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2781.332914] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-013b4356-54cd-4dda-8a37-f14b11442d5f {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2781.362076] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09df6871-dde5-4a55-ac2c-43609740dd74 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2781.368874] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f3dcd77-d1e7-48a9-aa21-a529916d2b85 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2781.382638] env[62824]: DEBUG nova.compute.provider_tree [None req-65e37330-681b-4bbe-a545-944f970dea6d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2781.886682] env[62824]: DEBUG nova.scheduler.client.report [None req-65e37330-681b-4bbe-a545-944f970dea6d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2782.391923] env[62824]: DEBUG oslo_concurrency.lockutils [None req-65e37330-681b-4bbe-a545-944f970dea6d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.611s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2782.411889] env[62824]: INFO nova.scheduler.client.report [None req-65e37330-681b-4bbe-a545-944f970dea6d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Deleted allocations for instance 02c41e06-f6e9-4b3c-bb29-bee33ab5e13c [ 2782.919271] env[62824]: DEBUG oslo_concurrency.lockutils [None req-65e37330-681b-4bbe-a545-944f970dea6d tempest-ServerActionsTestOtherB-274633131 tempest-ServerActionsTestOtherB-274633131-project-member] Lock "02c41e06-f6e9-4b3c-bb29-bee33ab5e13c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.067s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2798.774102] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager.update_available_resource {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2799.277073] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2799.277320] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2799.277468] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2799.277623] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62824) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 2799.278530] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e145f496-ece5-4570-af3c-edf1f732d014 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2799.286860] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-babbb172-eabf-48bf-a312-fabd5938f25c {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2799.300669] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bcdc281-75ed-4929-b942-e4d93949d10a {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2799.306775] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbd7fcfc-287a-4037-8c27-c4006db79144 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2799.335324] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181229MB free_disk=175GB free_vcpus=48 pci_devices=None {{(pid=62824) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 2799.335448] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2799.335633] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2800.358756] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Instance c68974df-ba6e-4f13-b682-737dfe78f265 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 2800.359036] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 2800.359135] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=62824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 2800.383696] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60b576ec-ded2-43df-b9e0-0902193d2b7e {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2800.391183] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcf2533e-19d7-4197-a244-e42ee45aefdd {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2800.420934] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11425cb5-0073-44f8-9c0b-0adcc478b543 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2800.428188] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ea0dc83-0866-4ede-bfe4-d0fffff669d1 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2800.441965] env[62824]: DEBUG nova.compute.provider_tree [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2800.945356] env[62824]: DEBUG nova.scheduler.client.report [None req-41517485-76de-4e8f-a181-a012faec848d None None] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2801.451077] env[62824]: DEBUG nova.compute.resource_tracker [None req-41517485-76de-4e8f-a181-a012faec848d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62824) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 2801.451423] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.116s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2803.446527] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2805.772806] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2805.773211] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] Starting heal instance info cache {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10445}} [ 2806.305333] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquiring lock "refresh_cache-c68974df-ba6e-4f13-b682-737dfe78f265" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 2806.305499] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Acquired lock "refresh_cache-c68974df-ba6e-4f13-b682-737dfe78f265" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 2806.305648] env[62824]: DEBUG nova.network.neutron [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Forcefully refreshing network info cache for instance {{(pid=62824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 2807.527485] env[62824]: DEBUG nova.network.neutron [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Updating instance_info_cache with network_info: [{"id": "18202605-3a10-4c27-989e-74430907f582", "address": "fa:16:3e:29:4a:15", "network": {"id": "db233b0c-27f3-4a76-ae9c-2b5889bf3d3b", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1825671448-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.192", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3240be4f32254f54aa87cb045e3d8f31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7041d198-66a3-40de-bf7d-cfc036e6ed69", "external-id": "nsx-vlan-transportzone-278", "segmentation_id": 278, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap18202605-3a", "ovs_interfaceid": "18202605-3a10-4c27-989e-74430907f582", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2808.030183] env[62824]: DEBUG oslo_concurrency.lockutils [None req-41517485-76de-4e8f-a181-a012faec848d None None] Releasing lock "refresh_cache-c68974df-ba6e-4f13-b682-737dfe78f265" {{(pid=62824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 2808.030407] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Updated the network info_cache for instance {{(pid=62824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10516}} [ 2808.030610] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2808.030772] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2808.030918] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2808.031078] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2808.772584] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2813.772674] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 2813.773108] env[62824]: DEBUG nova.compute.manager [None req-41517485-76de-4e8f-a181-a012faec848d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62824) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11064}} [ 2814.430466] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a8b77d6f-e40f-460e-ba2d-38fd17086b16 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Acquiring lock "c68974df-ba6e-4f13-b682-737dfe78f265" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2814.430466] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a8b77d6f-e40f-460e-ba2d-38fd17086b16 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Lock "c68974df-ba6e-4f13-b682-737dfe78f265" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2814.430680] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a8b77d6f-e40f-460e-ba2d-38fd17086b16 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Acquiring lock "c68974df-ba6e-4f13-b682-737dfe78f265-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2814.430805] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a8b77d6f-e40f-460e-ba2d-38fd17086b16 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Lock "c68974df-ba6e-4f13-b682-737dfe78f265-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2814.430990] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a8b77d6f-e40f-460e-ba2d-38fd17086b16 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Lock "c68974df-ba6e-4f13-b682-737dfe78f265-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2814.434675] env[62824]: INFO nova.compute.manager [None req-a8b77d6f-e40f-460e-ba2d-38fd17086b16 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Terminating instance [ 2814.938612] env[62824]: DEBUG nova.compute.manager [None req-a8b77d6f-e40f-460e-ba2d-38fd17086b16 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Start destroying the instance on the hypervisor. {{(pid=62824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 2814.938993] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-a8b77d6f-e40f-460e-ba2d-38fd17086b16 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Destroying instance {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 2814.939810] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a48d632-a473-4abf-ab41-fdd801396ddc {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2814.947788] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8b77d6f-e40f-460e-ba2d-38fd17086b16 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Powering off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 2814.948028] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b556fd71-ab63-4b6d-9177-12dc03070f09 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2814.954716] env[62824]: DEBUG oslo_vmware.api [None req-a8b77d6f-e40f-460e-ba2d-38fd17086b16 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Waiting for the task: (returnval){ [ 2814.954716] env[62824]: value = "task-2147250" [ 2814.954716] env[62824]: _type = "Task" [ 2814.954716] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2814.962505] env[62824]: DEBUG oslo_vmware.api [None req-a8b77d6f-e40f-460e-ba2d-38fd17086b16 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147250, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2815.464098] env[62824]: DEBUG oslo_vmware.api [None req-a8b77d6f-e40f-460e-ba2d-38fd17086b16 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147250, 'name': PowerOffVM_Task, 'duration_secs': 0.160833} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2815.464376] env[62824]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8b77d6f-e40f-460e-ba2d-38fd17086b16 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Powered off the VM {{(pid=62824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 2815.464548] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-a8b77d6f-e40f-460e-ba2d-38fd17086b16 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Unregistering the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 2815.464810] env[62824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-16b86191-43ba-4531-a600-bd50708d3b23 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2815.630389] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-a8b77d6f-e40f-460e-ba2d-38fd17086b16 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Unregistered the VM {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 2815.630645] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-a8b77d6f-e40f-460e-ba2d-38fd17086b16 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Deleting contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 2815.630792] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8b77d6f-e40f-460e-ba2d-38fd17086b16 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Deleting the datastore file [datastore1] c68974df-ba6e-4f13-b682-737dfe78f265 {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 2815.631080] env[62824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8ff858a5-64b6-40e1-85f4-63c9efa445c8 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2815.638572] env[62824]: DEBUG oslo_vmware.api [None req-a8b77d6f-e40f-460e-ba2d-38fd17086b16 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Waiting for the task: (returnval){ [ 2815.638572] env[62824]: value = "task-2147252" [ 2815.638572] env[62824]: _type = "Task" [ 2815.638572] env[62824]: } to complete. {{(pid=62824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 2815.646111] env[62824]: DEBUG oslo_vmware.api [None req-a8b77d6f-e40f-460e-ba2d-38fd17086b16 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147252, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 2816.148616] env[62824]: DEBUG oslo_vmware.api [None req-a8b77d6f-e40f-460e-ba2d-38fd17086b16 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Task: {'id': task-2147252, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.148266} completed successfully. {{(pid=62824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 2816.149061] env[62824]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8b77d6f-e40f-460e-ba2d-38fd17086b16 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Deleted the datastore file {{(pid=62824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 2816.149061] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-a8b77d6f-e40f-460e-ba2d-38fd17086b16 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Deleted contents of the VM from datastore datastore1 {{(pid=62824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 2816.149233] env[62824]: DEBUG nova.virt.vmwareapi.vmops [None req-a8b77d6f-e40f-460e-ba2d-38fd17086b16 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Instance destroyed {{(pid=62824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 2816.149405] env[62824]: INFO nova.compute.manager [None req-a8b77d6f-e40f-460e-ba2d-38fd17086b16 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Took 1.21 seconds to destroy the instance on the hypervisor. [ 2816.149642] env[62824]: DEBUG oslo.service.loopingcall [None req-a8b77d6f-e40f-460e-ba2d-38fd17086b16 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 2816.149833] env[62824]: DEBUG nova.compute.manager [-] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Deallocating network for instance {{(pid=62824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 2816.149931] env[62824]: DEBUG nova.network.neutron [-] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] deallocate_for_instance() {{(pid=62824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 2816.592180] env[62824]: DEBUG nova.compute.manager [req-8a1e7cef-3ef4-4406-96ac-7b4b276f6965 req-2b3770f5-13d0-417e-b9d0-9c6a77cbf3ce service nova] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Received event network-vif-deleted-18202605-3a10-4c27-989e-74430907f582 {{(pid=62824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11648}} [ 2816.592367] env[62824]: INFO nova.compute.manager [req-8a1e7cef-3ef4-4406-96ac-7b4b276f6965 req-2b3770f5-13d0-417e-b9d0-9c6a77cbf3ce service nova] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Neutron deleted interface 18202605-3a10-4c27-989e-74430907f582; detaching it from the instance and deleting it from the info cache [ 2816.592676] env[62824]: DEBUG nova.network.neutron [req-8a1e7cef-3ef4-4406-96ac-7b4b276f6965 req-2b3770f5-13d0-417e-b9d0-9c6a77cbf3ce service nova] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2817.071208] env[62824]: DEBUG nova.network.neutron [-] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Updating instance_info_cache with network_info: [] {{(pid=62824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 2817.095592] env[62824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-275be9cf-6c3e-4600-ab21-4e343a0236a6 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2817.106189] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1794633-bb24-4c24-ae65-5ec55f31dcb0 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2817.129579] env[62824]: DEBUG nova.compute.manager [req-8a1e7cef-3ef4-4406-96ac-7b4b276f6965 req-2b3770f5-13d0-417e-b9d0-9c6a77cbf3ce service nova] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Detach interface failed, port_id=18202605-3a10-4c27-989e-74430907f582, reason: Instance c68974df-ba6e-4f13-b682-737dfe78f265 could not be found. {{(pid=62824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11482}} [ 2817.574558] env[62824]: INFO nova.compute.manager [-] [instance: c68974df-ba6e-4f13-b682-737dfe78f265] Took 1.42 seconds to deallocate network for instance. [ 2818.082128] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a8b77d6f-e40f-460e-ba2d-38fd17086b16 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 2818.082429] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a8b77d6f-e40f-460e-ba2d-38fd17086b16 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 2818.082650] env[62824]: DEBUG nova.objects.instance [None req-a8b77d6f-e40f-460e-ba2d-38fd17086b16 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Lazy-loading 'resources' on Instance uuid c68974df-ba6e-4f13-b682-737dfe78f265 {{(pid=62824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 2818.622884] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f10128f-1ff8-4d81-a86e-6540615a8d90 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2818.628744] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0105770a-1293-4d12-8012-550d1bac0788 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2818.661560] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ba3fa99-e472-4067-bdf5-2c8d16cd4bd4 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2818.670519] env[62824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae6afa3d-99fb-472d-8b9d-7a4886af8a13 {{(pid=62824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 2818.685544] env[62824]: DEBUG nova.compute.provider_tree [None req-a8b77d6f-e40f-460e-ba2d-38fd17086b16 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Inventory has not changed in ProviderTree for provider: bf95157b-a274-42de-9ccf-9851128a44a1 {{(pid=62824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 2819.189060] env[62824]: DEBUG nova.scheduler.client.report [None req-a8b77d6f-e40f-460e-ba2d-38fd17086b16 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Inventory has not changed for provider bf95157b-a274-42de-9ccf-9851128a44a1 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 2819.694638] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a8b77d6f-e40f-460e-ba2d-38fd17086b16 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.612s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2819.715199] env[62824]: INFO nova.scheduler.client.report [None req-a8b77d6f-e40f-460e-ba2d-38fd17086b16 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Deleted allocations for instance c68974df-ba6e-4f13-b682-737dfe78f265 [ 2820.223182] env[62824]: DEBUG oslo_concurrency.lockutils [None req-a8b77d6f-e40f-460e-ba2d-38fd17086b16 tempest-AttachVolumeShelveTestJSON-913590524 tempest-AttachVolumeShelveTestJSON-913590524-project-member] Lock "c68974df-ba6e-4f13-b682-737dfe78f265" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.793s {{(pid=62824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 2821.768541] env[62824]: DEBUG oslo_service.periodic_task [None req-41517485-76de-4e8f-a181-a012faec848d None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}}